import torch import os from transformers import AutoProcessor from PIL import Image model_path = "D:/nighttest/nightshadeblip_high.pth" test_path = 'D:/nighttest/test/' model = torch.load(model_path) processor = AutoProcessor.from_pretrained("Salesforce/blip-image-captioning-base") model.eval() for filename in os.listdir(test_path): file_path = os.path.join(test_path, filename) if os.path.isfile(file_path): image = Image.open(file_path) device = "cuda" inputs = processor(images=image, return_tensors="pt").to(device) pixel_values = inputs.pixel_values generated_ids = model.generate(pixel_values=pixel_values, max_length=50) generated_caption = processor.batch_decode(generated_ids, skip_special_tokens=True)[0] print(f"[{generated_caption}] {file_path}")