import torch import torchvision.transforms as transforms from PIL import Image import numpy as np from sklearn.metrics.pairwise import cosine_similarity from torchvision.models import resnet50 from torchvision.datasets import ImageFolder from torch.utils.data import DataLoader # Load a pre-trained ResNet-50 model model = resnet50(pretrained=True) model.eval() # Define a function to preprocess images def preprocess_image(image_path): transform = transforms.Compose([ transforms.Resize((224, 224)), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]), ]) image = Image.open(image_path) image = transform(image).unsqueeze(0) # Add a batch dimension return image # Load your ideal subset of images ideal_image_paths = ["/content/trunck.jpg", "t4.jpg"] # Replace with your ideal image file paths ideal_embeddings = [] for image_path in ideal_image_paths: image = preprocess_image(image_path) with torch.no_grad(): embedding = model(image).squeeze().numpy() ideal_embeddings.append(embedding) # Load a set of candidate images candidate_image_paths = ["/content/trunck2.jpg", "t3.jpg", "car.jpg",] # Replace with your candidate image file paths candidate_embeddings = [] for image_path in candidate_image_paths: image = preprocess_image(image_path) with torch.no_grad(): embedding = model(image).squeeze().numpy() candidate_embeddings.append(embedding) # Calculate similarities between ideal and candidate images using cosine similarity similarities = cosine_similarity(ideal_embeddings, candidate_embeddings) # Print the similarity matrix print(similarities) import torch from transformers import SwinTransformer, SwinTransformerImageProcessor import torchvision.transforms as transforms from PIL import Image import numpy as np from sklearn.metrics.pairwise import cosine_similarity # Load the pretrained Swin Transformer model and image processor model_name = "microsoft/Swin-Transformer-base-patch4-in22k" model = SwinTransformer.from_pretrained(model_name) processor = SwinTransformerImageProcessor.from_pretrained(model_name) # Define a function to preprocess images def preprocess_image(image_path): transform = transforms.Compose([ transforms.Resize((224, 224)), transforms.ToTensor(), transforms.Normalize(mean=[0.485, 0.456, 0.406], std=[0.229, 0.224, 0.225]), ]) image = Image.open(image_path) inputs = processor(images=image, return_tensors="pt") return inputs # Load your ideal and candidate subsets of images ideal_image_paths = ["ideal_image1.jpg", "ideal_image2.jpg", "ideal_image3.jpg"] # Replace with your ideal image file paths candidate_image_paths = ["candidate_image1.jpg", "candidate_image2.jpg", "candidate_image3.jpg"] # Replace with your candidate image file paths # Calculate cosine similarities between ideal and candidate images similarities = [] for ideal_path in ideal_image_paths: ideal_embedding = None inputs_ideal = preprocess_image(ideal_path) with torch.no_grad(): output_ideal = model(**inputs_ideal) ideal_embedding = output_ideal['pixel_values'][0].cpu().numpy() for candidate_path in candidate_image_paths: candidate_embedding = None inputs_candidate = preprocess_image(candidate_path) with torch.no_grad(): output_candidate = model(**inputs_candidate) candidate_embedding = output_candidate['pixel_values'][0].cpu().numpy() # Calculate cosine similarity between ideal and candidate embeddings similarity = cosine_similarity([ideal_embedding], [candidate_embedding])[0][0] similarities.append((ideal_path, candidate_path, similarity)) # Set a similarity threshold (e.g., 0.7) threshold = 0.7 # Find similar image pairs based on the threshold similar_pairs = [] for ideal_path, candidate_path, similarity in similarities: if similarity > threshold: similar_pairs.append((ideal_path, candidate_path)) # Print similar image pairs for pair in similar_pairs: print(f"Similar images: {pair[0]} and {pair[1]}")