""" Visual QA Tool - A tool for answering questions about images This module provides functionality to analyze images and answer questions about them. It leverages powerful vision-language models (VLMs) to understand image content and respond to natural language questions about the images. The module offers two implementations: 1. VisualQATool class - Uses Hugging Face's IDEFICS-2 model 2. visualizer function - Uses OpenAI's GPT-4o model with vision capabilities Both implementations handle image loading, processing, and API communication to provide detailed responses about image content. Environment variables required: - OPENAI_API_KEY: API key for OpenAI (for the visualizer function) """ import base64 import json import mimetypes import os import uuid from io import BytesIO import PIL.Image import requests from dotenv import load_dotenv from huggingface_hub import InferenceClient from smolagents import Tool, tool # Load environment variables from .env file load_dotenv(override=True) def process_images_and_text(image_path, query, client): """ Process images and text using the IDEFICS-2 model from Hugging Face. This function handles the formatting of prompts and images for the IDEFICS-2 model, which is a powerful vision-language model capable of understanding images and text. Args: image_path (str): Path to the image file to analyze query (str): The question or instruction about the image client (InferenceClient): Hugging Face inference client for the model Returns: str: The model's response to the query about the image """ from transformers import AutoProcessor # Format messages for the chat template messages = [ { "role": "user", "content": [ {"type": "image"}, {"type": "text", "text": query}, ], }, ] # Load the processor for the IDEFICS-2 model idefics_processor = AutoProcessor.from_pretrained("HuggingFaceM4/idefics2-8b-chatty") prompt_with_template = idefics_processor.apply_chat_template(messages, add_generation_prompt=True) # Define a nested function to encode local images def encode_local_image(image_path): """ Encode a local image file to a base64 string for API transmission. Args: image_path (str): Path to the local image file Returns: str: Base64-encoded image with proper formatting for the API """ # Load image and convert to RGB format image = PIL.Image.open(image_path).convert("RGB") # Convert the image to a base64 string buffer = BytesIO() image.save(buffer, format="JPEG") # Use the appropriate format (e.g., JPEG, PNG) base64_image = base64.b64encode(buffer.getvalue()).decode("utf-8") # Add string formatting required by the endpoint image_string = f"data:image/jpeg;base64,{base64_image}" return image_string # Encode the image and insert it into the prompt template image_string = encode_local_image(image_path) prompt_with_images = prompt_with_template.replace("", "![]({}) ").format(image_string) # Prepare the payload for the API request payload = { "inputs": prompt_with_images, "parameters": { "return_full_text": False, "max_new_tokens": 200, # Limit response length }, } # Send the request to the API and parse the response return json.loads(client.post(json=payload).decode())[0] # Function to encode images for API transmission def encode_image(image_path): """ Encode an image for API transmission, handling both URLs and local files. If the image_path is a URL, the function will download the image first. Args: image_path (str): Path or URL to the image Returns: str: Base64-encoded image string """ # Handle URL-based images by downloading them first if image_path.startswith("http"): # Set up a user agent to avoid being blocked by websites user_agent = "Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/119.0.0.0 Safari/537.36 Edg/119.0.0.0" request_kwargs = { "headers": {"User-Agent": user_agent}, "stream": True, # Stream the download for large files } # Send a HTTP request to the URL response = requests.get(image_path, **request_kwargs) response.raise_for_status() # Raise an exception for HTTP errors content_type = response.headers.get("content-type", "") # Determine the file extension from the content type extension = mimetypes.guess_extension(content_type) if extension is None: extension = ".download" # Default extension if unknown # Generate a unique filename and save the downloaded image fname = str(uuid.uuid4()) + extension download_path = os.path.abspath(os.path.join("downloads", fname)) with open(download_path, "wb") as fh: for chunk in response.iter_content(chunk_size=512): fh.write(chunk) # Update the image_path to the local downloaded file image_path = download_path # Encode the local image file to base64 with open(image_path, "rb") as image_file: return base64.b64encode(image_file.read()).decode("utf-8") def resize_image(image_path): """ Resize an image to half its original dimensions. This function is used when the original image is too large for the API. Args: image_path (str): Path to the image file Returns: str: Path to the resized image """ # Open and get dimensions of the image img = PIL.Image.open(image_path) width, height = img.size # Resize to half the original dimensions img = img.resize((int(width / 2), int(height / 2))) # Save with a new filename new_image_path = f"resized_{image_path}" img.save(new_image_path) return new_image_path class VisualQATool(Tool): """ A tool that can answer questions about images using the IDEFICS-2 model. This class implements the Tool interface from smolagents and provides functionality to analyze images and answer questions about them. """ name = "visualizer" description = "A tool that can answer questions about attached images." inputs = { "image_path": { "description": "The path to the image on which to answer the question", "type": "string", }, "question": {"description": "the question to answer", "type": "string", "nullable": True}, } output_type = "string" # Initialize the Hugging Face inference client for IDEFICS-2 client = InferenceClient("HuggingFaceM4/idefics2-8b-chatty") def forward(self, image_path: str, question: str | None = None) -> str: """ Process an image and answer a question about it. If no question is provided, the function will generate a detailed caption. Args: image_path (str): Path to the image file question (str, optional): Question to answer about the image Returns: str: Answer to the question or a caption for the image """ output = "" add_note = False # If no question is provided, default to generating a caption if not question: add_note = True question = "Please write a detailed caption for this image." try: # Try to process the image and question output = process_images_and_text(image_path, question, self.client) except Exception as e: print(e) # If the image is too large, resize it and try again if "Payload Too Large" in str(e): new_image_path = resize_image(image_path) output = process_images_and_text(new_image_path, question, self.client) # Add a note if we generated a caption instead of answering a question if add_note: output = ( f"You did not provide a particular question, so here is a detailed caption for the image: {output}" ) return output @tool def visualizer(image_path: str, question: str | None = None) -> str: """ A tool that can answer questions about attached images using OpenAI's GPT-4o model. This function provides an alternative implementation using OpenAI's vision capabilities instead of the Hugging Face model used in VisualQATool. Args: image_path: The path to the image on which to answer the question. This should be a local path to downloaded image. question: The question to answer. Returns: str: Answer to the question or a caption for the image """ import mimetypes import os import requests from .visual_qa import encode_image # If no question is provided, default to generating a caption add_note = False if not question: add_note = True question = "Please write a detailed caption for this image." # Validate input if not isinstance(image_path, str): raise Exception("You should provide at least `image_path` string argument to this tool!") # Determine the MIME type and encode the image mime_type, _ = mimetypes.guess_type(image_path) base64_image = encode_image(image_path) # Prepare the payload for the OpenAI API request payload = { "model": "gpt-4o", # Using GPT-4o with vision capabilities "messages": [ { "role": "user", "content": [ {"type": "text", "text": question}, {"type": "image_url", "image_url": {"url": f"data:{mime_type};base64,{base64_image}"}}, ], } ], "max_tokens": 1000, # Limit response length } # Set up headers with API key headers = {"Content-Type": "application/json", "Authorization": f"Bearer {os.getenv('OPENAI_API_KEY')}"} # Send the request to the OpenAI API response = requests.post("https://api.openai.com/v1/chat/completions", headers=headers, json=payload) # Parse the response try: output = response.json()["choices"][0]["message"]["content"] except Exception: raise Exception(f"Response format unexpected: {response.json()}") # Add a note if we generated a caption instead of answering a question if add_note: output = f"You did not provide a particular question, so here is a detailed caption for the image: {output}" return output