import gradio as gr import requests import io from PIL import Image import json import os import logging import math from tqdm import tqdm import time logging.basicConfig(level=logging.DEBUG) with open('loras.json', 'r') as f: loras = json.load(f) def update_selection(selected_state: gr.SelectData): logging.debug(f"Inside update_selection, selected_state: {selected_state}") selected_lora_index = selected_state.index # Changed this line selected_lora = loras[selected_lora_index] new_placeholder = f"Type a prompt for {selected_lora['title']}" return ( gr.update(placeholder=new_placeholder), selected_state ) def run_lora(prompt, selected_state, progress=gr.Progress(track_tqdm=True)): logging.debug(f"Inside run_lora, selected_state: {selected_state}") if not selected_state: logging.error("selected_state is None or empty.") raise gr.Error("You must select a LoRA before proceeding.") # Popup error when no LoRA is selected selected_lora_index = selected_state.index # Changed this line selected_lora = loras[selected_lora_index] api_url = f"https://api-inference.huggingface.co/models/{selected_lora['repo']}" trigger_word = selected_lora["trigger_word"] #token = os.getenv("API_TOKEN") payload = { "inputs": f"{prompt} {trigger_word}", "parameters":{"negative_prompt": "bad art, ugly, watermark, deformed"}, } #headers = {"Authorization": f"Bearer {token}"} # Add a print statement to display the API request print(f"API Request: {api_url}") #print(f"API Headers: {headers}") print(f"API Payload: {payload}") error_count = 0 pbar = tqdm(total=None, desc="Loading model") while(True): response = requests.post(api_url, json=payload) if response.status_code == 200: return Image.open(io.BytesIO(response.content)) elif response.status_code == 503: #503 is triggered when the model is doing cold boot. It also gives you a time estimate from when the model is loaded but it is not super precise time.sleep(1) pbar.update(1) elif response.status_code == 500 and error_count < 5: print(response.content) time.sleep(1) error_count += 1 continue else: logging.error(f"API Error: {response.status_code}") raise gr.Error("API Error: Unable to fetch the image.") # Raise a Gradio error here with gr.Blocks(css="custom.css") as app: title = gr.Markdown("# artificialguybr LoRA portfolio") description = gr.Markdown( "### This is my portfolio. Follow me on Twitter [@artificialguybr](https://twitter.com/artificialguybr). \n" "**Note**: The speed and generation quality are for demonstration purposes. " "For best quality, use Auto or Comfy. \n" "**Warning**: The API might take some time to deliver the image. \n" # Added warning about API delay "Special thanks to Hugging Face for their free inference API." ) selected_state = gr.State() with gr.Row(): gallery = gr.Gallery( [(item["image"], item["title"]) for item in loras], label="LoRA Gallery", allow_preview=False, columns=3 ) with gr.Column(): prompt_title = gr.Markdown("### Click on a LoRA in the gallery to select it") with gr.Row(): prompt = gr.Textbox(label="Prompt", show_label=False, lines=1, max_lines=1, placeholder="Type a prompt after selecting a LoRA") button = gr.Button("Run") result = gr.Image(interactive=False, label="Generated Image") gallery.select( update_selection, outputs=[prompt, selected_state] ) prompt.submit( fn=run_lora, inputs=[prompt, selected_state], outputs=[result] ) button.click( fn=run_lora, inputs=[prompt, selected_state], outputs=[result] ) app.queue(max_size=20, concurrency_count=5) app.launch()