import os import sys import ctypes import pathlib from typing import Optional, List import enum from pathlib import Path import argparse import gradio as gr import minigpt4_library from huggingface_hub import hf_hub_download model_path = hf_hub_download(repo_id='maknee/minigpt4-13b-ggml', filename='minigpt4-13B-f16.bin', repo_type='dataset') llm_model_path = hf_hub_download(repo_id='maknee/ggml-vicuna-v0-quantized', filename='ggml-vicuna-13B-v0-q5_k.bin', repo_type='dataset') title = """

MiniGPT-4.cpp Demo

""" description = """

This is the demo of MiniGPT-4 with ggml (cpu only!). Upload your images and start chatting!

""" article = """
""" global minigpt4_chatbot minigpt4_chatbot: minigpt4_library.MiniGPT4ChatBot def user(message, history): history = history or [] # Append the user's message to the conversation history history.append([message, ""]) return "", history def chat(history, limit: int = 1024, temp: float = 0.8, top_k: int = 40, top_p: float = 0.9, repeat_penalty: float = 1.1): history = history or [] message = history[-1][0] history[-1][1] = "" for output in minigpt4_chatbot.generate( message, limit = int(limit), temp = float(temp), top_k = int(top_k), top_p = float(top_p), ): answer = output history[-1][1] += answer # stream the response yield history, history def clear_state(history, chat_message, image): history = [] minigpt4_chatbot.reset_chat() return history, gr.update(value=None, interactive=True), gr.update(placeholder='Upload image first', interactive=False), gr.update(value="Upload & Start Chat", interactive=True) def upload_image(image, history): if image is None: return None, None, gr.update(interactive=True), history history = [] minigpt4_chatbot.upload_image(image.convert('RGB')) return gr.update(interactive=False), gr.update(interactive=True, placeholder='Type and press Enter'), gr.update(value="Start Chatting", interactive=False), history def start(): with gr.Blocks() as demo: gr.Markdown(title) gr.Markdown(description) gr.Markdown(article) with gr.Row(): with gr.Column(scale=0.5): image = gr.Image(type="pil") upload_button = gr.Button(value="Upload & Start Chat", interactive=True, variant="primary") max_tokens = gr.Slider(1, 1024, label="Max Tokens", step=1, value=128) temperature = gr.Slider(0.0, 1.0, label="Temperature", step=0.05, value=0.8) top_p = gr.Slider(0.0, 1.0, label="Top P", step=0.05, value=0.95) top_k = gr.Slider(0, 100, label="Top K", step=1, value=40) repeat_penalty = gr.Slider(0.0, 2.0, label="Repetition Penalty", step=0.1, value=1.1) with gr.Column(): chatbot = gr.Chatbot(label='MiniGPT-4') message = gr.Textbox(label='User', placeholder='Upload image first', interactive=False) history = gr.State() with gr.Row(): submit = gr.Button(value="Send message", variant="secondary").style(full_width=True) clear = gr.Button(value="Reset", variant="secondary").style(full_width=False) # stop = gr.Button(value="Stop", variant="secondary").style(full_width=False) clear.click(clear_state, inputs=[history, image, message], outputs=[history, image, message, upload_button], queue=False) upload_button.click(upload_image, inputs=[image, history], outputs=[image, message, upload_button, history]) submit_click_event = submit.click( fn=user, inputs=[message, history], outputs=[message, history], queue=True ).then( fn=chat, inputs=[history, max_tokens, temperature, top_p, top_k, repeat_penalty], outputs=[chatbot, history], queue=True ) message_submit_event = message.submit( fn=user, inputs=[message, history], outputs=[message, history], queue=True ).then( fn=chat, inputs=[history, max_tokens, temperature, top_p, top_k, repeat_penalty], outputs=[chatbot, history], queue=True ) # stop.click(fn=None, inputs=None, outputs=None, cancels=[submit_click_event, message_submit_event], queue=False) demo.launch(enable_queue=True) minigpt4_chatbot = minigpt4_library.MiniGPT4ChatBot(model_path, llm_model_path, verbosity=minigpt4_library.Verbosity.SILENT) start()