api_for_unity / app.py
ldhldh's picture
Update app.py
98e337c
raw
history blame
No virus
1.77 kB
from threading import Thread
import gradio as gr
import inspect
from gradio import routes
from typing import List, Type
import requests, os, re, asyncio, queue
import math
import time
import datetime
import requests, json
from huggingface_hub import hf_hub_download
from llama_cpp import Llama
loop = asyncio.get_event_loop()
# Monkey patch
def get_types(cls_set: List[Type], component: str):
docset = []
types = []
if component == "input":
for cls in cls_set:
doc = inspect.getdoc(cls)
doc_lines = doc.split("\n")
docset.append(doc_lines[1].split(":")[-1])
types.append(doc_lines[1].split(")")[0].split("(")[-1])
else:
for cls in cls_set:
doc = inspect.getdoc(cls)
doc_lines = doc.split("\n")
docset.append(doc_lines[-1].split(":")[-1])
types.append(doc_lines[-1].split(")")[0].split("(")[-1])
return docset, types
routes.get_types = get_types
hf_hub_download(repo_id='StarFox7/Llama-2-ko-7B-chat-gguf', filename='Llama-2-ko-7B-chat-gguf-q4_0.bin', local_dir='./')
llm = Llama(model_path = 'Llama-2-ko-7B-chat-gguf-q4_0.bin',
n_ctx=512,
)
# App code
def chat(x):
#prom = f"λ‹€μŒμ€ A와 B의 역할극이야. λ„ˆλŠ” Bμ•Ό. A와 λŒ€ν™”ν•˜κ³  μžˆμ–΄. μΉœκ΅¬μ—κ²Œ μΉœκ·Όν•˜κ³  κ°„κ²°ν•˜κ²Œ 잘 λŒ€λ‹΅ν•΄μ€˜.\n\n### A:\n{x}\n\n### B:\n"
#output = llm(prom, max_tokens=20, stop=["###"], echo=True)
#return output['choices'][0]['text'][len(prom):-1]
return "AI μ‘λ‹΅μž…λ‹ˆλ‹€."
with gr.Blocks() as demo:
count = 0
aa = gr.Interface(
fn=chat,
inputs=["text"],
outputs="text",
description="call",
)
demo.queue(max_size=32).launch(enable_queue=True)