chansung's picture
Update app/main.py
6008614 verified
raw
history blame
1.64 kB
import gradio as gr
from openai import AsyncOpenAI
base_url = "http://127.0.0.1:8080/v1"
client = AsyncOpenAI(base_url=base_url, api_key="-")
"""
frequency_penalty: Optional[float] = None,
logit_bias: Optional[List[float]] = None,
logprobs: Optional[bool] = None,
top_logprobs: Optional[int] = None,
max_tokens: Optional[int] = None,
n: Optional[int] = None,
presence_penalty: Optional[float] = None,
stream: bool = False,
seed: Optional[int] = None,
temperature: Optional[float] = None,
top_p: Optional[float] = None,
tools: Optional[List[Tool]] = None,
tool_choice: Optional[str] = None,
"""
def _default_parameters():
return {
"max_tokens": 256,
"stream": True,
"temperature": 0.9,
}
def _translate_messages(history):
messages = []
for conv in history:
messages.append({"role":"user", "content":conv[0]})
messages.append({"role":"assistant", "content":conv[1]})
return messages
async def echo(message, history):
parameters = _default_parameters()
messages = _translate_messages(history)
messages.append({"role":"user", "content":message})
responses = await client.chat.completions.create(
model="tgi", messages=messages, **parameters
)
full_resp = ""
async for resp in responses:
full_resp = full_resp + resp.choices[0].delta.content
yield full_resp
demo = gr.ChatInterface(
fn=echo,
examples=["hello", "how are you?", "What is Large Language Model?"],
title="Space of Gradio βž• Text Generation Inference",
multimodal=False
)
demo.queue().launch(server_name="0.0.0.0", server_port=3000)