File size: 1,458 Bytes
3603925
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
import gradio as gr
import requests

def generate_message(prompt, model):
    url = 'https://api.together.xyz/inference'
    headers = {
        'Authorization': 'Bearer ee16b2ea8440ef7142ac5e21f5d096ee2dfb9c0c206acff4818a35b91ae59cef',
        'accept': 'application/json',
        'content-type': 'application/json'
    }
    data = {
        "model": model,
        "prompt": prompt,
        "max_tokens": 512,
        "stop": [
            "</s>",
            "[INST]",
            "<<SYS>>",
            "<</SYS>>",
            "[/INST]",
            "<s>"
        ],
        "temperature": 0.75,
        "top_p": 0.2,
        "top_k": 10
    }
    response = requests.post(url, headers=headers, json=data)
    result = response.json()
    return result['output']['choices'][0]['text'].strip()

def generate_message_with_submit(input_text, model):
    return generate_message(input_text, model)

iface = gr.Interface(
    fn=generate_message_with_submit,
    inputs=[
        gr.TextArea(lines=20, label="Enter Prompt"),
        gr.Radio(['meta-llama/Llama-2-70b-chat-hf', 'iamplus/model1-70b'], label="Model",value='meta-llama/Llama-2-70b-chat-hf')
    ],
    outputs=gr.TextArea(label="AI Assistant Message"),
    title="AI Assistant Message Generator",
    description="Generate an appropriate message for an AI assistant to let the user know that it's working on a task and will get back soon.",
    allow_flagging=False
)

iface.launch(share=True)