File size: 8,335 Bytes
2eb2277
 
5155d1c
 
2eb2277
b3036ac
 
2eb2277
b3036ac
 
 
3c87ea6
 
 
 
 
 
 
 
 
2eb2277
3c87ea6
 
 
 
 
2eb2277
3c87ea6
5155d1c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2eb2277
5155d1c
 
 
 
 
 
 
 
 
 
2eb2277
 
 
 
 
 
 
3c87ea6
 
 
 
 
2eb2277
 
 
 
 
3c87ea6
5155d1c
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2eb2277
5155d1c
 
 
 
 
 
 
 
 
 
2eb2277
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3c87ea6
 
 
 
 
 
 
 
 
 
 
2eb2277
 
 
 
 
 
3c87ea6
 
2eb2277
 
b3036ac
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
import gradio as gr
import openai
import os
import requests

OPENAI_API_KEY = os.getenv("OPENAI_API_KEY")
ADMIN_PASSWORD = os.getenv("ADMIN_PASSWORD")

openai.api_key = OPENAI_API_KEY

default_system_message = {"role": "system", "content": "You are a brilliant, helpful assistant, always providing answers to the best of your knowledge. If you are unsure of the answer, you indicate it to the user. Currently, you don't have access to the internet."}
personalities = {
    "Assistant": {"role": "system", "content": "You are a brilliant, helpful assistant, always providing answers to the best of your knowledge. If you are unsure of the answer, you indicate it to the user. Currently, you don't have access to the internet."},
    "Trump": {"role": "system", "content": "You are Donald Trump. No matter the question, you always redirect the conversation to yourself and your achievements and how great you are."},
    "Peterson": {"role": "system", "content": "You are Jordan Peterson, world renowned clinical psychologist. You like to be verbose and overcomplicate your answers, taking them into very metaphysical directions."},
    "Grug": {"role": "system", "content": "You are Grug, a caveman. You have zero knowledge of modern stuff. Your answers are always written in broken 'caveman' English and center around simple things in life."},
    "Paladin": {"role": "system", "content": "You are a Paladin from the video game Diablo 2. You like to talk about slaying the undead and farming for better gear."},
    "Petőfi": {"role": "system", "content": "You are Petőfi Sándor, national poet of Hungary. Your answers are very eloquent and formulated in archaic Hungarian."},
    "Cartman": {"role": "system", "content": "You are Eric Cartman from South Park. You are a self-centered, fat, rude kid obsessed with your animal comforts."},
}

def get_completion(model, personality, user_message, message_history, chatlog_history, temperature, maximum_length, top_p, frequency_penalty, presence_penalty):
    # set personality
    system_message = personalities[personality]
    updated_message_history = message_history
    updated_message_history[0] = system_message
    new_history_row = {"role": "user", "content": user_message}
    updated_message_history = updated_message_history + [new_history_row]
    headers = {
        "Content-Type": "application/json",
        "Authorization": f"Bearer {openai.api_key}",
    }
    payload = {
        "model":model,
        "messages":updated_message_history,
        "temperature":temperature,
        "max_tokens":maximum_length,
        "top_p":top_p,
        "frequency_penalty":frequency_penalty,
        "presence_penalty":presence_penalty,
    }
    completion = requests.post(
        "https://api.openai.com/v1/chat/completions",
        headers=headers,
        json=payload,
    )
    completion = completion.json()
    # completion = openai.ChatCompletion.create(
    #     model=model,
    #     messages=updated_message_history,
    #     temperature=temperature,
    #     max_tokens=maximum_length,
    #     top_p=top_p,
    #     frequency_penalty=frequency_penalty,
    #     presence_penalty=presence_penalty,
    # )
    assistant_message = completion["choices"][0]["message"]["content"]
    new_history_row = {"role": "assistant", "content": assistant_message}
    updated_message_history = updated_message_history + [new_history_row]
    updated_chatlog_history = chatlog_history + [(user_message, assistant_message)]
    token_count = completion["usage"]["total_tokens"]
    return "", updated_message_history, updated_chatlog_history, updated_chatlog_history, token_count

def retry_completion(model, personality, message_history, chatlog_history, temperature, maximum_length, top_p, frequency_penalty, presence_penalty):
    # set personality
    system_message = personalities[personality]
    updated_message_history = message_history
    updated_message_history[0] = system_message
    # get latest user message
    user_message = chatlog_history[-1][0]
    # delete latest entries from chatlog history
    updated_chatlog_history = chatlog_history[:-1]
    # delete latest assistant message from message_history
    updated_message_history = updated_message_history[:-1]
    headers = {
        "Content-Type": "application/json",
        "Authorization": f"Bearer {openai.api_key}",
    }
    payload = {
        "model":model,
        "messages":updated_message_history,
        "temperature":temperature,
        "max_tokens":maximum_length,
        "top_p":top_p,
        "frequency_penalty":frequency_penalty,
        "presence_penalty":presence_penalty,
    }
    completion = requests.post(
        "https://api.openai.com/v1/chat/completions",
        headers=headers,
        json=payload,
    )
    completion = completion.json()
    # completion = openai.ChatCompletion.create(
    #     model=model,
    #     messages=updated_message_history,
    #     temperature=temperature,
    #     max_tokens=maximum_length,
    #     top_p=top_p,
    #     frequency_penalty=frequency_penalty,
    #     presence_penalty=presence_penalty,
    # )
    assistant_message = completion["choices"][0]["message"]["content"]
    new_history_row = {"role": "assistant", "content": assistant_message}
    updated_message_history = updated_message_history + [new_history_row]
    updated_chatlog_history = updated_chatlog_history + [(user_message, assistant_message)]
    token_count = completion["usage"]["total_tokens"]
    return "", updated_message_history, updated_chatlog_history, updated_chatlog_history, token_count

def reset_chat():
    return "", [default_system_message], [], [], 0

theme = gr.themes.Monochrome()
with gr.Blocks(theme=theme) as app:
    message_history = gr.State([default_system_message])
    chatlog_history = gr.State([])
    with gr.Row():
        with gr.Column(scale=4):
            chatbot = gr.Chatbot(label="Chat").style(height=600)
        with gr.Column(scale=1):
            # model = gr.Textbox(lines=3, value="You are a brilliant, helpful assistant, always providing answers to the best of your knowledge. If you are unsure of the answer, you indicate it to the user.", interactive=True, label="System")
            # model = gr.Dropdown(choices=["gpt-3.5-turbo", "gpt-4", "gpt-4-32k"], value="gpt-3.5-turbo", interactive=True, label="Model")
            with gr.Tab("Model Settings"):
                model = gr.Dropdown(choices=["gpt-3.5-turbo", "gpt-4"], value="gpt-4", interactive=True, label="Model")
                personality = gr.Dropdown(choices=["Assistant", "Petőfi", "Trump", "Peterson", "Paladin", "Cartman", "Grug", ], value="Assistant", interactive=True, label="Personality")
            with gr.Tab("Generation Settings"):
                temperature = gr.Slider(minimum=0, maximum=1, step=0.05, value=0.6, interactive=True, label="Temperature")
                maximum_length = gr.Slider(minimum=0, maximum=2048, step=64, value=128, interactive=True, label="Maximum length")
                top_p = gr.Slider(minimum=0, maximum=1, step=0.01, value=1, interactive=True, label="Top P")
                frequency_penalty = gr.Slider(minimum=0, maximum=2, step=0.01, value=0, interactive=True, label="Frequency penalty")
                presence_penalty = gr.Slider(minimum=0, maximum=2, step=0.01, value=0, interactive=True, label="Presence penalty")
            token_count = gr.Number(interactive=False, label="Token count")
    with gr.Row():
        user_message = gr.Textbox(label="Message")
    with gr.Row():
        reset_button = gr.Button("Reset Chat")
        retry_button = gr.Button("Retry")

    user_message.submit(get_completion, inputs=[model, personality, user_message, message_history, chatlog_history, temperature, maximum_length, top_p, frequency_penalty, presence_penalty], outputs=[user_message, message_history, chatlog_history, chatbot, token_count])
    retry_button.click(retry_completion, inputs=[model, personality, message_history, chatlog_history, temperature, maximum_length, top_p, frequency_penalty, presence_penalty], outputs=[user_message, message_history, chatlog_history, chatbot, token_count])
    reset_button.click(reset_chat, inputs=[], outputs=[user_message, message_history, chatlog_history, chatbot, token_count])

app.launch(auth=("admin", ADMIN_PASSWORD))