Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -15,21 +15,26 @@ def respond(
|
|
15 |
temperature,
|
16 |
top_p,
|
17 |
):
|
18 |
-
messages = [{"role": "system", "content": system_message}]
|
19 |
|
20 |
-
# Add history to messages
|
21 |
for val in history:
|
22 |
-
if val[0]
|
23 |
-
|
24 |
-
if val[1]:
|
25 |
-
messages.append({"role": "assistant", "content": val[1]})
|
26 |
|
27 |
-
|
28 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
29 |
|
30 |
# Generate the response using the Mistral model
|
31 |
response = llm.create_chat_completion(messages=messages)
|
32 |
-
|
|
|
33 |
return response["choices"][0]["message"]["content"] # Adjust based on your model's output format
|
34 |
|
35 |
# Set up Gradio Chat Interface
|
|
|
15 |
temperature,
|
16 |
top_p,
|
17 |
):
|
18 |
+
messages = [{"role": "system", "content": system_message or "You are a friendly Chatbot."}]
|
19 |
|
20 |
+
# Add history to messages, ensuring no None values
|
21 |
for val in history:
|
22 |
+
user_message = val[0] if val[0] is not None else ""
|
23 |
+
assistant_message = val[1] if val[1] is not None else ""
|
|
|
|
|
24 |
|
25 |
+
if user_message:
|
26 |
+
messages.append({"role": "user", "content": user_message})
|
27 |
+
if assistant_message:
|
28 |
+
messages.append({"role": "assistant", "content": assistant_message})
|
29 |
+
|
30 |
+
# Add the current user message, ensure it's not None
|
31 |
+
if message:
|
32 |
+
messages.append({"role": "user", "content": message})
|
33 |
|
34 |
# Generate the response using the Mistral model
|
35 |
response = llm.create_chat_completion(messages=messages)
|
36 |
+
print("response:", response)
|
37 |
+
|
38 |
return response["choices"][0]["message"]["content"] # Adjust based on your model's output format
|
39 |
|
40 |
# Set up Gradio Chat Interface
|