Spaces:
Runtime error
Runtime error
Change model from mistral to MistralTrix
Browse files
app.py
CHANGED
@@ -13,7 +13,7 @@ def respond(message, chat_history,instruction,model,temperature):
|
|
13 |
if model == "Llama2-Chat":
|
14 |
model = "meta-llama/Llama-2-7b-chat-hf"
|
15 |
else:
|
16 |
-
model="
|
17 |
|
18 |
client = InferenceClient(model=f"{model}",timeout=30)
|
19 |
formatted_prompt = format_chat_prompt(message,chat_history,instruction)
|
|
|
13 |
if model == "Llama2-Chat":
|
14 |
model = "meta-llama/Llama-2-7b-chat-hf"
|
15 |
else:
|
16 |
+
model="CultriX/MistralTrix-v1"
|
17 |
|
18 |
client = InferenceClient(model=f"{model}",timeout=30)
|
19 |
formatted_prompt = format_chat_prompt(message,chat_history,instruction)
|