Spaces:
Running
Running
Update model temperature setting
Browse files
app.py
CHANGED
@@ -14,7 +14,7 @@ from utils import COMMUNITY_POSTFIX_URL, get_model_config, check_format, models_
|
|
14 |
logged_event_handler, DEBUG_MODE, DEBUG_MODEL, log_debug, log_info, log_error, log_warning
|
15 |
from log_chat import log_chat
|
16 |
|
17 |
-
|
18 |
BUTTON_WIDTH = 160
|
19 |
DEFAULT_OPT_OUT_VALUE = DEBUG_MODE
|
20 |
|
@@ -114,10 +114,11 @@ def run_chat_inference(history, message, state):
|
|
114 |
state["stop_flag"] = False
|
115 |
error = None
|
116 |
model_name = model_config.get('MODEL_NAME')
|
|
|
117 |
|
118 |
# Reinitialize the OpenAI client with a random endpoint from the list
|
119 |
setup_model(model_config.get('MODEL_KEY'))
|
120 |
-
log_info(f"Using model {model_name} with endpoint {model_config.get('base_url')}")
|
121 |
|
122 |
if len(history) == 0:
|
123 |
state["chat_id"] = uuid4().hex
|
@@ -372,7 +373,7 @@ def run_chat_inference(history, message, state):
|
|
372 |
stream = openai_client.chat.completions.create(
|
373 |
model=model_name,
|
374 |
messages=api_messages,
|
375 |
-
temperature=
|
376 |
stream=True
|
377 |
)
|
378 |
except Exception as e:
|
@@ -387,7 +388,7 @@ def run_chat_inference(history, message, state):
|
|
387 |
model_name=model_name,
|
388 |
prompt=message,
|
389 |
history=history,
|
390 |
-
info={"is_reasoning": model_config.get("REASONING"), "temperature":
|
391 |
"stopped": True, "error": str(e)},
|
392 |
)
|
393 |
else:
|
@@ -477,7 +478,7 @@ def run_chat_inference(history, message, state):
|
|
477 |
model_name=model_name,
|
478 |
prompt=message,
|
479 |
history=history,
|
480 |
-
info={"is_reasoning": model_config.get("REASONING"), "temperature":
|
481 |
"stopped": state["stop_flag"]},
|
482 |
)
|
483 |
|
|
|
14 |
logged_event_handler, DEBUG_MODE, DEBUG_MODEL, log_debug, log_info, log_error, log_warning
|
15 |
from log_chat import log_chat
|
16 |
|
17 |
+
DEFAULT_MODEL_TEMPERATURE = 0.6
|
18 |
BUTTON_WIDTH = 160
|
19 |
DEFAULT_OPT_OUT_VALUE = DEBUG_MODE
|
20 |
|
|
|
114 |
state["stop_flag"] = False
|
115 |
error = None
|
116 |
model_name = model_config.get('MODEL_NAME')
|
117 |
+
temperature = model_config.get('TEMPERATURE', DEFAULT_MODEL_TEMPERATURE)
|
118 |
|
119 |
# Reinitialize the OpenAI client with a random endpoint from the list
|
120 |
setup_model(model_config.get('MODEL_KEY'))
|
121 |
+
log_info(f"Using model {model_name} (temperature: {temperature}) with endpoint {model_config.get('base_url')}")
|
122 |
|
123 |
if len(history) == 0:
|
124 |
state["chat_id"] = uuid4().hex
|
|
|
373 |
stream = openai_client.chat.completions.create(
|
374 |
model=model_name,
|
375 |
messages=api_messages,
|
376 |
+
temperature=temperature,
|
377 |
stream=True
|
378 |
)
|
379 |
except Exception as e:
|
|
|
388 |
model_name=model_name,
|
389 |
prompt=message,
|
390 |
history=history,
|
391 |
+
info={"is_reasoning": model_config.get("REASONING"), "temperature": temperature,
|
392 |
"stopped": True, "error": str(e)},
|
393 |
)
|
394 |
else:
|
|
|
478 |
model_name=model_name,
|
479 |
prompt=message,
|
480 |
history=history,
|
481 |
+
info={"is_reasoning": model_config.get("REASONING"), "temperature": temperature,
|
482 |
"stopped": state["stop_flag"]},
|
483 |
)
|
484 |
|
utils.py
CHANGED
@@ -20,7 +20,8 @@ models_config = {
|
|
20 |
"VLLM_API_URL_LIST": os.environ.get("VLLM_API_URL_LIST_APRIEL_1_5_15B"),
|
21 |
"AUTH_TOKEN": os.environ.get("AUTH_TOKEN"),
|
22 |
"REASONING": True,
|
23 |
-
"MULTIMODAL": True
|
|
|
24 |
},
|
25 |
# "Apriel-Nemotron-15b-Thinker": {
|
26 |
# "MODEL_DISPLAY_NAME": "Apriel-Nemotron-15b-Thinker",
|
|
|
20 |
"VLLM_API_URL_LIST": os.environ.get("VLLM_API_URL_LIST_APRIEL_1_5_15B"),
|
21 |
"AUTH_TOKEN": os.environ.get("AUTH_TOKEN"),
|
22 |
"REASONING": True,
|
23 |
+
"MULTIMODAL": True,
|
24 |
+
"TEMPERATURE": 0.6
|
25 |
},
|
26 |
# "Apriel-Nemotron-15b-Thinker": {
|
27 |
# "MODEL_DISPLAY_NAME": "Apriel-Nemotron-15b-Thinker",
|