Spaces:
Sleeping
Sleeping
动态endpoint
Browse files- crazy_functions/询问多个大语言模型.py +1 -1
- request_llm/bridge_chatgpt.py +6 -2
- toolbox.py +0 -1
crazy_functions/询问多个大语言模型.py
CHANGED
@@ -16,7 +16,7 @@ def 同时问询(txt, llm_kwargs, plugin_kwargs, chatbot, history, system_prompt
|
|
16 |
chatbot.append((txt, "正在同时咨询ChatGPT和ChatGLM……"))
|
17 |
yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 # 由于请求gpt需要一段时间,我们先及时地做一次界面更新
|
18 |
|
19 |
-
llm_kwargs['llm_model'] = '
|
20 |
gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
|
21 |
inputs=txt, inputs_show_user=txt,
|
22 |
llm_kwargs=llm_kwargs, chatbot=chatbot, history=history,
|
|
|
16 |
chatbot.append((txt, "正在同时咨询ChatGPT和ChatGLM……"))
|
17 |
yield from update_ui(chatbot=chatbot, history=history) # 刷新界面 # 由于请求gpt需要一段时间,我们先及时地做一次界面更新
|
18 |
|
19 |
+
llm_kwargs['llm_model'] = 'gpt-3.5-turbo&api2d-gpt-3.5-turbo'
|
20 |
gpt_say = yield from request_gpt_model_in_new_thread_with_ui_alive(
|
21 |
inputs=txt, inputs_show_user=txt,
|
22 |
llm_kwargs=llm_kwargs, chatbot=chatbot, history=history,
|
request_llm/bridge_chatgpt.py
CHANGED
@@ -60,7 +60,9 @@ def predict_no_ui_long_connection(inputs, llm_kwargs, history=[], sys_prompt="",
|
|
60 |
while True:
|
61 |
try:
|
62 |
# make a POST request to the API endpoint, stream=False
|
63 |
-
|
|
|
|
|
64 |
json=payload, stream=True, timeout=TIMEOUT_SECONDS); break
|
65 |
except requests.exceptions.ReadTimeout as e:
|
66 |
retry += 1
|
@@ -148,7 +150,9 @@ def predict(inputs, llm_kwargs, plugin_kwargs, chatbot, history=[], system_promp
|
|
148 |
while True:
|
149 |
try:
|
150 |
# make a POST request to the API endpoint, stream=True
|
151 |
-
|
|
|
|
|
152 |
json=payload, stream=True, timeout=TIMEOUT_SECONDS);break
|
153 |
except:
|
154 |
retry += 1
|
|
|
60 |
while True:
|
61 |
try:
|
62 |
# make a POST request to the API endpoint, stream=False
|
63 |
+
from .bridge_all import model_info
|
64 |
+
endpoint = model_info[llm_kwargs['llm_model']]['endpoint']
|
65 |
+
response = requests.post(endpoint, headers=headers, proxies=proxies,
|
66 |
json=payload, stream=True, timeout=TIMEOUT_SECONDS); break
|
67 |
except requests.exceptions.ReadTimeout as e:
|
68 |
retry += 1
|
|
|
150 |
while True:
|
151 |
try:
|
152 |
# make a POST request to the API endpoint, stream=True
|
153 |
+
from .bridge_all import model_info
|
154 |
+
endpoint = model_info[llm_kwargs['llm_model']]['endpoint']
|
155 |
+
response = requests.post(endpoint, headers=headers, proxies=proxies,
|
156 |
json=payload, stream=True, timeout=TIMEOUT_SECONDS);break
|
157 |
except:
|
158 |
retry += 1
|
toolbox.py
CHANGED
@@ -36,7 +36,6 @@ def ArgsGeneralWrapper(f):
|
|
36 |
llm_kwargs = {
|
37 |
'api_key': cookies['api_key'],
|
38 |
'llm_model': llm_model,
|
39 |
-
'endpoint': model_info[llm_model]['endpoint'],
|
40 |
'top_p':top_p,
|
41 |
'max_length': max_length,
|
42 |
'temperature':temperature,
|
|
|
36 |
llm_kwargs = {
|
37 |
'api_key': cookies['api_key'],
|
38 |
'llm_model': llm_model,
|
|
|
39 |
'top_p':top_p,
|
40 |
'max_length': max_length,
|
41 |
'temperature':temperature,
|