:zap: [Enhance] Quieter openai auth, use cffi to request hf-chat id, and console token count for exceeds
Browse files
messagers/token_checker.py
CHANGED
@@ -40,5 +40,7 @@ class TokenChecker:
|
|
40 |
|
41 |
def check_token_limit(self):
|
42 |
if self.get_token_redundancy() <= 0:
|
43 |
-
raise ValueError(
|
|
|
|
|
44 |
return True
|
|
|
40 |
|
41 |
def check_token_limit(self):
|
42 |
if self.get_token_redundancy() <= 0:
|
43 |
+
raise ValueError(
|
44 |
+
f"Prompt exceeded token limit: {self.count_tokens()} > {self.get_token_limit()}"
|
45 |
+
)
|
46 |
return True
|
networks/huggingchat_streamer.py
CHANGED
@@ -1,7 +1,9 @@
|
|
1 |
import copy
|
2 |
import json
|
3 |
import re
|
|
|
4 |
import requests
|
|
|
5 |
|
6 |
from tclogger import logger
|
7 |
|
@@ -30,7 +32,7 @@ class HuggingchatRequester:
|
|
30 |
request_body.update(extra_body)
|
31 |
logger.note(f"> hf-chat ID:", end=" ")
|
32 |
|
33 |
-
res =
|
34 |
request_url,
|
35 |
headers=HUGGINGCHAT_POST_HEADERS,
|
36 |
json=request_body,
|
|
|
1 |
import copy
|
2 |
import json
|
3 |
import re
|
4 |
+
|
5 |
import requests
|
6 |
+
from curl_cffi import requests as cffi_requests
|
7 |
|
8 |
from tclogger import logger
|
9 |
|
|
|
32 |
request_body.update(extra_body)
|
33 |
logger.note(f"> hf-chat ID:", end=" ")
|
34 |
|
35 |
+
res = cffi_requests.post(
|
36 |
request_url,
|
37 |
headers=HUGGINGCHAT_POST_HEADERS,
|
38 |
json=request_body,
|
networks/openai_streamer.py
CHANGED
@@ -171,18 +171,21 @@ class OpenaiStreamer:
|
|
171 |
|
172 |
def check_token_limit(self, messages: list[dict]):
|
173 |
token_limit = TOKEN_LIMIT_MAP[self.model]
|
174 |
-
|
175 |
-
|
176 |
-
)
|
177 |
if token_redundancy <= 0:
|
178 |
-
raise ValueError(
|
|
|
|
|
179 |
return True
|
180 |
|
181 |
-
def chat_response(self, messages: list[dict]):
|
182 |
self.check_token_limit(messages)
|
|
|
183 |
requester = OpenaiRequester()
|
184 |
requester.auth()
|
185 |
-
|
|
|
186 |
|
187 |
def chat_return_generator(self, stream_response: requests.Response, verbose=False):
|
188 |
content_offset = 0
|
|
|
171 |
|
172 |
def check_token_limit(self, messages: list[dict]):
|
173 |
token_limit = TOKEN_LIMIT_MAP[self.model]
|
174 |
+
token_count = self.count_tokens(messages)
|
175 |
+
token_redundancy = int(token_limit - TOKEN_RESERVED - token_count)
|
|
|
176 |
if token_redundancy <= 0:
|
177 |
+
raise ValueError(
|
178 |
+
f"Prompt exceeded token limit: {token_count} > {token_limit}"
|
179 |
+
)
|
180 |
return True
|
181 |
|
182 |
+
def chat_response(self, messages: list[dict], verbose=False):
|
183 |
self.check_token_limit(messages)
|
184 |
+
logger.enter_quiet(not verbose)
|
185 |
requester = OpenaiRequester()
|
186 |
requester.auth()
|
187 |
+
logger.exit_quiet(not verbose)
|
188 |
+
return requester.chat_completions(messages, verbose=verbose)
|
189 |
|
190 |
def chat_return_generator(self, stream_response: requests.Response, verbose=False):
|
191 |
content_offset = 0
|