Spaces:
Runtime error
Runtime error
seawolf2357
commited on
Commit
โข
9e7ded1
1
Parent(s):
3ca345f
Update app.py
Browse files
app.py
CHANGED
@@ -72,22 +72,26 @@ async def generate_response(message):
|
|
72 |
conversation_history.append({"role": "user", "content": user_input})
|
73 |
logging.debug(f'๋ํ ํ์คํ ๋ฆฌ ์
๋ฐ์ดํธ๋จ: {conversation_history}')
|
74 |
|
75 |
-
#
|
76 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
77 |
logging.debug(f'๋ชจ๋ธ์ ๋ณด๋ผ ๋ฉ์์ง: {messages}')
|
78 |
|
79 |
-
# ์ฌ๋ฐ๋ฅด๊ฒ ๊ต์ฐจํ๋๋ก ๋ํ ๋ด์ญ์ ํํฐ๋ง
|
80 |
-
filtered_messages = []
|
81 |
-
for i in range(len(messages)):
|
82 |
-
if i == 0 or (messages[i]['role'] != messages[i-1]['role']):
|
83 |
-
filtered_messages.append(messages[i])
|
84 |
-
|
85 |
-
logging.debug(f'ํํฐ๋ง๋ ๋ฉ์์ง: {filtered_messages}')
|
86 |
-
|
87 |
# ๋ชจ๋ธ ํธ์ถ
|
88 |
loop = asyncio.get_event_loop()
|
89 |
response = await loop.run_in_executor(None, lambda: hf_client.chat_completion(
|
90 |
-
messages=
|
91 |
|
92 |
full_response = []
|
93 |
for part in response:
|
|
|
72 |
conversation_history.append({"role": "user", "content": user_input})
|
73 |
logging.debug(f'๋ํ ํ์คํ ๋ฆฌ ์
๋ฐ์ดํธ๋จ: {conversation_history}')
|
74 |
|
75 |
+
# ๋ํ ๋ด์ญ์ด ๋๋ฌด ๊ธธ ๊ฒฝ์ฐ, ์ต๋ ๊ธธ์ด๋ฅผ ์ ์งํ๊ธฐ ์ํด ์๋ถ๋ถ์ ์๋ผ๋
|
76 |
+
if len(conversation_history) > 20:
|
77 |
+
conversation_history = conversation_history[-20:]
|
78 |
+
|
79 |
+
# ๋ํ ๋ด์ญ์์ ์ญํ ์ด ์ ๋๋ก ๊ต์ฐจํ๋๋ก ํ์ธ
|
80 |
+
filtered_conversation = []
|
81 |
+
last_role = None
|
82 |
+
for message in conversation_history:
|
83 |
+
if message['role'] != last_role:
|
84 |
+
filtered_conversation.append(message)
|
85 |
+
last_role = message['role']
|
86 |
+
|
87 |
+
# ์์คํ
๋ฉ์์ง์ ํํฐ๋ง๋ ๋ํ ํ์คํ ๋ฆฌ๋ฅผ ๊ฒฐํฉํ์ฌ ๋ชจ๋ธ์ ๋ณด๋ผ ๋ฉ์์ง ๊ตฌ์ฑ
|
88 |
+
messages = [{"role": "system", "content": f"{system_prefix} {system_message}"}] + filtered_conversation
|
89 |
logging.debug(f'๋ชจ๋ธ์ ๋ณด๋ผ ๋ฉ์์ง: {messages}')
|
90 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
91 |
# ๋ชจ๋ธ ํธ์ถ
|
92 |
loop = asyncio.get_event_loop()
|
93 |
response = await loop.run_in_executor(None, lambda: hf_client.chat_completion(
|
94 |
+
messages=messages, max_tokens=1000, temperature=0.7, top_p=0.85))
|
95 |
|
96 |
full_response = []
|
97 |
for part in response:
|