Update main.py
Browse files
main.py
CHANGED
@@ -149,6 +149,10 @@ async def Search(input, categorie):
|
|
149 |
async def setup_conversationalChain():
|
150 |
model = await LLModel()
|
151 |
retriever = await Retriever(cl.user_session.get("selectRequest"))
|
|
|
|
|
|
|
|
|
152 |
qa = ConversationalRetrievalChain.from_llm(
|
153 |
model,
|
154 |
memory=cl.user_session.get("memory"),
|
@@ -218,9 +222,6 @@ async def on_chat_start():
|
|
218 |
await cl.Message(f"Vous pouvez requêter sur la thématique : {res.get('value')}").send()
|
219 |
cl.user_session.set("selectRequest", res.get("name"))
|
220 |
|
221 |
-
########## Chain with streaming ##########
|
222 |
-
message_history = ChatMessageHistory()
|
223 |
-
memory = ConversationBufferMemory(memory_key="chat_history",output_key="answer",chat_memory=message_history,return_messages=True)
|
224 |
await setup_conversationalChain()
|
225 |
|
226 |
@cl.on_chat_resume
|
|
|
149 |
async def setup_conversationalChain():
|
150 |
model = await LLModel()
|
151 |
retriever = await Retriever(cl.user_session.get("selectRequest"))
|
152 |
+
########## Chain with streaming ##########
|
153 |
+
message_history = ChatMessageHistory()
|
154 |
+
memory = ConversationBufferMemory(memory_key="chat_history",output_key="answer",chat_memory=message_history,return_messages=True)
|
155 |
+
|
156 |
qa = ConversationalRetrievalChain.from_llm(
|
157 |
model,
|
158 |
memory=cl.user_session.get("memory"),
|
|
|
222 |
await cl.Message(f"Vous pouvez requêter sur la thématique : {res.get('value')}").send()
|
223 |
cl.user_session.set("selectRequest", res.get("name"))
|
224 |
|
|
|
|
|
|
|
225 |
await setup_conversationalChain()
|
226 |
|
227 |
@cl.on_chat_resume
|