Update app.py
Browse files
app.py
CHANGED
@@ -60,10 +60,8 @@ def predict(message):
|
|
60 |
# Obtener la pregunta del usuario
|
61 |
user_message = full_prompt[-1]["content"]
|
62 |
|
63 |
-
# Utilizar el nuevo index Doc_1 para generar la respuesta
|
64 |
respuesta = engine.query(user_message)
|
65 |
|
66 |
-
# Inicializa la respuesta de OpenAI
|
67 |
response = openai.ChatCompletion.create(
|
68 |
model='gpt-3.5-turbo-16k', # Use the new model
|
69 |
max_tokens=1000,
|
@@ -71,16 +69,12 @@ def predict(message):
|
|
71 |
temperature=0.5,
|
72 |
stream=True
|
73 |
)
|
74 |
-
|
75 |
partial_message = ""
|
76 |
for chunk in respuesta:
|
77 |
if len(chunk['choices'][0]['delta']) != 0:
|
78 |
partial_message = partial_message + chunk['choices'][0]['delta']['content']
|
79 |
-
|
80 |
-
# Concatenar la respuesta de OpenAI con la respuesta del index
|
81 |
-
full_response = partial_message + '\n\n' + response.choices[0].message['content']
|
82 |
-
|
83 |
-
return full_response
|
84 |
|
85 |
# Define la interfaz de Gradio para el nuevo index Doc_1
|
86 |
gr.Interface(
|
|
|
60 |
# Obtener la pregunta del usuario
|
61 |
user_message = full_prompt[-1]["content"]
|
62 |
|
|
|
63 |
respuesta = engine.query(user_message)
|
64 |
|
|
|
65 |
response = openai.ChatCompletion.create(
|
66 |
model='gpt-3.5-turbo-16k', # Use the new model
|
67 |
max_tokens=1000,
|
|
|
69 |
temperature=0.5,
|
70 |
stream=True
|
71 |
)
|
72 |
+
|
73 |
partial_message = ""
|
74 |
for chunk in respuesta:
|
75 |
if len(chunk['choices'][0]['delta']) != 0:
|
76 |
partial_message = partial_message + chunk['choices'][0]['delta']['content']
|
77 |
+
yield partial_message
|
|
|
|
|
|
|
|
|
78 |
|
79 |
# Define la interfaz de Gradio para el nuevo index Doc_1
|
80 |
gr.Interface(
|