geronimo-pericoli
commited on
Commit
•
aef499a
1
Parent(s):
bb63346
Update app.py
Browse files
app.py
CHANGED
@@ -16,6 +16,7 @@ import gradio as gr
|
|
16 |
from gradio import components
|
17 |
import textwrap
|
18 |
import datetime
|
|
|
19 |
|
20 |
openai.api_key = os.environ.get('openai_key')
|
21 |
os.environ["OPENAI_API_KEY"] = os.environ.get('openai_key')
|
@@ -57,17 +58,20 @@ def responder(pregunta):
|
|
57 |
if len(prompt_question.split()) > max_tokens:
|
58 |
prompt_question = ' '.join(prompt_question.split()[:max_tokens])
|
59 |
|
60 |
-
|
61 |
-
|
62 |
-
|
63 |
-
# Muestra la cadena que se va formando palabra por palabra
|
64 |
-
partial_message = ""
|
65 |
-
for chunk in respuesta.response_gen:
|
66 |
-
partial_message += chunk
|
67 |
-
yield partial_message
|
68 |
-
|
69 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
70 |
|
|
|
71 |
# Definir la interfaz de usuario con Gradio
|
72 |
with gr.Blocks(theme='sudeepshouche/minimalist') as demo:
|
73 |
|
|
|
16 |
from gradio import components
|
17 |
import textwrap
|
18 |
import datetime
|
19 |
+
from llama_index.core import Settings
|
20 |
|
21 |
openai.api_key = os.environ.get('openai_key')
|
22 |
os.environ["OPENAI_API_KEY"] = os.environ.get('openai_key')
|
|
|
58 |
if len(prompt_question.split()) > max_tokens:
|
59 |
prompt_question = ' '.join(prompt_question.split()[:max_tokens])
|
60 |
|
61 |
+
try:
|
62 |
+
# Respuesta con prompt sumado a la pregunta
|
63 |
+
respuesta = query_engine.query(prompt_question)
|
|
|
|
|
|
|
|
|
|
|
|
|
64 |
|
65 |
+
# Muestra la cadena que se va formando palabra por palabra
|
66 |
+
partial_message = ""
|
67 |
+
for chunk in respuesta.response_gen:
|
68 |
+
partial_message += chunk
|
69 |
+
yield partial_message
|
70 |
+
except Exception as e:
|
71 |
+
# Si hay algún error, devolver un mensaje de error
|
72 |
+
yield "Error, inténtelo nuevamente!"
|
73 |
|
74 |
+
|
75 |
# Definir la interfaz de usuario con Gradio
|
76 |
with gr.Blocks(theme='sudeepshouche/minimalist') as demo:
|
77 |
|