Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
|
@@ -13,9 +13,9 @@ def chatbot(message, history):
|
|
| 13 |
try:
|
| 14 |
prompt = "User: " + message + "\nAssistant: "
|
| 15 |
response = client.post(
|
| 16 |
-
prompt, #
|
| 17 |
-
model=MODEL_NAME,
|
| 18 |
-
{"max_new_tokens": 128, "temperature": 0.7, "do_sample": True} #
|
| 19 |
)
|
| 20 |
return response.strip()
|
| 21 |
except Exception as e:
|
|
@@ -35,5 +35,6 @@ with gr.Blocks(theme=gr.themes.Soft()) as iface:
|
|
| 35 |
|
| 36 |
gr.Markdown("---")
|
| 37 |
gr.Markdown(f"Modelo utilizado: `{MODEL_NAME}`")
|
|
|
|
| 38 |
|
| 39 |
iface.launch()
|
|
|
|
| 13 |
try:
|
| 14 |
prompt = "User: " + message + "\nAssistant: "
|
| 15 |
response = client.post(
|
| 16 |
+
prompt, # Argumento posicional: el prompt
|
| 17 |
+
model=MODEL_NAME, # Argumento de palabra clave
|
| 18 |
+
parameters={"max_new_tokens": 128, "temperature": 0.7, "do_sample": True} # Argumento de palabra clave
|
| 19 |
)
|
| 20 |
return response.strip()
|
| 21 |
except Exception as e:
|
|
|
|
| 35 |
|
| 36 |
gr.Markdown("---")
|
| 37 |
gr.Markdown(f"Modelo utilizado: `{MODEL_NAME}`")
|
| 38 |
+
gr.Markdown("Desarrollado con Gradio y Hugging Face Inference API")
|
| 39 |
|
| 40 |
iface.launch()
|