Update app.py
Browse files
app.py
CHANGED
@@ -141,7 +141,7 @@ def add_text(history, prompt, file):
|
|
141 |
else:
|
142 |
history = history + [((file.name,), None), (prompt, None)]
|
143 |
|
144 |
-
return history, prompt, "", gr.File( label=None, interactive=False, height=20, min_width=20, visible=False, scale=2) #gr.Textbox(value="", interactive=False)
|
145 |
|
146 |
def add_file(history, file, prompt):
|
147 |
if (prompt == ""):
|
@@ -323,6 +323,9 @@ def generate_prompt_with_history_openai(prompt, history):
|
|
323 |
history_openai_format.append({"role": "assistant", "content":assistant})
|
324 |
|
325 |
history_openai_format.append({"role": "user", "content": prompt})
|
|
|
|
|
|
|
326 |
return history_openai_format
|
327 |
|
328 |
#Prompt und History für Hugging Face Schnittstelle
|
@@ -380,7 +383,7 @@ def umwandeln_fuer_anzeige(image):
|
|
380 |
image.save(buffer, format='PNG')
|
381 |
return buffer.getvalue()
|
382 |
|
383 |
-
def generate_auswahl(prompt, file, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,):
|
384 |
if (prompt.find('zeichnen') != -1):
|
385 |
response = generate_bild(prompt)
|
386 |
result = response.content
|
@@ -394,10 +397,10 @@ def generate_auswahl(prompt, file, history, rag_option, model_option, openai_api
|
|
394 |
else:
|
395 |
result = generate_text(prompt, file, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,)
|
396 |
#Antwort als Stream ausgeben... wenn Textantwort gefordert
|
397 |
-
print("history vor Zusatz...........")
|
398 |
-
print(history)
|
399 |
history[-1][1] = result
|
400 |
-
|
|
|
|
|
401 |
"""
|
402 |
for character in result:
|
403 |
history[-1][1] += character
|
@@ -417,7 +420,7 @@ def generate_bild(prompt):
|
|
417 |
print("fertig Bild")
|
418 |
return response
|
419 |
|
420 |
-
def generate_text (prompt, file, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,):
|
421 |
global splittet
|
422 |
print(splittet)
|
423 |
|
@@ -586,7 +589,7 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
|
|
586 |
user_question,
|
587 |
upload,
|
588 |
chatbot,
|
589 |
-
|
590 |
rag_option,
|
591 |
model_option,
|
592 |
openai_key,
|
@@ -597,7 +600,7 @@ with gr.Blocks(css=customCSS, theme=small_and_beautiful_theme) as demo:
|
|
597 |
max_context_length_tokens,
|
598 |
repetition_penalty
|
599 |
],
|
600 |
-
outputs=[chatbot, status_display], #[chatbot, history, status_display]
|
601 |
show_progress=True,
|
602 |
postprocess=False
|
603 |
)
|
|
|
141 |
else:
|
142 |
history = history + [((file.name,), None), (prompt, None)]
|
143 |
|
144 |
+
return history, history, prompt, "", gr.File( label=None, interactive=False, height=20, min_width=20, visible=False, scale=2) #gr.Textbox(value="", interactive=False)
|
145 |
|
146 |
def add_file(history, file, prompt):
|
147 |
if (prompt == ""):
|
|
|
323 |
history_openai_format.append({"role": "assistant", "content":assistant})
|
324 |
|
325 |
history_openai_format.append({"role": "user", "content": prompt})
|
326 |
+
print("openai history und prompt................")
|
327 |
+
print(history_openai_format)
|
328 |
+
print (history_openai_format)
|
329 |
return history_openai_format
|
330 |
|
331 |
#Prompt und History für Hugging Face Schnittstelle
|
|
|
383 |
image.save(buffer, format='PNG')
|
384 |
return buffer.getvalue()
|
385 |
|
386 |
+
def generate_auswahl(prompt, file, chatbot, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,):
|
387 |
if (prompt.find('zeichnen') != -1):
|
388 |
response = generate_bild(prompt)
|
389 |
result = response.content
|
|
|
397 |
else:
|
398 |
result = generate_text(prompt, file, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,)
|
399 |
#Antwort als Stream ausgeben... wenn Textantwort gefordert
|
|
|
|
|
400 |
history[-1][1] = result
|
401 |
+
print("history nach Zusatz und mit KI Antwort...........")
|
402 |
+
print(history)
|
403 |
+
return history, history, "Success"
|
404 |
"""
|
405 |
for character in result:
|
406 |
history[-1][1] += character
|
|
|
420 |
print("fertig Bild")
|
421 |
return response
|
422 |
|
423 |
+
def generate_text (prompt, file, chatbot, history, rag_option, model_option, openai_api_key, k=3, top_p=0.6, temperature=0.5, max_new_tokens=4048, max_context_length_tokens=2048, repetition_penalty=1.3,):
|
424 |
global splittet
|
425 |
print(splittet)
|
426 |
|
|
|
589 |
user_question,
|
590 |
upload,
|
591 |
chatbot,
|
592 |
+
history,
|
593 |
rag_option,
|
594 |
model_option,
|
595 |
openai_key,
|
|
|
600 |
max_context_length_tokens,
|
601 |
repetition_penalty
|
602 |
],
|
603 |
+
outputs=[chatbot, history, status_display], #[chatbot, history, status_display]
|
604 |
show_progress=True,
|
605 |
postprocess=False
|
606 |
)
|