Spaces:
Running
Running
ryanrwatkins
commited on
Commit
•
4f009dc
1
Parent(s):
660ac3a
Update app.py
Browse files
app.py
CHANGED
@@ -783,7 +783,7 @@ Standalone question:""")
|
|
783 |
|
784 |
|
785 |
|
786 |
-
def
|
787 |
|
788 |
|
789 |
history = state['messages']
|
@@ -793,38 +793,8 @@ def submit_message_OLD(prompt, prompt_template, temperature, max_tokens, context
|
|
793 |
|
794 |
prompt_template = prompt_templates[prompt_template]
|
795 |
|
796 |
-
with open("prompts_archive.csv", "a") as csvfile:
|
797 |
-
writer = csv.DictWriter(csvfile, fieldnames=["prompt", "time"])
|
798 |
-
writer.writerow(
|
799 |
-
{"prompt": str(prompt), "time": str(datetime.now())}
|
800 |
-
)
|
801 |
-
|
802 |
-
# with open(prompts_archive_file, "a") as csvfile:
|
803 |
-
# writer = csv.DictWriter(csvfile, fieldnames=["prompt", "time"])
|
804 |
-
# writer.writerow(
|
805 |
-
# {"prompt": str(prompt), "time": str(datetime.now())}
|
806 |
-
# )
|
807 |
-
# commit_url = repo.push_to_hub()
|
808 |
-
# print(commit_url)
|
809 |
-
|
810 |
-
system_prompt = []
|
811 |
-
if prompt_template:
|
812 |
-
system_prompt = [{ "role": "system", "content": prompt_template }]
|
813 |
-
|
814 |
-
prompt_msg = { "role": "user", "content": prompt }
|
815 |
|
816 |
-
|
817 |
-
#try:
|
818 |
-
|
819 |
-
with open("embeddings.pkl", 'rb') as f:
|
820 |
-
new_docsearch = pickle.load(f)
|
821 |
-
|
822 |
-
|
823 |
-
query = str(system_prompt + history + [prompt_msg])
|
824 |
-
|
825 |
-
docs = new_docsearch.similarity_search(query)
|
826 |
-
|
827 |
-
chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
|
828 |
#completion = chain.run(input_documents=docs, question=query)
|
829 |
|
830 |
|
@@ -835,20 +805,8 @@ def submit_message_OLD(prompt, prompt_template, temperature, max_tokens, context
|
|
835 |
|
836 |
completion = { "content": completion }
|
837 |
|
838 |
-
|
839 |
-
#state['total_tokens'] += completion['usage']['total_tokens']
|
840 |
-
|
841 |
|
842 |
-
|
843 |
-
#except Exception as e:
|
844 |
-
# history.append(prompt_msg.copy())
|
845 |
-
# error = {
|
846 |
-
# "role": "system",
|
847 |
-
# "content": f"Error: {e}"
|
848 |
-
# }
|
849 |
-
# history.append(error.copy())
|
850 |
|
851 |
-
#total_tokens_used_msg = f"Total tokens used: {state['total_tokens']}"
|
852 |
|
853 |
chat_messages = [(prompt_msg['content'], completion['content'])]
|
854 |
return '', chat_messages, state # total_tokens_used_msg,
|
|
|
783 |
|
784 |
|
785 |
|
786 |
+
def submit_message(prompt, prompt_template, temperature, max_tokens, context_length, state):
|
787 |
|
788 |
|
789 |
history = state['messages']
|
|
|
793 |
|
794 |
prompt_template = prompt_templates[prompt_template]
|
795 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
796 |
|
797 |
+
#chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
798 |
#completion = chain.run(input_documents=docs, question=query)
|
799 |
|
800 |
|
|
|
805 |
|
806 |
completion = { "content": completion }
|
807 |
|
|
|
|
|
|
|
808 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
809 |
|
|
|
810 |
|
811 |
chat_messages = [(prompt_msg['content'], completion['content'])]
|
812 |
return '', chat_messages, state # total_tokens_used_msg,
|