ryanrwatkins commited on
Commit
4f009dc
1 Parent(s): 660ac3a

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -44
app.py CHANGED
@@ -783,7 +783,7 @@ Standalone question:""")
783
 
784
 
785
 
786
- def submit_message_OLD(prompt, prompt_template, temperature, max_tokens, context_length, state):
787
 
788
 
789
  history = state['messages']
@@ -793,38 +793,8 @@ def submit_message_OLD(prompt, prompt_template, temperature, max_tokens, context
793
 
794
  prompt_template = prompt_templates[prompt_template]
795
 
796
- with open("prompts_archive.csv", "a") as csvfile:
797
- writer = csv.DictWriter(csvfile, fieldnames=["prompt", "time"])
798
- writer.writerow(
799
- {"prompt": str(prompt), "time": str(datetime.now())}
800
- )
801
-
802
- # with open(prompts_archive_file, "a") as csvfile:
803
- # writer = csv.DictWriter(csvfile, fieldnames=["prompt", "time"])
804
- # writer.writerow(
805
- # {"prompt": str(prompt), "time": str(datetime.now())}
806
- # )
807
- # commit_url = repo.push_to_hub()
808
- # print(commit_url)
809
-
810
- system_prompt = []
811
- if prompt_template:
812
- system_prompt = [{ "role": "system", "content": prompt_template }]
813
-
814
- prompt_msg = { "role": "user", "content": prompt }
815
 
816
-
817
- #try:
818
-
819
- with open("embeddings.pkl", 'rb') as f:
820
- new_docsearch = pickle.load(f)
821
-
822
-
823
- query = str(system_prompt + history + [prompt_msg])
824
-
825
- docs = new_docsearch.similarity_search(query)
826
-
827
- chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
828
  #completion = chain.run(input_documents=docs, question=query)
829
 
830
 
@@ -835,20 +805,8 @@ def submit_message_OLD(prompt, prompt_template, temperature, max_tokens, context
835
 
836
  completion = { "content": completion }
837
 
838
-
839
- #state['total_tokens'] += completion['usage']['total_tokens']
840
-
841
 
842
-
843
- #except Exception as e:
844
- # history.append(prompt_msg.copy())
845
- # error = {
846
- # "role": "system",
847
- # "content": f"Error: {e}"
848
- # }
849
- # history.append(error.copy())
850
 
851
- #total_tokens_used_msg = f"Total tokens used: {state['total_tokens']}"
852
 
853
  chat_messages = [(prompt_msg['content'], completion['content'])]
854
  return '', chat_messages, state # total_tokens_used_msg,
 
783
 
784
 
785
 
786
+ def submit_message(prompt, prompt_template, temperature, max_tokens, context_length, state):
787
 
788
 
789
  history = state['messages']
 
793
 
794
  prompt_template = prompt_templates[prompt_template]
795
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
796
 
797
+ #chain = load_qa_chain(ChatOpenAI(temperature=temperature, max_tokens=max_tokens, model_name="gpt-3.5-turbo"), chain_type="stuff")
 
 
 
 
 
 
 
 
 
 
 
798
  #completion = chain.run(input_documents=docs, question=query)
799
 
800
 
 
805
 
806
  completion = { "content": completion }
807
 
 
 
 
808
 
 
 
 
 
 
 
 
 
809
 
 
810
 
811
  chat_messages = [(prompt_msg['content'], completion['content'])]
812
  return '', chat_messages, state # total_tokens_used_msg,