NealCaren commited on
Commit
183108b
1 Parent(s): 1d1d706
Files changed (1) hide show
  1. app.py +15 -7
app.py CHANGED
@@ -4,7 +4,7 @@ from scipy import spatial
4
  from collections import defaultdict
5
 
6
  import tiktoken
7
- import openai
8
  import gradio as gr
9
  from tenacity import retry, stop_after_attempt, wait_random_exponential
10
 
@@ -20,12 +20,14 @@ def ask_naive(query):
20
  {"role": "system", "content": "You are a college sociology professor. Provide a very brief answer to this student question."},
21
  {"role": "user", "content": query},
22
  ]
23
- response = openai.ChatCompletion.create(
 
 
24
  model='gpt-3.5-turbo',
25
  messages=messages,
26
  )
27
 
28
- response_message = response["choices"][0]["message"]["content"]
29
  return response_message
30
 
31
  # search function
@@ -38,11 +40,15 @@ def strings_ranked_by_relatedness(
38
  top_n: int = 100
39
  ) -> tuple[list[str], list[float]]:
40
  """Returns a list of strings and relatednesses, sorted from most related to least."""
41
- query_embedding_response = openai.Embedding.create(
 
 
 
 
42
  model=EMBEDDING_MODEL,
43
  input=query,
44
  )
45
- query_embedding = query_embedding_response["data"][0]["embedding"]
46
  strings_and_relatednesses = [
47
  (row["text"], relatedness_fn(query_embedding, row["embedding"]))
48
  for i, row in df.iterrows()
@@ -81,11 +87,13 @@ def respond(question, textbook_samples):
81
  """ }
82
  ]
83
 
84
- response = openai.ChatCompletion.create(
 
 
85
  model='gpt-3.5-turbo',
86
  n=1,
87
  messages=messages)
88
- return response["choices"][0]["message"]["content"]
89
 
90
  def ask(query):
91
  psuedo_answer = ask_naive(query)
 
4
  from collections import defaultdict
5
 
6
  import tiktoken
7
+ from openai import OpenAI
8
  import gradio as gr
9
  from tenacity import retry, stop_after_attempt, wait_random_exponential
10
 
 
20
  {"role": "system", "content": "You are a college sociology professor. Provide a very brief answer to this student question."},
21
  {"role": "user", "content": query},
22
  ]
23
+
24
+ client = OpenAI()
25
+ response = client.chat.completions.create(
26
  model='gpt-3.5-turbo',
27
  messages=messages,
28
  )
29
 
30
+ response_message = response.choices[0].message.content
31
  return response_message
32
 
33
  # search function
 
40
  top_n: int = 100
41
  ) -> tuple[list[str], list[float]]:
42
  """Returns a list of strings and relatednesses, sorted from most related to least."""
43
+
44
+ client = OpenAI()
45
+
46
+
47
+ query_embedding_response = client.embeddings.create(
48
  model=EMBEDDING_MODEL,
49
  input=query,
50
  )
51
+ query_embedding = query_embedding_response.data[0].embedding
52
  strings_and_relatednesses = [
53
  (row["text"], relatedness_fn(query_embedding, row["embedding"]))
54
  for i, row in df.iterrows()
 
87
  """ }
88
  ]
89
 
90
+ client = OpenAI()
91
+
92
+ response = client.chat.completions.create(
93
  model='gpt-3.5-turbo',
94
  n=1,
95
  messages=messages)
96
+ return response.choices[0].message.content
97
 
98
  def ask(query):
99
  psuedo_answer = ask_naive(query)