Spaces:
Runtime error
Runtime error
- gpt_based_function.py +6 -6
gpt_based_function.py
CHANGED
@@ -11,7 +11,7 @@ def gpt_summary_generator(user_text):
|
|
11 |
user_prompt = f"{user_text}"
|
12 |
messages = [{"role": "system", "content": task_description_fs}, {"role": "user", "content": user_prompt}]
|
13 |
response = openai.ChatCompletion.create(
|
14 |
-
model="gpt-3.5-turbo-
|
15 |
messages=messages,
|
16 |
temperature=0.06,
|
17 |
max_tokens=8362,
|
@@ -34,7 +34,7 @@ def gpt_keyword_highlighter(user_text):
|
|
34 |
user_prompt = r"{input_text}=" + f"{user_text}"
|
35 |
messages = [{"role": "system", "content": task_description}, {"role": "user", "content": user_prompt}]
|
36 |
response = openai.ChatCompletion.create(
|
37 |
-
model="gpt-3.5-turbo",
|
38 |
messages=messages,
|
39 |
temperature=0,
|
40 |
max_tokens=2006,
|
@@ -60,7 +60,7 @@ def gpt_text_naturalizer(user_input):
|
|
60 |
]
|
61 |
|
62 |
response = openai.ChatCompletion.create(
|
63 |
-
model="gpt-3.5-turbo-
|
64 |
messages=messages,
|
65 |
temperature=0.1,
|
66 |
max_tokens=2500
|
@@ -80,7 +80,7 @@ def gpt_explanation_generator(user_input, text_to_consider):
|
|
80 |
]
|
81 |
|
82 |
response = openai.ChatCompletion.create(
|
83 |
-
model="gpt-3.5-turbo-
|
84 |
messages=messages,
|
85 |
temperature=0.1,
|
86 |
max_tokens=200
|
@@ -98,7 +98,7 @@ def gpt_easier_text_generator(user_input):
|
|
98 |
]
|
99 |
|
100 |
response = openai.ChatCompletion.create(
|
101 |
-
model="gpt-3.5-turbo",
|
102 |
messages=messages,
|
103 |
temperature=0,
|
104 |
max_tokens=2048
|
@@ -119,7 +119,7 @@ def gpt_rater(user_input):
|
|
119 |
]
|
120 |
|
121 |
response = openai.ChatCompletion.create(
|
122 |
-
model="gpt-3.5-turbo-
|
123 |
messages=messages,
|
124 |
temperature=0.1,
|
125 |
max_tokens=2500
|
|
|
11 |
user_prompt = f"{user_text}"
|
12 |
messages = [{"role": "system", "content": task_description_fs}, {"role": "user", "content": user_prompt}]
|
13 |
response = openai.ChatCompletion.create(
|
14 |
+
model="gpt-3.5-turbo-0125",
|
15 |
messages=messages,
|
16 |
temperature=0.06,
|
17 |
max_tokens=8362,
|
|
|
34 |
user_prompt = r"{input_text}=" + f"{user_text}"
|
35 |
messages = [{"role": "system", "content": task_description}, {"role": "user", "content": user_prompt}]
|
36 |
response = openai.ChatCompletion.create(
|
37 |
+
model="gpt-3.5-turbo-0125",
|
38 |
messages=messages,
|
39 |
temperature=0,
|
40 |
max_tokens=2006,
|
|
|
60 |
]
|
61 |
|
62 |
response = openai.ChatCompletion.create(
|
63 |
+
model="gpt-3.5-turbo-0125",
|
64 |
messages=messages,
|
65 |
temperature=0.1,
|
66 |
max_tokens=2500
|
|
|
80 |
]
|
81 |
|
82 |
response = openai.ChatCompletion.create(
|
83 |
+
model="gpt-3.5-turbo-0125",
|
84 |
messages=messages,
|
85 |
temperature=0.1,
|
86 |
max_tokens=200
|
|
|
98 |
]
|
99 |
|
100 |
response = openai.ChatCompletion.create(
|
101 |
+
model="gpt-3.5-turbo-0125",
|
102 |
messages=messages,
|
103 |
temperature=0,
|
104 |
max_tokens=2048
|
|
|
119 |
]
|
120 |
|
121 |
response = openai.ChatCompletion.create(
|
122 |
+
model="gpt-3.5-turbo-0125",
|
123 |
messages=messages,
|
124 |
temperature=0.1,
|
125 |
max_tokens=2500
|