KarthickAdopleAI commited on
Commit
4fed321
1 Parent(s): ae9b65c

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -6
app.py CHANGED
@@ -67,7 +67,7 @@ class Resume_Overall:
67
  # Generate a response from the GPT-3 model
68
  chat_completion = self.client.chat.completions.create(
69
  model = "ChatGPT",
70
- prompt=prompt,
71
  max_tokens=200,
72
  temperature=0,
73
  n=1,
@@ -94,7 +94,7 @@ class Resume_Overall:
94
  # Generate a response from the GPT-3 model
95
  chat_completion = self.client.chat.completions.create(
96
  model = "ChatGPT",
97
- prompt=prompt,
98
  max_tokens=200,
99
  temperature=0,
100
  n=1,
@@ -123,7 +123,7 @@ class Resume_Overall:
123
  # Generate a response from the GPT-3 model
124
  chat_completion = self.client.chat.completions.create(
125
  model = "ChatGPT", # Choose the GPT-3 engine you want to use
126
- prompt=prompt,
127
  max_tokens=100, # Set the maximum number of tokens in the generated response
128
  temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
129
  n=1, # Generate a single response
@@ -143,7 +143,7 @@ class Resume_Overall:
143
 
144
  chat_completion = self.client.chat.completions.create(
145
  model = "ChatGPT",
146
- prompt=prompt,
147
  max_tokens=100,
148
  temperature=0,
149
  n=1,
@@ -176,7 +176,7 @@ class Resume_Overall:
176
 
177
  chat_completion = self.client.chat.completions.create(
178
  model = "ChatGPT", # Choose the GPT-3 engine you want to use
179
- prompt=prompt,
180
  max_tokens=500, # Set the maximum number of tokens in the generated response
181
  temperature=0.5, # Controls the randomness of the output. Higher values = more random, lower values = more focused
182
  )
@@ -203,7 +203,7 @@ class Resume_Overall:
203
  # Generate a response from the GPT-3 model
204
  chat_completion = self.client.chat.completions.create(
205
  model = "ChatGPT", # Choose the GPT-3 engine you want to use
206
- prompt=prompt,
207
  max_tokens=200, # Set the maximum number of tokens in the generated response
208
  temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
209
  n=1, # Generate a single response
 
67
  # Generate a response from the GPT-3 model
68
  chat_completion = self.client.chat.completions.create(
69
  model = "ChatGPT",
70
+ messages = conversation,
71
  max_tokens=200,
72
  temperature=0,
73
  n=1,
 
94
  # Generate a response from the GPT-3 model
95
  chat_completion = self.client.chat.completions.create(
96
  model = "ChatGPT",
97
+ messages = conversation,
98
  max_tokens=200,
99
  temperature=0,
100
  n=1,
 
123
  # Generate a response from the GPT-3 model
124
  chat_completion = self.client.chat.completions.create(
125
  model = "ChatGPT", # Choose the GPT-3 engine you want to use
126
+ messages = conversation,
127
  max_tokens=100, # Set the maximum number of tokens in the generated response
128
  temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
129
  n=1, # Generate a single response
 
143
 
144
  chat_completion = self.client.chat.completions.create(
145
  model = "ChatGPT",
146
+ messages = conversation,
147
  max_tokens=100,
148
  temperature=0,
149
  n=1,
 
176
 
177
  chat_completion = self.client.chat.completions.create(
178
  model = "ChatGPT", # Choose the GPT-3 engine you want to use
179
+ messages = conversation,
180
  max_tokens=500, # Set the maximum number of tokens in the generated response
181
  temperature=0.5, # Controls the randomness of the output. Higher values = more random, lower values = more focused
182
  )
 
203
  # Generate a response from the GPT-3 model
204
  chat_completion = self.client.chat.completions.create(
205
  model = "ChatGPT", # Choose the GPT-3 engine you want to use
206
+ messages = conversation,
207
  max_tokens=200, # Set the maximum number of tokens in the generated response
208
  temperature=0, # Controls the randomness of the output. Higher values = more random, lower values = more focused
209
  n=1, # Generate a single response