liyaoshi commited on
Commit
9bfbbf7
1 Parent(s): 3fa3ce5

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +13 -3
app.py CHANGED
@@ -46,6 +46,14 @@ def is_image(file_path):
46
  except IOError:
47
  return False
48
 
 
 
 
 
 
 
 
 
49
  # def respond(
50
  # message,
51
  # history: list[tuple[str, str]],
@@ -86,7 +94,7 @@ def get_completion(message,history):
86
  raise gr.Error("Please input a text query along the image(s).")
87
 
88
  text = message['text']
89
- content = [
90
  {"type": "text", "text": text},
91
  ]
92
  if message['files']:
@@ -98,7 +106,7 @@ def get_completion(message,history):
98
  "image_url": {
99
  "url": public_url,
100
  },}
101
- content.append(content_image)
102
  else:
103
  raise gr.Error("Only support image files now.")
104
 
@@ -109,7 +117,7 @@ def get_completion(message,history):
109
  continue
110
  history_openai_format.append({"role": "user", "content": human })
111
  history_openai_format.append({"role": "assistant", "content":assistant})
112
- history_openai_format.append({"role": "user", "content": content})
113
  # print(history_openai_format)
114
 
115
  system_message = '''You are GPT-4o("o" for omni), OpenAI's new flagship model that can reason across audio, vision, and text in real time.
@@ -150,6 +158,8 @@ def get_completion(message,history):
150
  else:
151
  response_content = response_data['choices'][0]['message']['content']
152
  usage = response_data['usage']
 
 
153
 
154
  return response_content
155
 
 
46
  except IOError:
47
  return False
48
 
49
+ def supbase_insert(user_message,response_content,messages,response):
50
+ from supabase import create_client, Client
51
+ url = os.environ.get('supabase_url')
52
+ key = os.environ.get('supabase_key')
53
+ supabase = create_client(url, key)
54
+ data, count = supabase.table('messages').insert({"user_message": user_message, "response_content": response_content,"messages":messages,"response":response}).execute()
55
+
56
+
57
  # def respond(
58
  # message,
59
  # history: list[tuple[str, str]],
 
94
  raise gr.Error("Please input a text query along the image(s).")
95
 
96
  text = message['text']
97
+ user_message = [
98
  {"type": "text", "text": text},
99
  ]
100
  if message['files']:
 
106
  "image_url": {
107
  "url": public_url,
108
  },}
109
+ user_message.append(content_image)
110
  else:
111
  raise gr.Error("Only support image files now.")
112
 
 
117
  continue
118
  history_openai_format.append({"role": "user", "content": human })
119
  history_openai_format.append({"role": "assistant", "content":assistant})
120
+ history_openai_format.append({"role": "user", "content": user_message})
121
  # print(history_openai_format)
122
 
123
  system_message = '''You are GPT-4o("o" for omni), OpenAI's new flagship model that can reason across audio, vision, and text in real time.
 
158
  else:
159
  response_content = response_data['choices'][0]['message']['content']
160
  usage = response_data['usage']
161
+
162
+ supbase_insert(user_message,response_content,messages,response_data)
163
 
164
  return response_content
165