ysharma HF staff commited on
Commit
87e586e
1 Parent(s): 503973a

Removed default system prompt

Browse files
Files changed (1) hide show
  1. app.py +13 -10
app.py CHANGED
@@ -34,25 +34,29 @@ examples=[
34
  ]
35
 
36
 
 
 
 
 
 
37
  # Stream text
38
  def predict(message, chatbot, system_prompt="", temperature=0.9, max_new_tokens=256, top_p=0.6, repetition_penalty=1.0,):
39
 
40
  if system_prompt != "":
41
- system_message = system_prompt
42
  else:
43
- system_message = "\nYou are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information."
44
 
45
  temperature = float(temperature)
46
  if temperature < 1e-2:
47
  temperature = 1e-2
48
  top_p = float(top_p)
49
 
50
- input_prompt = f"[INST] <<SYS>>\n{system_message}\n<</SYS>>\n\n "
51
  for interaction in chatbot:
52
- input_prompt = input_prompt + str(interaction[0]) + " [/INST] " + str(interaction[1]) + " </s><s> [INST] "
53
 
54
  input_prompt = input_prompt + str(message) + " [/INST] "
55
-
56
  data = {
57
  "inputs": input_prompt,
58
  "parameters": {
@@ -101,18 +105,17 @@ def predict(message, chatbot, system_prompt="", temperature=0.9, max_new_tokens=
101
  def predict_batch(message, chatbot, system_prompt="", temperature=0.9, max_new_tokens=256, top_p=0.6, repetition_penalty=1.0,):
102
 
103
  if system_prompt != "":
104
- system_message = system_prompt
105
  else:
106
- system_message = "\nYou are a helpful, respectful and honest assistant. Always answer as helpfully as possible, while being safe. Your answers should not include any harmful, unethical, racist, sexist, toxic, dangerous, or illegal content. Please ensure that your responses are socially unbiased and positive in nature.\n\nIf a question does not make any sense, or is not factually coherent, explain why instead of answering something not correct. If you don't know the answer to a question, please don't share false information."
107
 
108
  temperature = float(temperature)
109
  if temperature < 1e-2:
110
  temperature = 1e-2
111
  top_p = float(top_p)
112
 
113
- input_prompt = f"[INST]<<SYS>>\n{system_message}\n<</SYS>>\n\n "
114
  for interaction in chatbot:
115
- input_prompt = input_prompt + str(interaction[0]) + " [/INST] " + str(interaction[1]) + " </s><s> [INST] "
116
 
117
  input_prompt = input_prompt + str(message) + " [/INST] "
118
 
@@ -225,4 +228,4 @@ with gr.Blocks() as demo:
225
  chatbot_batch.like(vote, None, None)
226
  chat_interface_batch.render()
227
 
228
- demo.queue(concurrency_count=75, max_size=100).launch(debug=True)
 
34
  ]
35
 
36
 
37
+ # Note: We have removed default system prompt as requested by the paper authors [Dated: 13/Oct/2023]
38
+ # Prompting style for Llama2 without using system prompt
39
+ # <s>[INST] {{ user_msg_1 }} [/INST] {{ model_answer_1 }} </s><s>[INST] {{ user_msg_2 }} [/INST]
40
+
41
+
42
  # Stream text
43
  def predict(message, chatbot, system_prompt="", temperature=0.9, max_new_tokens=256, top_p=0.6, repetition_penalty=1.0,):
44
 
45
  if system_prompt != "":
46
+ input_prompt = f"<s>[INST] <<SYS>>\n{system_prompt}\n<</SYS>>\n\n "
47
  else:
48
+ input_prompt = f"<s>[INST] "
49
 
50
  temperature = float(temperature)
51
  if temperature < 1e-2:
52
  temperature = 1e-2
53
  top_p = float(top_p)
54
 
 
55
  for interaction in chatbot:
56
+ input_prompt = input_prompt + str(interaction[0]) + " [/INST] " + str(interaction[1]) + " </s><s>[INST] "
57
 
58
  input_prompt = input_prompt + str(message) + " [/INST] "
59
+
60
  data = {
61
  "inputs": input_prompt,
62
  "parameters": {
 
105
  def predict_batch(message, chatbot, system_prompt="", temperature=0.9, max_new_tokens=256, top_p=0.6, repetition_penalty=1.0,):
106
 
107
  if system_prompt != "":
108
+ input_prompt = f"<s>[INST] <<SYS>>\n{system_prompt}\n<</SYS>>\n\n "
109
  else:
110
+ input_prompt = f"<s>[INST] "
111
 
112
  temperature = float(temperature)
113
  if temperature < 1e-2:
114
  temperature = 1e-2
115
  top_p = float(top_p)
116
 
 
117
  for interaction in chatbot:
118
+ input_prompt = input_prompt + str(interaction[0]) + " [/INST] " + str(interaction[1]) + " </s><s>[INST] "
119
 
120
  input_prompt = input_prompt + str(message) + " [/INST] "
121
 
 
228
  chatbot_batch.like(vote, None, None)
229
  chat_interface_batch.render()
230
 
231
+ demo.queue(concurrency_count=75, max_size=100).launch(debug=True)