Zhengyi commited on
Commit
42ce289
·
verified ·
1 Parent(s): 1a20162

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -2
app.py CHANGED
@@ -138,8 +138,9 @@ def chat_llama3_8b(message: str,
138
  input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
139
 
140
  streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
141
- print(max_new_tokens)
142
  max_new_tokens=4096
 
143
  generate_kwargs = dict(
144
  input_ids= input_ids,
145
  streamer=streamer,
@@ -180,14 +181,16 @@ with gr.Blocks(fill_height=True, css=css) as demo:
180
  gr.Slider(minimum=0,
181
  maximum=1,
182
  step=0.1,
183
- value=0.95,
184
  label="Temperature",
 
185
  render=False),
186
  gr.Slider(minimum=128,
187
  maximum=4096,
188
  step=1,
189
  value=4096,
190
  label="Max new tokens",
 
191
  render=False),
192
  ],
193
  examples=[
 
138
  input_ids = tokenizer.apply_chat_template(conversation, return_tensors="pt").to(model.device)
139
 
140
  streamer = TextIteratorStreamer(tokenizer, timeout=10.0, skip_prompt=True, skip_special_tokens=True)
141
+ # print(max_new_tokens)
142
  max_new_tokens=4096
143
+ temperature=0.9
144
  generate_kwargs = dict(
145
  input_ids= input_ids,
146
  streamer=streamer,
 
181
  gr.Slider(minimum=0,
182
  maximum=1,
183
  step=0.1,
184
+ value=0.9,
185
  label="Temperature",
186
+ interactive = False,
187
  render=False),
188
  gr.Slider(minimum=128,
189
  maximum=4096,
190
  step=1,
191
  value=4096,
192
  label="Max new tokens",
193
+ interactive = False,
194
  render=False),
195
  ],
196
  examples=[