Felladrin commited on
Commit
cca23d7
·
1 Parent(s): a99204e

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -5
app.py CHANGED
@@ -23,15 +23,15 @@ def generate(
23
  prompt = pipe.tokenizer.apply_chat_template(message_template, tokenize=False, add_generation_prompt=True)
24
 
25
  if model_name == "Felladrin/Pythia-31M-Chat-v1":
26
- outputs = pipe(prompt, max_new_tokens=1024, do_sample=True, temperature=0.4, top_k=7, top_p=0.25, repetition_penalty=1.0016)
27
  elif model_name == "Felladrin/Llama-160M-Chat-v1":
28
- outputs = pipe(prompt, max_new_tokens=1024, penalty_alpha=0.5, top_k=5, repetition_penalty=1.01)
29
  elif model_name == "Felladrin/TinyMistral-248M-SFT-v4":
30
- outputs = pipe(prompt, max_new_tokens=1024, penalty_alpha=0.5, top_k=5, repetition_penalty=1.0)
31
  elif model_name == "Felladrin/Smol-Llama-101M-Chat-v1":
32
- outputs = pipe(prompt, max_new_tokens=1024, penalty_alpha=0.5, top_k=5, repetition_penalty=1.0, add_special_tokens=True)
33
  else:
34
- outputs = pipe(prompt, max_new_tokens=1024, do_sample=True, temperature=0.9, top_k=50, top_p=0.95, repetition_penalty=1.2)
35
 
36
  return outputs[0]["generated_text"]
37
 
 
23
  prompt = pipe.tokenizer.apply_chat_template(message_template, tokenize=False, add_generation_prompt=True)
24
 
25
  if model_name == "Felladrin/Pythia-31M-Chat-v1":
26
+ outputs = pipe(prompt, max_length=1024, use_cache=True, penalty_alpha=0.5, top_k=2, repetition_penalty=1.0016)
27
  elif model_name == "Felladrin/Llama-160M-Chat-v1":
28
+ outputs = pipe(prompt, max_length=1024, use_cache=True, penalty_alpha=0.5, top_k=5, repetition_penalty=1.01)
29
  elif model_name == "Felladrin/TinyMistral-248M-SFT-v4":
30
+ outputs = pipe(prompt, max_length=1024, use_cache=True, penalty_alpha=0.5, top_k=5, repetition_penalty=1.0)
31
  elif model_name == "Felladrin/Smol-Llama-101M-Chat-v1":
32
+ outputs = pipe(prompt, max_length=1024, use_cache=True, penalty_alpha=0.5, top_k=4, repetition_penalty=1.105)
33
  else:
34
+ outputs = pipe(prompt, max_length=1024, do_sample=True, temperature=0.9, top_k=50, top_p=0.95, repetition_penalty=1.2)
35
 
36
  return outputs[0]["generated_text"]
37