DeepMount00 commited on
Commit
c8016b4
1 Parent(s): 42f7935

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -2
README.md CHANGED
@@ -43,11 +43,11 @@ model = transformers.LlamaForCausalLM.from_pretrained(model_name, torch_dtype=to
43
  def stream(user_prompt):
44
  runtimeFlag = "cuda:0"
45
  system_prompt = ''
46
- B_INST, E_INST = "[INST]", "[/INST]"
47
  prompt = f"{system_prompt}{B_INST}{user_prompt.strip()}\n{E_INST}"
48
  inputs = tokenizer([prompt], return_tensors="pt").to(runtimeFlag)
49
  streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
50
- _ = model.generate(**inputs, streamer=streamer, max_new_tokens=100, num_return_sequences=1)
51
 
52
  domanda = """Quale è la media tra questi valori?"""
53
  contesto = """
 
43
  def stream(user_prompt):
44
  runtimeFlag = "cuda:0"
45
  system_prompt = ''
46
+ B_INST, E_INST = "<s> [INST]", "[/INST]"
47
  prompt = f"{system_prompt}{B_INST}{user_prompt.strip()}\n{E_INST}"
48
  inputs = tokenizer([prompt], return_tensors="pt").to(runtimeFlag)
49
  streamer = TextStreamer(tokenizer, skip_prompt=True, skip_special_tokens=True)
50
+ _ = model.generate(**inputs, streamer=streamer, max_new_tokens=200, temperature=0.0001, eos_token_id=2, num_return_sequences=1)
51
 
52
  domanda = """Quale è la media tra questi valori?"""
53
  contesto = """