Yash Sachdeva commited on
Commit
72231f4
·
1 Parent(s): 18dd69a
Files changed (1) hide show
  1. question_paper.py +8 -11
question_paper.py CHANGED
@@ -12,15 +12,12 @@ TOKENIZER = None
12
 
13
  @app.get("/")
14
  def llama():
15
- text = "Hi, my name is "
16
- inputs = TOKENIZER(text, return_tensors="pt").input_ids
17
- outputs = MODEL.generate(
18
- inputs,
19
- max_length=256,
20
- pad_token_id=TOKENIZER.pad_token_id,
21
- eos_token_id=TOKENIZER.eos_token_id,
22
- )
23
- tresponse = TOKENIZER.decode(outputs[0], skip_special_tokens=True)
24
  print(tresponse)
25
 
26
  return tresponse
@@ -31,6 +28,6 @@ def init_model():
31
  global TOKENIZER
32
  if not MODEL:
33
  print("loading model")
34
- TOKENIZER = AutoTokenizer.from_pretrained("berkeley-nest/Starling-LM-7B-alpha")
35
- MODEL = AutoModelForCausalLM.from_pretrained("berkeley-nest/Starling-LM-7B-alpha")
36
  print("loaded model")
 
12
 
13
  @app.get("/")
14
  def llama():
15
+ prompt = [{'role': 'user', 'content': 'List 3 synonyms for the word "tiny"'}]
16
+ inputs = TOKENIZER.apply_chat_template( prompt, add_generation_prompt=True, return_tensors='pt' )
17
+
18
+ tokens = MODEL.generate( inputs.to(MODEL.device), max_new_tokens=1024, temperature=0.3, do_sample=True)
19
+
20
+ tresponse = TOKENIZER.decode(tokens[0], skip_special_tokens=False)
 
 
 
21
  print(tresponse)
22
 
23
  return tresponse
 
28
  global TOKENIZER
29
  if not MODEL:
30
  print("loading model")
31
+ TOKENIZER = AutoTokenizer.from_pretrained('stabilityai/stablelm-zephyr-3b')
32
+ MODEL = AutoModelForCausalLM.from_pretrained('stabilityai/stablelm-zephyr-3b', device_map="auto")
33
  print("loaded model")