Tonic commited on
Commit
5f4522c
·
verified ·
1 Parent(s): db52422

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +2 -3
app.py CHANGED
@@ -18,14 +18,13 @@ quantization_config = BitsAndBytesConfig(load_in_8bit=True)
18
  model = AutoModelForCausalLM.from_pretrained(model_path, device_map='cuda', quantization_config=quantization_config)
19
 
20
  @spaces.GPU
21
- def generate_text(usertitle, content, max_new_tokens=512, temperature=0.7, model=model, tokenizer=tokenizer):
22
  # msg = [{
23
  # 'title': title,
24
  # 'content': content
25
  # }]
26
  input_text = f"[[[Title:]]]{usertitle}\n[[[Content:]]]{content}\n"
27
- inputs = tokenizer(input_text, return_tensors='pt').to('cuda')
28
-
29
  generated_ids = model.generate(inputs['input_ids'], max_new_tokens=max_length, temperature=temperature, pad_token_id=tokenizer.eos_token_id)
30
  generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=False)
31
 
 
18
  model = AutoModelForCausalLM.from_pretrained(model_path, device_map='cuda', quantization_config=quantization_config)
19
 
20
  @spaces.GPU
21
+ def generate_text(usertitle, content, max_new_tokens=512, temperature=0.7):
22
  # msg = [{
23
  # 'title': title,
24
  # 'content': content
25
  # }]
26
  input_text = f"[[[Title:]]]{usertitle}\n[[[Content:]]]{content}\n"
27
+ inputs = tokenizer(input_text, return_tensors='pt').to('cuda')
 
28
  generated_ids = model.generate(inputs['input_ids'], max_new_tokens=max_length, temperature=temperature, pad_token_id=tokenizer.eos_token_id)
29
  generated_text = tokenizer.decode(generated_ids[0], skip_special_tokens=False)
30