imperialwool commited on
Commit
6522af3
1 Parent(s): d8751a6

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +7 -6
app.py CHANGED
@@ -18,12 +18,13 @@ async def echo():
18
  userPrompt = prompt + "\n\nUser: " + data['request'] + "\nAssistant: "
19
  input_ids = tokenizer.encode(prompt, return_tensors='pt')
20
  with torch.no_grad():
21
- output_ids = model.generate(
22
- input_ids=input_ids,
23
- do_sample=random.choice([True, False]), temperature=float(random.randint(7,20)) / 10.0,
24
- max_new_tokens=data.get("max_tokens") or random.randomint(200,500),
25
- eos_token_id=tokenizer.eos_token_id, return_full_text = False)
26
- return {"output": tokenizer.decode(output_ids[0], skip_special_tokens=True)}
 
27
 
28
  @app.get("/")
29
  async def get():
 
18
  userPrompt = prompt + "\n\nUser: " + data['request'] + "\nAssistant: "
19
  input_ids = tokenizer.encode(prompt, return_tensors='pt')
20
  with torch.no_grad():
21
+ output_ids = model.generate(
22
+ input_ids=input_ids,
23
+ do_sample=random.choice([True, False]), temperature=float(random.randint(7,20)) / 10.0,
24
+ max_new_tokens=data.get("max_tokens") or random.randomint(200,500),
25
+ eos_token_id=tokenizer.eos_token_id, return_full_text = False)
26
+ output = tokenizer.decode(output_ids[0], skip_special_tokens=True)
27
+ return {"output": output}
28
 
29
  @app.get("/")
30
  async def get():