jason9693 commited on
Commit
84c1f5e
1 Parent(s): dde792a

modified_config

Browse files
Files changed (1) hide show
  1. app.py +6 -3
app.py CHANGED
@@ -15,9 +15,9 @@ examples = [
15
  ["2040년 미국은, "]
16
  ]
17
 
18
- tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-ko-2.7b", use_auth_token=os.environ['TOKEN'])
19
  model = AutoModelForCausalLM.from_pretrained(
20
- "EleutherAI/gpt-neox-ko-2.7b", use_auth_token=os.environ['TOKEN'],
21
  pad_token_id=tokenizer.pad_token_id,
22
  eos_token_id=tokenizer.eos_token_id,
23
  bos_token_id=tokenizer.bos_token_id
@@ -30,7 +30,10 @@ def predict(text):
30
  stopping_cond = StoppingCriteriaList([tokenizer.encode('<|endoftext|>')])
31
  with torch.no_grad():
32
  tokens = tokenizer(text, return_tensors="pt").input_ids
33
- gen_tokens = model.generate(tokens, do_sample=True, temperature=0.8, max_new_tokens=64)
 
 
 
34
  generated = tokenizer.batch_decode(gen_tokens)[0]
35
  return generated
36
  # return pipe(text)[0]['generated_text']
 
15
  ["2040년 미국은, "]
16
  ]
17
 
18
+ tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-ko-1.3b", use_auth_token=os.environ['TOKEN'])
19
  model = AutoModelForCausalLM.from_pretrained(
20
+ "EleutherAI/gpt-neox-ko-1.3b", use_auth_token=os.environ['TOKEN'],
21
  pad_token_id=tokenizer.pad_token_id,
22
  eos_token_id=tokenizer.eos_token_id,
23
  bos_token_id=tokenizer.bos_token_id
 
30
  stopping_cond = StoppingCriteriaList([tokenizer.encode('<|endoftext|>')])
31
  with torch.no_grad():
32
  tokens = tokenizer(text, return_tensors="pt").input_ids
33
+ gen_tokens = model.generate(
34
+ tokens, do_sample=True, temperature=0.8, max_new_tokens=64, top_k=50, top_p=0.8,
35
+ no_repeat_ngram_size=3, repetition_penalty=1.2
36
+ )
37
  generated = tokenizer.batch_decode(gen_tokens)[0]
38
  return generated
39
  # return pipe(text)[0]['generated_text']