jason9693 commited on
Commit
621210f
1 Parent(s): e3fd9e9

modified some arguments

Browse files
Files changed (1) hide show
  1. app.py +19 -3
app.py CHANGED
@@ -1,6 +1,7 @@
1
  import gradio as gr
2
- from transformers import pipeline
3
  import os
 
4
 
5
  theme = "darkgrass"
6
  title = "GPT-NeoX(Korean) Demo"
@@ -14,10 +15,25 @@ examples = [
14
  ["2040년 미국은, "]
15
  ]
16
 
17
- pipe = pipeline('text-generation', model=model_name, use_auth_token=os.environ['TOKEN'])
 
 
 
 
 
 
 
 
 
18
 
19
  def predict(text):
20
- return pipe(text, do_sample=True, return_full_text=False, max_length=100)[0]['generated_text']
 
 
 
 
 
 
21
 
22
  iface = gr.Interface(
23
  fn=predict,
 
1
  import gradio as gr
2
+ from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM, StoppingCriteriaList
3
  import os
4
+ import torch
5
 
6
  theme = "darkgrass"
7
  title = "GPT-NeoX(Korean) Demo"
 
15
  ["2040년 미국은, "]
16
  ]
17
 
18
+ tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-ko-1.3b", use_auth_token=os.environ['TOKEN'])
19
+ model = AutoModelForCausalLM.from_pretrained(
20
+ "EleutherAI/gpt-neox-ko-1.3b", use_auth_token=os.environ['TOKEN'],
21
+ pad_token_id=tokenizer.pad_token_id,
22
+ eos_token_id=tokenizer.eos_token_id,
23
+ bos_token_id=tokenizer.bos_token_id
24
+ )
25
+ model.eval()
26
+
27
+ pipe = pipeline('text-generation', model=model, tokenizer=tokenizer)
28
 
29
  def predict(text):
30
+ stopping_cond = StoppingCriteriaList([tokenizer.encode('<|endoftext|>')])
31
+ with torch.no_grad():
32
+ tokens = tokenizer(text, return_tensors="pt").input_ids
33
+ gen_tokens = model.generate(tokens, do_sample=True, temperature=0.8, max_new_tokens=64)
34
+ generated = tokenizer.batch_decode(gen_tokens)[0]
35
+ return generated
36
+ # return pipe(text)[0]['generated_text']
37
 
38
  iface = gr.Interface(
39
  fn=predict,