Spaces:
Build error
Build error
modified some arguments
Browse files
app.py
CHANGED
@@ -1,6 +1,7 @@
|
|
1 |
import gradio as gr
|
2 |
-
from transformers import pipeline
|
3 |
import os
|
|
|
4 |
|
5 |
theme = "darkgrass"
|
6 |
title = "GPT-NeoX(Korean) Demo"
|
@@ -14,10 +15,25 @@ examples = [
|
|
14 |
["2040년 미국은, "]
|
15 |
]
|
16 |
|
17 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
18 |
|
19 |
def predict(text):
|
20 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
21 |
|
22 |
iface = gr.Interface(
|
23 |
fn=predict,
|
|
|
1 |
import gradio as gr
|
2 |
+
from transformers import pipeline, AutoTokenizer, AutoModelForCausalLM, StoppingCriteriaList
|
3 |
import os
|
4 |
+
import torch
|
5 |
|
6 |
theme = "darkgrass"
|
7 |
title = "GPT-NeoX(Korean) Demo"
|
|
|
15 |
["2040년 미국은, "]
|
16 |
]
|
17 |
|
18 |
+
tokenizer = AutoTokenizer.from_pretrained("EleutherAI/gpt-neox-ko-1.3b", use_auth_token=os.environ['TOKEN'])
|
19 |
+
model = AutoModelForCausalLM.from_pretrained(
|
20 |
+
"EleutherAI/gpt-neox-ko-1.3b", use_auth_token=os.environ['TOKEN'],
|
21 |
+
pad_token_id=tokenizer.pad_token_id,
|
22 |
+
eos_token_id=tokenizer.eos_token_id,
|
23 |
+
bos_token_id=tokenizer.bos_token_id
|
24 |
+
)
|
25 |
+
model.eval()
|
26 |
+
|
27 |
+
pipe = pipeline('text-generation', model=model, tokenizer=tokenizer)
|
28 |
|
29 |
def predict(text):
|
30 |
+
stopping_cond = StoppingCriteriaList([tokenizer.encode('<|endoftext|>')])
|
31 |
+
with torch.no_grad():
|
32 |
+
tokens = tokenizer(text, return_tensors="pt").input_ids
|
33 |
+
gen_tokens = model.generate(tokens, do_sample=True, temperature=0.8, max_new_tokens=64)
|
34 |
+
generated = tokenizer.batch_decode(gen_tokens)[0]
|
35 |
+
return generated
|
36 |
+
# return pipe(text)[0]['generated_text']
|
37 |
|
38 |
iface = gr.Interface(
|
39 |
fn=predict,
|