modified: app.py
Browse filesnew file: requirements.txt
- app.py +3 -3
- requirements.txt +0 -0
app.py
CHANGED
@@ -1,11 +1,12 @@
|
|
1 |
-
# git clone https://huggingface.co/Pipatpong/
|
2 |
|
3 |
import gradio as gr
|
4 |
import re
|
|
|
5 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
6 |
|
7 |
checkpoint = "Pipatpong/vcm_santa"
|
8 |
-
device = "cpu"
|
9 |
tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
|
10 |
model = AutoModelForCausalLM.from_pretrained(checkpoint, trust_remote_code=True, device_map="auto", load_in_8bit=True)
|
11 |
|
@@ -13,7 +14,6 @@ def generate(text, max_length, num_return_sequences=1):
|
|
13 |
inputs = tokenizer.encode(text, padding=False, add_special_tokens=False, return_tensors="pt")
|
14 |
outputs = model.generate(inputs, max_length=max_length, num_return_sequences=num_return_sequences)
|
15 |
gen_text = "Assignment : " + tokenizer.decode(outputs[0]).split("#")[0] if "#" else "Assignment : " + tokenizer.decode(outputs[0])
|
16 |
-
|
17 |
return gen_text
|
18 |
|
19 |
|
|
|
1 |
+
# git clone https://huggingface.co/spaces/Pipatpong/VCM_Demo
|
2 |
|
3 |
import gradio as gr
|
4 |
import re
|
5 |
+
import torch
|
6 |
from transformers import AutoModelForCausalLM, AutoTokenizer
|
7 |
|
8 |
checkpoint = "Pipatpong/vcm_santa"
|
9 |
+
device = "cuda" if torch.cuda.is_available() else "cpu"
|
10 |
tokenizer = AutoTokenizer.from_pretrained(checkpoint, trust_remote_code=True)
|
11 |
model = AutoModelForCausalLM.from_pretrained(checkpoint, trust_remote_code=True, device_map="auto", load_in_8bit=True)
|
12 |
|
|
|
14 |
inputs = tokenizer.encode(text, padding=False, add_special_tokens=False, return_tensors="pt")
|
15 |
outputs = model.generate(inputs, max_length=max_length, num_return_sequences=num_return_sequences)
|
16 |
gen_text = "Assignment : " + tokenizer.decode(outputs[0]).split("#")[0] if "#" else "Assignment : " + tokenizer.decode(outputs[0])
|
|
|
17 |
return gen_text
|
18 |
|
19 |
|
requirements.txt
ADDED
Binary file (5.03 kB). View file
|
|