Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -1,5 +1,4 @@
|
|
1 |
import torch
|
2 |
-
from PIL import Image
|
3 |
import gradio as gr
|
4 |
import spaces
|
5 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
@@ -9,16 +8,16 @@ import random
|
|
9 |
from datasets import load_dataset
|
10 |
|
11 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
12 |
-
MODEL_ID = "
|
13 |
MODELS = os.environ.get("MODELS")
|
14 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
15 |
|
16 |
TITLE = "<h1><center>New japanese LLM model webui</center></h1>"
|
17 |
|
18 |
DESCRIPTION = f"""
|
19 |
-
<h3>MODEL: <a href="https://huggingface.co/
|
20 |
<center>
|
21 |
-
<p>
|
22 |
<br>
|
23 |
Feel free to test without log.
|
24 |
</p>
|
@@ -46,7 +45,7 @@ h3 {
|
|
46 |
# モデルとトークナイザーの読み込み
|
47 |
model = AutoModelForCausalLM.from_pretrained(
|
48 |
MODEL_ID,
|
49 |
-
torch_dtype=torch.
|
50 |
device_map="auto",
|
51 |
)
|
52 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
@@ -86,7 +85,7 @@ def stream_chat(message: str, history: list, temperature: float, max_new_tokens:
|
|
86 |
max_new_tokens=max_new_tokens,
|
87 |
do_sample=True,
|
88 |
temperature=temperature,
|
89 |
-
eos_token_id=[128001, 128009],
|
90 |
)
|
91 |
|
92 |
thread = Thread(target=model.generate, kwargs=generate_kwargs)
|
|
|
1 |
import torch
|
|
|
2 |
import gradio as gr
|
3 |
import spaces
|
4 |
from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
|
|
|
8 |
from datasets import load_dataset
|
9 |
|
10 |
HF_TOKEN = os.environ.get("HF_TOKEN", None)
|
11 |
+
MODEL_ID = "Aratako/sarashina2.1-1b-sft"
|
12 |
MODELS = os.environ.get("MODELS")
|
13 |
MODEL_NAME = MODEL_ID.split("/")[-1]
|
14 |
|
15 |
TITLE = "<h1><center>New japanese LLM model webui</center></h1>"
|
16 |
|
17 |
DESCRIPTION = f"""
|
18 |
+
<h3>MODEL: <a href="https://huggingface.co/Aratako/sarashina2.1-1b-sft">Aratako/sarashina2.1-1b-sft</a></h3>
|
19 |
<center>
|
20 |
+
<p>
|
21 |
<br>
|
22 |
Feel free to test without log.
|
23 |
</p>
|
|
|
45 |
# モデルとトークナイザーの読み込み
|
46 |
model = AutoModelForCausalLM.from_pretrained(
|
47 |
MODEL_ID,
|
48 |
+
torch_dtype=torch.bfloat16,
|
49 |
device_map="auto",
|
50 |
)
|
51 |
tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
|
|
|
85 |
max_new_tokens=max_new_tokens,
|
86 |
do_sample=True,
|
87 |
temperature=temperature,
|
88 |
+
#eos_token_id=[128001, 128009],
|
89 |
)
|
90 |
|
91 |
thread = Thread(target=model.generate, kwargs=generate_kwargs)
|