aixsatoshi commited on
Commit
bc71774
·
verified ·
1 Parent(s): 6c16cee

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -6
app.py CHANGED
@@ -1,5 +1,4 @@
1
  import torch
2
- from PIL import Image
3
  import gradio as gr
4
  import spaces
5
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
@@ -9,16 +8,16 @@ import random
9
  from datasets import load_dataset
10
 
11
  HF_TOKEN = os.environ.get("HF_TOKEN", None)
12
- MODEL_ID = "aixsatoshi/Meta-Llama-3.1-8B-Instruct-plus-Swallow"
13
  MODELS = os.environ.get("MODELS")
14
  MODEL_NAME = MODEL_ID.split("/")[-1]
15
 
16
  TITLE = "<h1><center>New japanese LLM model webui</center></h1>"
17
 
18
  DESCRIPTION = f"""
19
- <h3>MODEL: <a href="https://huggingface.co/aixsatoshi/Meta-Llama-3.1-8B-Instruct-plus-Swallow">Meta-Llama-3.1-8B-Instruct-plus-Swallow</a></h3>
20
  <center>
21
- <p>aixsatoshi/Meta-Llama-3.1-8B-Instruct-plus-Swallow is the merged model.
22
  <br>
23
  Feel free to test without log.
24
  </p>
@@ -46,7 +45,7 @@ h3 {
46
  # モデルとトークナイザーの読み込み
47
  model = AutoModelForCausalLM.from_pretrained(
48
  MODEL_ID,
49
- torch_dtype=torch.float16,
50
  device_map="auto",
51
  )
52
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
@@ -86,7 +85,7 @@ def stream_chat(message: str, history: list, temperature: float, max_new_tokens:
86
  max_new_tokens=max_new_tokens,
87
  do_sample=True,
88
  temperature=temperature,
89
- eos_token_id=[128001, 128009],
90
  )
91
 
92
  thread = Thread(target=model.generate, kwargs=generate_kwargs)
 
1
  import torch
 
2
  import gradio as gr
3
  import spaces
4
  from transformers import AutoModelForCausalLM, AutoTokenizer, TextIteratorStreamer
 
8
  from datasets import load_dataset
9
 
10
  HF_TOKEN = os.environ.get("HF_TOKEN", None)
11
+ MODEL_ID = "Aratako/sarashina2.1-1b-sft"
12
  MODELS = os.environ.get("MODELS")
13
  MODEL_NAME = MODEL_ID.split("/")[-1]
14
 
15
  TITLE = "<h1><center>New japanese LLM model webui</center></h1>"
16
 
17
  DESCRIPTION = f"""
18
+ <h3>MODEL: <a href="https://huggingface.co/Aratako/sarashina2.1-1b-sft">Aratako/sarashina2.1-1b-sft</a></h3>
19
  <center>
20
+ <p>
21
  <br>
22
  Feel free to test without log.
23
  </p>
 
45
  # モデルとトークナイザーの読み込み
46
  model = AutoModelForCausalLM.from_pretrained(
47
  MODEL_ID,
48
+ torch_dtype=torch.bfloat16,
49
  device_map="auto",
50
  )
51
  tokenizer = AutoTokenizer.from_pretrained(MODEL_ID)
 
85
  max_new_tokens=max_new_tokens,
86
  do_sample=True,
87
  temperature=temperature,
88
+ #eos_token_id=[128001, 128009],
89
  )
90
 
91
  thread = Thread(target=model.generate, kwargs=generate_kwargs)