Update app.py
Browse files
app.py
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
from transformers import
|
2 |
import torch
|
3 |
import os
|
4 |
import gradio as gr
|
@@ -6,15 +6,8 @@ import sentencepiece
|
|
6 |
# from tokenization_yi import YiTokenizer
|
7 |
|
8 |
|
9 |
-
|
10 |
-
|
11 |
-
import os
|
12 |
-
import gradio as gr
|
13 |
-
import sentencepiece
|
14 |
-
from tokenization_yi import YiTokenizer
|
15 |
-
|
16 |
-
os.environ['PYTORCH_CUDA_ALLOC_CONF'] = 'max_split_size_mb:50'
|
17 |
-
model_id = "TheBloke/Yi-34B-200K-Llamafied-GPTQ"
|
18 |
|
19 |
|
20 |
tokenizer = AutoTokenizer.from_pretrained("larryvrh/Yi-34B-200K-Llamafied")
|
@@ -54,7 +47,7 @@ with gr.Blocks(theme='ParityError/Anime') as demo:
|
|
54 |
chatbot = gr.Chatbot(label='TonicYi-30B-200K')
|
55 |
|
56 |
with gr.Accordion(label='Advanced options', open=False):
|
57 |
-
max_new_tokens = gr.Slider(label='Max New Tokens', minimum=1, maximum=
|
58 |
temperature = gr.Slider(label='Temperature', minimum=0.1, maximum=4.0, step=0.1, value=1.2)
|
59 |
top_p = gr.Slider(label='Top-P (nucleus sampling)', minimum=0.05, maximum=1.0, step=0.05, value=0.9)
|
60 |
top_k = gr.Slider(label='Top-K', minimum=1, maximum=1000, step=1, value=900)
|
|
|
1 |
+
from transformers import AutoTokenizer, AutoModelForCausalLM
|
2 |
import torch
|
3 |
import os
|
4 |
import gradio as gr
|
|
|
6 |
# from tokenization_yi import YiTokenizer
|
7 |
|
8 |
|
9 |
+
os.environ['PYTORCH_CUDA_ALLOC_CONF'] = 'max_split_size_mb:120'
|
10 |
+
model_id = "larryvrh/Yi-34B-200K-Llamafied"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
11 |
|
12 |
|
13 |
tokenizer = AutoTokenizer.from_pretrained("larryvrh/Yi-34B-200K-Llamafied")
|
|
|
47 |
chatbot = gr.Chatbot(label='TonicYi-30B-200K')
|
48 |
|
49 |
with gr.Accordion(label='Advanced options', open=False):
|
50 |
+
max_new_tokens = gr.Slider(label='Max New Tokens', minimum=1, maximum=55000, step=1, value=3800)
|
51 |
temperature = gr.Slider(label='Temperature', minimum=0.1, maximum=4.0, step=0.1, value=1.2)
|
52 |
top_p = gr.Slider(label='Top-P (nucleus sampling)', minimum=0.05, maximum=1.0, step=0.05, value=0.9)
|
53 |
top_k = gr.Slider(label='Top-K', minimum=1, maximum=1000, step=1, value=900)
|