masanorihirano
commited on
Commit
•
724bbf4
1
Parent(s):
82d194e
update
Browse files
app.py
CHANGED
@@ -258,7 +258,7 @@ description = (
|
|
258 |
"It is a 13B-parameter LLaMA model finetuned to follow instructions. "
|
259 |
"It is trained on the [izumi-lab/llm-japanese-dataset](https://huggingface.co/datasets/izumi-lab/llm-japanese-dataset) dataset. "
|
260 |
"For more information, please visit [the project's website](https://llm.msuzuki.me). "
|
261 |
-
"This model can output up to 256 tokens, but the maximum number of tokens is
|
262 |
"It takes about **1 minute** to output. When access is concentrated, the operation may become slow."
|
263 |
)
|
264 |
with gr.Blocks(
|
@@ -293,8 +293,8 @@ with gr.Blocks(
|
|
293 |
)
|
294 |
max_tokens = gr.Slider(
|
295 |
minimum=20,
|
296 |
-
maximum=
|
297 |
-
value=
|
298 |
step=1,
|
299 |
interactive=True,
|
300 |
label="Max length (Pre-prompt + instruction + input + output)",
|
|
|
258 |
"It is a 13B-parameter LLaMA model finetuned to follow instructions. "
|
259 |
"It is trained on the [izumi-lab/llm-japanese-dataset](https://huggingface.co/datasets/izumi-lab/llm-japanese-dataset) dataset. "
|
260 |
"For more information, please visit [the project's website](https://llm.msuzuki.me). "
|
261 |
+
"This model can output up to 256 tokens, but the maximum number of tokens is 200 due to the GPU memory limit of HuggingFace Space. "
|
262 |
"It takes about **1 minute** to output. When access is concentrated, the operation may become slow."
|
263 |
)
|
264 |
with gr.Blocks(
|
|
|
293 |
)
|
294 |
max_tokens = gr.Slider(
|
295 |
minimum=20,
|
296 |
+
maximum=200,
|
297 |
+
value=100,
|
298 |
step=1,
|
299 |
interactive=True,
|
300 |
label="Max length (Pre-prompt + instruction + input + output)",
|