Spaces:
Running
Running
Update app.py
Browse files
app.py
CHANGED
@@ -8,7 +8,7 @@ from llama_cpp import Llama
|
|
8 |
from huggingface_hub import hf_hub_download
|
9 |
|
10 |
hf_hub_download(repo_id="Pi3141/alpaca-lora-7B-ggml", filename="ggml-model-q4_1.bin", local_dir=".")
|
11 |
-
llm = Llama(model_path="./ggml-model-q4_1.bin"
|
12 |
|
13 |
|
14 |
ins = '''Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
@@ -132,7 +132,7 @@ with gr.Blocks(theme=seafoam, analytics_enabled=False, css=css) as demo:
|
|
132 |
gr.Examples(
|
133 |
examples=examples,
|
134 |
inputs=[instruction],
|
135 |
-
cache_examples=
|
136 |
fn=process_example,
|
137 |
outputs=[output],
|
138 |
)
|
|
|
8 |
from huggingface_hub import hf_hub_download
|
9 |
|
10 |
hf_hub_download(repo_id="Pi3141/alpaca-lora-7B-ggml", filename="ggml-model-q4_1.bin", local_dir=".")
|
11 |
+
llm = Llama(model_path="./ggml-model-q4_1.bin")
|
12 |
|
13 |
|
14 |
ins = '''Below is an instruction that describes a task. Write a response that appropriately completes the request.
|
|
|
132 |
gr.Examples(
|
133 |
examples=examples,
|
134 |
inputs=[instruction],
|
135 |
+
cache_examples=False,
|
136 |
fn=process_example,
|
137 |
outputs=[output],
|
138 |
)
|