Spaces:
Sleeping
Sleeping
Update app.py
Browse files
app.py
CHANGED
@@ -3,8 +3,8 @@ import gradio as gr
|
|
3 |
from huggingface_hub import hf_hub_download
|
4 |
from llama_cpp import Llama
|
5 |
|
6 |
-
REPO_ID = "keitokei1994/
|
7 |
-
MODEL_NAME = "
|
8 |
MAX_CONTEXT_LENGTH = 8192
|
9 |
CUDA = True
|
10 |
SYSTEM_PROMPT = "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability."
|
@@ -130,7 +130,7 @@ def clear_chat(chat_history_state, chat_message):
|
|
130 |
|
131 |
def gui(llm_chat):
|
132 |
with gr.Blocks(theme="NoCrypt/miku", css=css) as app:
|
133 |
-
gr.Markdown("#
|
134 |
gr.Markdown(
|
135 |
f"""
|
136 |
### This demo utilizes the repository ID {REPO_ID} with the model {MODEL_NAME}, powered by the LLaMA.cpp backend.
|
|
|
3 |
from huggingface_hub import hf_hub_download
|
4 |
from llama_cpp import Llama
|
5 |
|
6 |
+
REPO_ID = "keitokei1994/swallow-3-8B-sqlcoder-2x8B-GGUF"
|
7 |
+
MODEL_NAME = "swallow-3-8b-sqlcoder-2x8b.Q8_0.gguf"
|
8 |
MAX_CONTEXT_LENGTH = 8192
|
9 |
CUDA = True
|
10 |
SYSTEM_PROMPT = "You are a helpful, smart, kind, and efficient AI assistant. You always fulfill the user's requests to the best of your ability."
|
|
|
130 |
|
131 |
def gui(llm_chat):
|
132 |
with gr.Blocks(theme="NoCrypt/miku", css=css) as app:
|
133 |
+
gr.Markdown("# swallow-3-8b-sqlcoder-2x8b.Q8_0.gguf")
|
134 |
gr.Markdown(
|
135 |
f"""
|
136 |
### This demo utilizes the repository ID {REPO_ID} with the model {MODEL_NAME}, powered by the LLaMA.cpp backend.
|