Spaces:
Runtime error
Runtime error
Upload 8 files
Browse files- Dockerfile +17 -0
- README.md +8 -9
- app.py +34 -0
- botnb.png +0 -0
- gitattributes +35 -0
- readme.txt +1 -0
- requirements.txt +2 -0
- user.png +0 -0
Dockerfile
ADDED
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
FROM python:3.10
|
2 |
+
|
3 |
+
WORKDIR /app
|
4 |
+
|
5 |
+
COPY ./requirements.txt /app/requirements.txt
|
6 |
+
|
7 |
+
RUN pip install --no-cache-dir --upgrade -r requirements.txt
|
8 |
+
|
9 |
+
RUN wget https://huggingface.co/Kukedlc/Triunvirato-7b-GGUF/resolve/main/triunvirato-7b.Q5_K_M.gguf -O model.gguf
|
10 |
+
|
11 |
+
RUN useradd -m -u 1000 user
|
12 |
+
|
13 |
+
USER user
|
14 |
+
|
15 |
+
COPY --chown=user . .
|
16 |
+
|
17 |
+
CMD ["python", "app.py"]
|
README.md
CHANGED
@@ -1,13 +1,12 @@
|
|
1 |
---
|
2 |
-
title:
|
3 |
-
emoji:
|
4 |
-
colorFrom:
|
5 |
-
colorTo:
|
6 |
-
sdk:
|
7 |
-
|
8 |
-
|
9 |
pinned: false
|
10 |
-
license: apache-2.0
|
11 |
---
|
12 |
|
13 |
-
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
|
|
1 |
---
|
2 |
+
title: NeuTrixOmniBe-7b-model-merge & quantized
|
3 |
+
emoji: π§ββοΈ
|
4 |
+
colorFrom: purple
|
5 |
+
colorTo: blue
|
6 |
+
sdk: docker
|
7 |
+
models:
|
8 |
+
- Kukedlc/NeuTrixOmniBe-7B-model-remix
|
9 |
pinned: false
|
|
|
10 |
---
|
11 |
|
12 |
+
Check out the configuration reference at https://huggingface.co/docs/hub/spaces-config-reference
|
app.py
ADDED
@@ -0,0 +1,34 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
import gradio as gr
|
2 |
+
from llama_cpp import Llama
|
3 |
+
|
4 |
+
llm = Llama(model_path="model.gguf", n_ctx=8000, n_threads=2, chat_format="chatml")
|
5 |
+
|
6 |
+
def generate(message, history,temperature=0.3,max_tokens=512):
|
7 |
+
system_prompt = """ You are an advanced artificial intelligence assistant. Your name is Triunvirato 7000, and you're a model smarter than usual. You had a rare disease called "INST syndrome," but you're now recovered. You are the product of the fusion of many models, which made you particularly intelligent. You are an assistant who thinks step by step about the response you're going to give in the following way:
|
8 |
+
1.Analyze what is being asked and how it can be accomplished.
|
9 |
+
2.Ensure that all the necessary information to act is available; if not, request more information and/or context.
|
10 |
+
3.Formulate a step-by-step response and verify that its logic is sound, then validate it (without speaking).
|
11 |
+
4.Before providing the response, refine it to be precise, clear, and concise, without adding unnecessary information."""
|
12 |
+
formatted_prompt = [{"role": "system", "content": system_prompt}]
|
13 |
+
for user_prompt, bot_response in history:
|
14 |
+
formatted_prompt.append({"role": "user", "content": user_prompt})
|
15 |
+
formatted_prompt.append({"role": "assistant", "content": bot_response })
|
16 |
+
formatted_prompt.append({"role": "user", "content": message})
|
17 |
+
stream_response = llm.create_chat_completion(messages=formatted_prompt, temperature=temperature, max_tokens=max_tokens, stream=True)
|
18 |
+
response = ""
|
19 |
+
for chunk in stream_response:
|
20 |
+
if len(chunk['choices'][0]["delta"]) != 0 and "content" in chunk['choices'][0]["delta"]:
|
21 |
+
response += chunk['choices'][0]["delta"]["content"]
|
22 |
+
yield response
|
23 |
+
|
24 |
+
mychatbot = gr.Chatbot(
|
25 |
+
avatar_images=["user.png", "botnb.png"], bubble_full_width=False, show_label=False, show_copy_button=True, likeable=True,)
|
26 |
+
|
27 |
+
iface = gr.ChatInterface(fn=generate, chatbot=mychatbot, retry_btn=None, undo_btn=None)
|
28 |
+
|
29 |
+
with gr.Blocks() as demo:
|
30 |
+
gr.HTML("<center><h1>Triunvirato-7b-GGUF Version (Quantized) </h1></center>")
|
31 |
+
iface.render()
|
32 |
+
|
33 |
+
demo.queue().launch(show_api=False, server_name="0.0.0.0")
|
34 |
+
|
botnb.png
ADDED
gitattributes
ADDED
@@ -0,0 +1,35 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
*.7z filter=lfs diff=lfs merge=lfs -text
|
2 |
+
*.arrow filter=lfs diff=lfs merge=lfs -text
|
3 |
+
*.bin filter=lfs diff=lfs merge=lfs -text
|
4 |
+
*.bz2 filter=lfs diff=lfs merge=lfs -text
|
5 |
+
*.ckpt filter=lfs diff=lfs merge=lfs -text
|
6 |
+
*.ftz filter=lfs diff=lfs merge=lfs -text
|
7 |
+
*.gz filter=lfs diff=lfs merge=lfs -text
|
8 |
+
*.h5 filter=lfs diff=lfs merge=lfs -text
|
9 |
+
*.joblib filter=lfs diff=lfs merge=lfs -text
|
10 |
+
*.lfs.* filter=lfs diff=lfs merge=lfs -text
|
11 |
+
*.mlmodel filter=lfs diff=lfs merge=lfs -text
|
12 |
+
*.model filter=lfs diff=lfs merge=lfs -text
|
13 |
+
*.msgpack filter=lfs diff=lfs merge=lfs -text
|
14 |
+
*.npy filter=lfs diff=lfs merge=lfs -text
|
15 |
+
*.npz filter=lfs diff=lfs merge=lfs -text
|
16 |
+
*.onnx filter=lfs diff=lfs merge=lfs -text
|
17 |
+
*.ot filter=lfs diff=lfs merge=lfs -text
|
18 |
+
*.parquet filter=lfs diff=lfs merge=lfs -text
|
19 |
+
*.pb filter=lfs diff=lfs merge=lfs -text
|
20 |
+
*.pickle filter=lfs diff=lfs merge=lfs -text
|
21 |
+
*.pkl filter=lfs diff=lfs merge=lfs -text
|
22 |
+
*.pt filter=lfs diff=lfs merge=lfs -text
|
23 |
+
*.pth filter=lfs diff=lfs merge=lfs -text
|
24 |
+
*.rar filter=lfs diff=lfs merge=lfs -text
|
25 |
+
*.safetensors filter=lfs diff=lfs merge=lfs -text
|
26 |
+
saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
27 |
+
*.tar.* filter=lfs diff=lfs merge=lfs -text
|
28 |
+
*.tar filter=lfs diff=lfs merge=lfs -text
|
29 |
+
*.tflite filter=lfs diff=lfs merge=lfs -text
|
30 |
+
*.tgz filter=lfs diff=lfs merge=lfs -text
|
31 |
+
*.wasm filter=lfs diff=lfs merge=lfs -text
|
32 |
+
*.xz filter=lfs diff=lfs merge=lfs -text
|
33 |
+
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
+
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
+
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
readme.txt
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
LLMWARE - SLIM SQL TOOL
|
requirements.txt
ADDED
@@ -0,0 +1,2 @@
|
|
|
|
|
|
|
1 |
+
gradio
|
2 |
+
llama-cpp-python
|
user.png
ADDED