Spaces:
Paused
Paused
bitsandbytes added, checking with gpu
Browse files- README.md +2 -1
- app.py +4 -4
- requirements.txt +3 -1
README.md
CHANGED
@@ -8,9 +8,10 @@ sdk_version: 4.26.0
|
|
8 |
app_file: app.py
|
9 |
pinned: false
|
10 |
preload_from_hub:
|
11 |
-
- "microsoft/phi-2"
|
12 |
- "BAAI/bge-small-en-v1.5"
|
13 |
- "HuggingFaceH4/zephyr-7b-alpha"
|
14 |
- "meta-llama/Meta-Llama-3-8B"
|
15 |
---
|
16 |
|
|
|
|
|
|
8 |
app_file: app.py
|
9 |
pinned: false
|
10 |
preload_from_hub:
|
|
|
11 |
- "BAAI/bge-small-en-v1.5"
|
12 |
- "HuggingFaceH4/zephyr-7b-alpha"
|
13 |
- "meta-llama/Meta-Llama-3-8B"
|
14 |
---
|
15 |
|
16 |
+
- "microsoft/phi-2"
|
17 |
+
|
app.py
CHANGED
@@ -20,12 +20,12 @@ CHEAPMODE = torch.cuda.is_available()
|
|
20 |
# LLM = "HuggingFaceH4/zephyr-7b-alpha" if not CHEAPMODE else "microsoft/phi-2"
|
21 |
|
22 |
config = {
|
23 |
-
|
24 |
-
"LLM": "microsoft/phi-2",
|
25 |
# "LLM": "HuggingFaceH4/zephyr-7b-alpha",
|
26 |
"embeddings": "BAAI/bge-small-en-v1.5",
|
27 |
"similarity_top_k": 2,
|
28 |
-
"context_window":
|
29 |
"max_new_tokens": 200,
|
30 |
"temperature": 0.7,
|
31 |
"top_k": 5,
|
@@ -184,7 +184,7 @@ def ask_my_thesis(
|
|
184 |
config["temperature"] = temperature
|
185 |
config["top_p"] = top_p
|
186 |
config["max_new_tokens"] = max_new_tokens
|
187 |
-
|
188 |
# config["embeddings"] = embeddings
|
189 |
config["similarity_top_k"] = similarity_top_k
|
190 |
config["context_window"] = context_window
|
|
|
20 |
# LLM = "HuggingFaceH4/zephyr-7b-alpha" if not CHEAPMODE else "microsoft/phi-2"
|
21 |
|
22 |
config = {
|
23 |
+
"LLM": "meta-llama/Meta-Llama-3-8B",
|
24 |
+
# "LLM": "microsoft/phi-2",
|
25 |
# "LLM": "HuggingFaceH4/zephyr-7b-alpha",
|
26 |
"embeddings": "BAAI/bge-small-en-v1.5",
|
27 |
"similarity_top_k": 2,
|
28 |
+
"context_window": 2048,
|
29 |
"max_new_tokens": 200,
|
30 |
"temperature": 0.7,
|
31 |
"top_k": 5,
|
|
|
184 |
config["temperature"] = temperature
|
185 |
config["top_p"] = top_p
|
186 |
config["max_new_tokens"] = max_new_tokens
|
187 |
+
config["LLM"] = LLM
|
188 |
# config["embeddings"] = embeddings
|
189 |
config["similarity_top_k"] = similarity_top_k
|
190 |
config["context_window"] = context_window
|
requirements.txt
CHANGED
@@ -7,4 +7,6 @@ tokenizers
|
|
7 |
llama-index
|
8 |
llama-index-embeddings-huggingface
|
9 |
llama-index-llms-huggingface
|
10 |
-
einops
|
|
|
|
|
|
7 |
llama-index
|
8 |
llama-index-embeddings-huggingface
|
9 |
llama-index-llms-huggingface
|
10 |
+
einops
|
11 |
+
accelerate
|
12 |
+
bitsandbytes
|