jordyvl commited on
Commit
496f360
1 Parent(s): 62d0d52

bitsandbytes added, checking with gpu

Browse files
Files changed (3) hide show
  1. README.md +2 -1
  2. app.py +4 -4
  3. requirements.txt +3 -1
README.md CHANGED
@@ -8,9 +8,10 @@ sdk_version: 4.26.0
8
  app_file: app.py
9
  pinned: false
10
  preload_from_hub:
11
- - "microsoft/phi-2"
12
  - "BAAI/bge-small-en-v1.5"
13
  - "HuggingFaceH4/zephyr-7b-alpha"
14
  - "meta-llama/Meta-Llama-3-8B"
15
  ---
16
 
 
 
 
8
  app_file: app.py
9
  pinned: false
10
  preload_from_hub:
 
11
  - "BAAI/bge-small-en-v1.5"
12
  - "HuggingFaceH4/zephyr-7b-alpha"
13
  - "meta-llama/Meta-Llama-3-8B"
14
  ---
15
 
16
+ - "microsoft/phi-2"
17
+
app.py CHANGED
@@ -20,12 +20,12 @@ CHEAPMODE = torch.cuda.is_available()
20
  # LLM = "HuggingFaceH4/zephyr-7b-alpha" if not CHEAPMODE else "microsoft/phi-2"
21
 
22
  config = {
23
- # "LLM": "meta-llama/Meta-Llama-3-8B",
24
- "LLM": "microsoft/phi-2",
25
  # "LLM": "HuggingFaceH4/zephyr-7b-alpha",
26
  "embeddings": "BAAI/bge-small-en-v1.5",
27
  "similarity_top_k": 2,
28
- "context_window": 4048,
29
  "max_new_tokens": 200,
30
  "temperature": 0.7,
31
  "top_k": 5,
@@ -184,7 +184,7 @@ def ask_my_thesis(
184
  config["temperature"] = temperature
185
  config["top_p"] = top_p
186
  config["max_new_tokens"] = max_new_tokens
187
- # config["LLM"] = LLM
188
  # config["embeddings"] = embeddings
189
  config["similarity_top_k"] = similarity_top_k
190
  config["context_window"] = context_window
 
20
  # LLM = "HuggingFaceH4/zephyr-7b-alpha" if not CHEAPMODE else "microsoft/phi-2"
21
 
22
  config = {
23
+ "LLM": "meta-llama/Meta-Llama-3-8B",
24
+ # "LLM": "microsoft/phi-2",
25
  # "LLM": "HuggingFaceH4/zephyr-7b-alpha",
26
  "embeddings": "BAAI/bge-small-en-v1.5",
27
  "similarity_top_k": 2,
28
+ "context_window": 2048,
29
  "max_new_tokens": 200,
30
  "temperature": 0.7,
31
  "top_k": 5,
 
184
  config["temperature"] = temperature
185
  config["top_p"] = top_p
186
  config["max_new_tokens"] = max_new_tokens
187
+ config["LLM"] = LLM
188
  # config["embeddings"] = embeddings
189
  config["similarity_top_k"] = similarity_top_k
190
  config["context_window"] = context_window
requirements.txt CHANGED
@@ -7,4 +7,6 @@ tokenizers
7
  llama-index
8
  llama-index-embeddings-huggingface
9
  llama-index-llms-huggingface
10
- einops
 
 
 
7
  llama-index
8
  llama-index-embeddings-huggingface
9
  llama-index-llms-huggingface
10
+ einops
11
+ accelerate
12
+ bitsandbytes