Spaces:
Sleeping
Sleeping
init
Browse files- app.py +4 -5
- packages.txt +1 -1
- requirements.txt +1 -1
app.py
CHANGED
@@ -160,16 +160,15 @@ def download_llama_7b(ckpt_dir, tokenizer_path):
|
|
160 |
|
161 |
def download_llama_adapter(instruct_adapter_path, caption_adapter_path):
|
162 |
if not os.path.exists(instruct_adapter_path):
|
163 |
-
os.system(f"wget -O {instruct_adapter_path} https://github.com/ZrrSkywalker/LLaMA-Adapter/releases/download/v.1.0.0/llama_adapter_len10_layer30_release.pth")
|
164 |
|
165 |
if not os.path.exists(caption_adapter_path):
|
166 |
-
os.system(f"wget -O {caption_adapter_path} https://github.com/ZrrSkywalker/LLaMA-Adapter/releases/download/v.1.0.0/llama_adapter_len10_layer30_caption_vit_l.pth")
|
167 |
|
168 |
|
169 |
-
#
|
|
|
170 |
# tokenizer_path = "/data1/llma/tokenizer.model"
|
171 |
-
# ckpt_dir = "LLaMA-7B/"
|
172 |
-
# tokenizer_path = "LLaMA-7B/tokenizer.model"
|
173 |
ckpt_path = hf_hub_download(repo_id="nyanko7/LLaMA-7B", filename="consolidated.00.pth")
|
174 |
param_path = hf_hub_download(repo_id="nyanko7/LLaMA-7B", filename="params.json")
|
175 |
tokenizer_path = hf_hub_download(repo_id="nyanko7/LLaMA-7B", filename="tokenizer.model")
|
|
|
160 |
|
161 |
def download_llama_adapter(instruct_adapter_path, caption_adapter_path):
|
162 |
if not os.path.exists(instruct_adapter_path):
|
163 |
+
os.system(f"wget -q -O {instruct_adapter_path} https://github.com/ZrrSkywalker/LLaMA-Adapter/releases/download/v.1.0.0/llama_adapter_len10_layer30_release.pth")
|
164 |
|
165 |
if not os.path.exists(caption_adapter_path):
|
166 |
+
os.system(f"wget -q -O {caption_adapter_path} https://github.com/ZrrSkywalker/LLaMA-Adapter/releases/download/v.1.0.0/llama_adapter_len10_layer30_caption_vit_l.pth")
|
167 |
|
168 |
|
169 |
+
# ckpt_path = "/data1/llma/7B/consolidated.00.pth"
|
170 |
+
# param_path = "/data1/llma/7B/params.json"
|
171 |
# tokenizer_path = "/data1/llma/tokenizer.model"
|
|
|
|
|
172 |
ckpt_path = hf_hub_download(repo_id="nyanko7/LLaMA-7B", filename="consolidated.00.pth")
|
173 |
param_path = hf_hub_download(repo_id="nyanko7/LLaMA-7B", filename="params.json")
|
174 |
tokenizer_path = hf_hub_download(repo_id="nyanko7/LLaMA-7B", filename="tokenizer.model")
|
packages.txt
CHANGED
@@ -1 +1 @@
|
|
1 |
-
|
|
|
1 |
+
cuda-toolkit-11-3
|
requirements.txt
CHANGED
@@ -1,4 +1,4 @@
|
|
1 |
-
|
2 |
fairscale
|
3 |
sentencepiece
|
4 |
Pillow
|
|
|
1 |
+
torch==1.12.0
|
2 |
fairscale
|
3 |
sentencepiece
|
4 |
Pillow
|