riverallzero
commited on
Commit
•
bb4ec8b
1
Parent(s):
f1de462
upload files
Browse files- .gitattributes +1 -0
- README.md +28 -0
- adapter_config.json +19 -0
- adapter_model.bin +3 -0
.gitattributes
CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
|
|
|
33 |
*.zip filter=lfs diff=lfs merge=lfs -text
|
34 |
*.zst filter=lfs diff=lfs merge=lfs -text
|
35 |
*tfevents* filter=lfs diff=lfs merge=lfs -text
|
36 |
+
adapter_model.bin filter=lfs diff=lfs merge=lfs -text
|
README.md
CHANGED
@@ -1,3 +1,31 @@
|
|
1 |
---
|
2 |
license: mit
|
|
|
|
|
3 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
---
|
2 |
license: mit
|
3 |
+
datasets:
|
4 |
+
- yahma/alpaca-cleaned
|
5 |
---
|
6 |
+
|
7 |
+
This repo contains a low-rank adapter for LLaMA-7b
|
8 |
+
fit on the [Stanford Alpaca](https://github.com/tatsu-lab/stanford_alpaca) dataset.
|
9 |
+
|
10 |
+
This version of the weights was trained with the following hyperparameters:
|
11 |
+
|
12 |
+
- Epochs: 10 (load from best epoch)
|
13 |
+
- Batch size: 128
|
14 |
+
- Cutoff length: 512
|
15 |
+
- Learning rate: 3e-4
|
16 |
+
- Lora _r_: 16
|
17 |
+
- Lora target modules: q_proj, k_proj, v_proj, o_proj
|
18 |
+
|
19 |
+
That is:
|
20 |
+
|
21 |
+
```
|
22 |
+
python finetune.py \
|
23 |
+
--base_model='baffo32/decapoda-research-llama-7B-hf' \
|
24 |
+
--num_epochs=10 \
|
25 |
+
--cutoff_len=512 \
|
26 |
+
--group_by_length \
|
27 |
+
--output_dir='./lora-alpaca-512-qkvo' \
|
28 |
+
--lora_target_modules='[q_proj,k_proj,v_proj,o_proj]' \
|
29 |
+
--lora_r=16 \
|
30 |
+
--micro_batch_size=8
|
31 |
+
```
|
adapter_config.json
ADDED
@@ -0,0 +1,19 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"base_model_name_or_path": "baffo32/decapoda-research-llama-7B-hf",
|
3 |
+
"bias": "none",
|
4 |
+
"fan_in_fan_out": false,
|
5 |
+
"inference_mode": true,
|
6 |
+
"lora_alpha": 16,
|
7 |
+
"lora_dropout": 0.05,
|
8 |
+
"merge_weights": false,
|
9 |
+
"modules_to_save": null,
|
10 |
+
"peft_type": "LORA",
|
11 |
+
"r": 16,
|
12 |
+
"target_modules": [
|
13 |
+
"q_proj",
|
14 |
+
"k_proj",
|
15 |
+
"v_proj",
|
16 |
+
"o_proj"
|
17 |
+
],
|
18 |
+
"task_type": "CAUSAL_LM"
|
19 |
+
}
|
adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2e7187f51fbdeff8815046d30f0a325e43491040e6eac8cec5e2ba64f1e87807
|
3 |
+
size 67201357
|