Muhammad Khalifa commited on
Commit
e1fbab3
Β·
1 Parent(s): 7d22281

organize text models

Browse files
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/adapter_config.json RENAMED
File without changes
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/adapter_model.bin RENAMED
File without changes
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/optimizer.pt RENAMED
File without changes
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/rng_state.pth RENAMED
File without changes
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/scheduler.pt RENAMED
File without changes
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/trainer_state.json RENAMED
File without changes
{alpaca_text_52K β†’ text_models/llama-7b}/best_model/training_args.bin RENAMED
File without changes
text_models/llama-7b/checkpoint-13200/adapter_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "huggyllama/llama-7b",
3
+ "bias": "none",
4
+ "enable_lora": null,
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "lora_alpha": 16,
9
+ "lora_dropout": 0.05,
10
+ "merge_weights": false,
11
+ "modules_to_save": null,
12
+ "peft_type": "LORA",
13
+ "r": 16,
14
+ "target_modules": [
15
+ "q_proj",
16
+ "k_proj",
17
+ "v_proj",
18
+ "o_proj"
19
+ ],
20
+ "task_type": "CAUSAL_LM"
21
+ }
text_models/llama-7b/checkpoint-13200/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8b6d477b267009b4786d2a63e2d5d61b8917b0a131f3235d830723fb0380743b
3
+ size 67201357
text_models/llama-7b/checkpoint-13200/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7bd6b1fb9a64d8514dfce7796a097fee932715f8ea649dca54610a4f2cd33de6
3
+ size 134433093
text_models/llama-7b/checkpoint-13200/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9f407c9f92ac9f4eb5ae6ca6bb1259d9c1f56c08d983914f32c3045ff8ead46
3
+ size 14575
text_models/llama-7b/checkpoint-13200/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75208e1f32c6e4de232bd88f689d7913bd828d91f34dbf36cfff26e02e76339c
3
+ size 627
text_models/llama-7b/checkpoint-13200/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
text_models/llama-7b/checkpoint-13200/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09cbbfbe6d6c002252235b4476abb50ae8f0d55947cd46a53991e96578a822e
3
+ size 4027
text_models/llama-7b/checkpoint-13400/adapter_config.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "base_model_name_or_path": "huggyllama/llama-7b",
3
+ "bias": "none",
4
+ "enable_lora": null,
5
+ "fan_in_fan_out": false,
6
+ "inference_mode": true,
7
+ "init_lora_weights": true,
8
+ "lora_alpha": 16,
9
+ "lora_dropout": 0.05,
10
+ "merge_weights": false,
11
+ "modules_to_save": null,
12
+ "peft_type": "LORA",
13
+ "r": 16,
14
+ "target_modules": [
15
+ "q_proj",
16
+ "k_proj",
17
+ "v_proj",
18
+ "o_proj"
19
+ ],
20
+ "task_type": "CAUSAL_LM"
21
+ }
text_models/llama-7b/checkpoint-13400/adapter_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0662963c53b42f28844640476614c450550c1ee38420627d2c0b89bf3d87ff7f
3
+ size 67201357
text_models/llama-7b/checkpoint-13400/optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5f36370ad8a45579106121b103e0be8cad5af9a5db6f0a86bd77e9d31893add5
3
+ size 134433093
text_models/llama-7b/checkpoint-13400/rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:150b5b8852d55b0c2d35a82e3b0dde547eef5e7a801f8ef3ff255e5bd7a581d2
3
+ size 14575
text_models/llama-7b/checkpoint-13400/scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b56ee2fcdb9065ff03101361b9a7fc08f144856ecdb5b7c5d351035a6ce9ed7
3
+ size 627
text_models/llama-7b/checkpoint-13400/trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
text_models/llama-7b/checkpoint-13400/training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a09cbbfbe6d6c002252235b4476abb50ae8f0d55947cd46a53991e96578a822e
3
+ size 4027