ZeroUniqueness
commited on
Commit
•
9f166b1
1
Parent(s):
a59cceb
Training in progress, step 300
Browse files- adapter_config.json +4 -4
- adapter_model.bin +2 -2
- checkpoint-300/README.md +20 -0
- checkpoint-300/adapter_config.json +26 -0
- checkpoint-300/adapter_model.bin +3 -0
- checkpoint-300/optimizer.pt +3 -0
- checkpoint-300/rng_state_0.pth +3 -0
- checkpoint-300/rng_state_1.pth +3 -0
- checkpoint-300/rng_state_10.pth +3 -0
- checkpoint-300/rng_state_11.pth +3 -0
- checkpoint-300/rng_state_12.pth +3 -0
- checkpoint-300/rng_state_13.pth +3 -0
- checkpoint-300/rng_state_2.pth +3 -0
- checkpoint-300/rng_state_3.pth +3 -0
- checkpoint-300/rng_state_4.pth +3 -0
- checkpoint-300/rng_state_5.pth +3 -0
- checkpoint-300/rng_state_6.pth +3 -0
- checkpoint-300/rng_state_7.pth +3 -0
- checkpoint-300/rng_state_8.pth +3 -0
- checkpoint-300/rng_state_9.pth +3 -0
- checkpoint-300/scheduler.pt +3 -0
- checkpoint-300/trainer_state.json +64 -0
- checkpoint-300/training_args.bin +3 -0
- training_args.bin +1 -1
adapter_config.json
CHANGED
@@ -3,7 +3,7 @@
|
|
3 |
"base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
|
4 |
"bias": "none",
|
5 |
"fan_in_fan_out": null,
|
6 |
-
"inference_mode":
|
7 |
"init_lora_weights": true,
|
8 |
"layers_pattern": null,
|
9 |
"layers_to_transform": null,
|
@@ -15,12 +15,12 @@
|
|
15 |
"revision": null,
|
16 |
"target_modules": [
|
17 |
"v_proj",
|
18 |
-
"k_proj",
|
19 |
"q_proj",
|
20 |
-
"down_proj",
|
21 |
"up_proj",
|
22 |
"gate_proj",
|
23 |
-
"o_proj"
|
|
|
|
|
24 |
],
|
25 |
"task_type": "CAUSAL_LM"
|
26 |
}
|
|
|
3 |
"base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
|
4 |
"bias": "none",
|
5 |
"fan_in_fan_out": null,
|
6 |
+
"inference_mode": true,
|
7 |
"init_lora_weights": true,
|
8 |
"layers_pattern": null,
|
9 |
"layers_to_transform": null,
|
|
|
15 |
"revision": null,
|
16 |
"target_modules": [
|
17 |
"v_proj",
|
|
|
18 |
"q_proj",
|
|
|
19 |
"up_proj",
|
20 |
"gate_proj",
|
21 |
+
"o_proj",
|
22 |
+
"k_proj",
|
23 |
+
"down_proj"
|
24 |
],
|
25 |
"task_type": "CAUSAL_LM"
|
26 |
}
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6361cf788eb2eecf574488f94018046acb7d636e2b8c01e3508587db412f3cd2
|
3 |
+
size 500897101
|
checkpoint-300/README.md
ADDED
@@ -0,0 +1,20 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
library_name: peft
|
3 |
+
---
|
4 |
+
## Training procedure
|
5 |
+
|
6 |
+
|
7 |
+
The following `bitsandbytes` quantization config was used during training:
|
8 |
+
- load_in_8bit: False
|
9 |
+
- load_in_4bit: True
|
10 |
+
- llm_int8_threshold: 6.0
|
11 |
+
- llm_int8_skip_modules: None
|
12 |
+
- llm_int8_enable_fp32_cpu_offload: False
|
13 |
+
- llm_int8_has_fp16_weight: False
|
14 |
+
- bnb_4bit_quant_type: nf4
|
15 |
+
- bnb_4bit_use_double_quant: True
|
16 |
+
- bnb_4bit_compute_dtype: bfloat16
|
17 |
+
### Framework versions
|
18 |
+
|
19 |
+
|
20 |
+
- PEFT 0.5.0.dev0
|
checkpoint-300/adapter_config.json
ADDED
@@ -0,0 +1,26 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"auto_mapping": null,
|
3 |
+
"base_model_name_or_path": "/workspace/webui/models/TheBloke_Llama-2-13B-fp16",
|
4 |
+
"bias": "none",
|
5 |
+
"fan_in_fan_out": null,
|
6 |
+
"inference_mode": true,
|
7 |
+
"init_lora_weights": true,
|
8 |
+
"layers_pattern": null,
|
9 |
+
"layers_to_transform": null,
|
10 |
+
"lora_alpha": 16,
|
11 |
+
"lora_dropout": 0.05,
|
12 |
+
"modules_to_save": null,
|
13 |
+
"peft_type": "LORA",
|
14 |
+
"r": 32,
|
15 |
+
"revision": null,
|
16 |
+
"target_modules": [
|
17 |
+
"v_proj",
|
18 |
+
"q_proj",
|
19 |
+
"up_proj",
|
20 |
+
"gate_proj",
|
21 |
+
"o_proj",
|
22 |
+
"k_proj",
|
23 |
+
"down_proj"
|
24 |
+
],
|
25 |
+
"task_type": "CAUSAL_LM"
|
26 |
+
}
|
checkpoint-300/adapter_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6361cf788eb2eecf574488f94018046acb7d636e2b8c01e3508587db412f3cd2
|
3 |
+
size 500897101
|
checkpoint-300/optimizer.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dd2fe5fdce6d5e52a0f219229ef5e391d0a99183721768728e8618542924bce8
|
3 |
+
size 1001752701
|
checkpoint-300/rng_state_0.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f73e30dba1eba7a17e72336e01b55d9f45cc4bc93ae7772a5dd6d51e3eb71bc0
|
3 |
+
size 27772
|
checkpoint-300/rng_state_1.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:032ef3580361793ee9ae976197f72bd9d02821e12612fe786271ce9e4dcd88b5
|
3 |
+
size 27772
|
checkpoint-300/rng_state_10.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:366a4f691f1434c9d766f311626db4378598647acca195bc6d82610e0bad0cb7
|
3 |
+
size 27789
|
checkpoint-300/rng_state_11.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e8338173967d92a4366c887b8a90ea8dd942131116116d87e8664c05dc74c5b8
|
3 |
+
size 27789
|
checkpoint-300/rng_state_12.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5cc877da8acd38bfe8a82ebad2bbdbbcb635a54553f3f0c69a81c306dd090b33
|
3 |
+
size 27789
|
checkpoint-300/rng_state_13.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce8e16c1236a73d6d5a05b3abccc1b499017b6d4e023e6a3fb938ee4da8e12e6
|
3 |
+
size 27789
|
checkpoint-300/rng_state_2.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9e538c7cbe0d4aa5816d8891148177e39eec3caba14d5d438714c5025dbb7b63
|
3 |
+
size 27772
|
checkpoint-300/rng_state_3.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:49c89b7ec9b7d44c53d61af4eb045658960f0e49e18017e5b57efd1ede7c9dd3
|
3 |
+
size 27772
|
checkpoint-300/rng_state_4.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:51d5c7202dac1a2747fe82418a615633cd49f78d8b50431a951d19cb06f5df1c
|
3 |
+
size 27772
|
checkpoint-300/rng_state_5.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c701298fb37793e83edaddc3029806245162b95a14c7e799b7d21c004ec3ad6
|
3 |
+
size 27772
|
checkpoint-300/rng_state_6.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:50ae123c5d62d3799cc74217f729393187b853cdd0560d27d2237f7a7305463c
|
3 |
+
size 27772
|
checkpoint-300/rng_state_7.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:352860b2c59c403aebafef8048b27372066afd9fb8461941c15fc3b27039dee7
|
3 |
+
size 27772
|
checkpoint-300/rng_state_8.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d43a150c0c1d9fae996e73e5aab606c75628ca1d714f5428c2efe9ab14b1d382
|
3 |
+
size 27772
|
checkpoint-300/rng_state_9.pth
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ea39e6e01b60b67bbf2308b1fa8ffb4f498f576373d684099f9db2ad38980129
|
3 |
+
size 27772
|
checkpoint-300/scheduler.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b902cbfc5361903023808ab50ffb35c827f082341566e01dcf44743fa063659
|
3 |
+
size 627
|
checkpoint-300/trainer_state.json
ADDED
@@ -0,0 +1,64 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"best_metric": null,
|
3 |
+
"best_model_checkpoint": null,
|
4 |
+
"epoch": 0.11632415664986429,
|
5 |
+
"global_step": 300,
|
6 |
+
"is_hyper_param_search": false,
|
7 |
+
"is_local_process_zero": true,
|
8 |
+
"is_world_process_zero": true,
|
9 |
+
"log_history": [
|
10 |
+
{
|
11 |
+
"epoch": 0.02,
|
12 |
+
"learning_rate": 0.0001999867761371633,
|
13 |
+
"loss": 1.0435,
|
14 |
+
"step": 50
|
15 |
+
},
|
16 |
+
{
|
17 |
+
"epoch": 0.04,
|
18 |
+
"learning_rate": 0.00019993306018843102,
|
19 |
+
"loss": 0.8918,
|
20 |
+
"step": 100
|
21 |
+
},
|
22 |
+
{
|
23 |
+
"epoch": 0.06,
|
24 |
+
"learning_rate": 0.00019983804784290833,
|
25 |
+
"loss": 0.8874,
|
26 |
+
"step": 150
|
27 |
+
},
|
28 |
+
{
|
29 |
+
"epoch": 0.08,
|
30 |
+
"learning_rate": 0.00019970177836355307,
|
31 |
+
"loss": 0.8839,
|
32 |
+
"step": 200
|
33 |
+
},
|
34 |
+
{
|
35 |
+
"epoch": 0.09,
|
36 |
+
"learning_rate": 0.00019961818913082012,
|
37 |
+
"loss": 0.8801,
|
38 |
+
"step": 225
|
39 |
+
},
|
40 |
+
{
|
41 |
+
"epoch": 0.1,
|
42 |
+
"learning_rate": 0.00019952430806244534,
|
43 |
+
"loss": 0.8753,
|
44 |
+
"step": 250
|
45 |
+
},
|
46 |
+
{
|
47 |
+
"epoch": 0.11,
|
48 |
+
"learning_rate": 0.00019942014485754635,
|
49 |
+
"loss": 0.8754,
|
50 |
+
"step": 275
|
51 |
+
},
|
52 |
+
{
|
53 |
+
"epoch": 0.12,
|
54 |
+
"learning_rate": 0.00019930571027751713,
|
55 |
+
"loss": 0.8751,
|
56 |
+
"step": 300
|
57 |
+
}
|
58 |
+
],
|
59 |
+
"max_steps": 7737,
|
60 |
+
"num_train_epochs": 3,
|
61 |
+
"total_flos": 1.290261776910254e+18,
|
62 |
+
"trial_name": null,
|
63 |
+
"trial_params": null
|
64 |
+
}
|
checkpoint-300/training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe6d7bbf6b8ed54bb30d4fb9291907ac865f24b396951fe9912d2874dc58e000
|
3 |
+
size 4027
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 4027
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fe6d7bbf6b8ed54bb30d4fb9291907ac865f24b396951fe9912d2874dc58e000
|
3 |
size 4027
|