Training in progress, step 16324, checkpoint
Browse files- last-checkpoint/adapter_model.safetensors +1 -1
- last-checkpoint/global_step16324/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step16324/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step16324/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step16324/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt +3 -0
- last-checkpoint/global_step16324/zero_pp_rank_0_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step16324/zero_pp_rank_1_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step16324/zero_pp_rank_2_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/global_step16324/zero_pp_rank_3_mp_rank_00_model_states.pt +3 -0
- last-checkpoint/latest +1 -1
- last-checkpoint/scheduler.pt +1 -1
- last-checkpoint/trainer_state.json +37 -5
last-checkpoint/adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 29034840
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6dd27678e81e52b77b203ee21948888d64924194d564a7da5da8608977a649ef
|
3 |
size 29034840
|
last-checkpoint/global_step16324/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:faca1e311a8daac12fe43436ced3bf489902c39e393afcaf02c7abbef8f24360
|
3 |
+
size 43429616
|
last-checkpoint/global_step16324/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:36ab456294a7d399bad30d81580434a28f6c7386aaedd027de42b53aeb7f853a
|
3 |
+
size 43429616
|
last-checkpoint/global_step16324/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c7922d760d5b1a91c02ebc1013ba311be44ed60980d5477a0e251e4483539249
|
3 |
+
size 43429616
|
last-checkpoint/global_step16324/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6d857b446e3a6357b3b0df18eb42ff50f717dac378e49b32526d9bb5046c4f8a
|
3 |
+
size 43429616
|
last-checkpoint/global_step16324/zero_pp_rank_0_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:734487e0fe33955a906e0aa5485a809829d38b8e91a508a471fab410c8a16aa9
|
3 |
+
size 637299
|
last-checkpoint/global_step16324/zero_pp_rank_1_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f4464de89a444833e29b3f20fe88cc45c3dc57e65cc1aa037ff5265410f7d659
|
3 |
+
size 637171
|
last-checkpoint/global_step16324/zero_pp_rank_2_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c4b3acdcc17ad92ef9613e4eb74aae6343516021b167e2173523df0a2ec08df
|
3 |
+
size 637171
|
last-checkpoint/global_step16324/zero_pp_rank_3_mp_rank_00_model_states.pt
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:196792206c368bd09992379cd8104b1403f3697d59854c5e2a8b5d1f8f9e27a7
|
3 |
+
size 637171
|
last-checkpoint/latest
CHANGED
@@ -1 +1 @@
|
|
1 |
-
|
|
|
1 |
+
global_step16324
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7fd84006555d36e2389fd01b027fd18ebefe9c980286dac28012c182eb4d23e3
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 0.04092838987708092,
|
3 |
"best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
|
4 |
-
"epoch": 1.
|
5 |
"eval_steps": 50,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -29021,11 +29021,43 @@
|
|
29021 |
"eval_steps_per_second": 0.787,
|
29022 |
"num_input_tokens_seen": 109112120,
|
29023 |
"step": 16300
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
29024 |
}
|
29025 |
],
|
29026 |
"logging_steps": 5,
|
29027 |
"max_steps": 16324,
|
29028 |
-
"num_input_tokens_seen":
|
29029 |
"num_train_epochs": 2,
|
29030 |
"save_steps": 50,
|
29031 |
"stateful_callbacks": {
|
@@ -29035,12 +29067,12 @@
|
|
29035 |
"should_evaluate": false,
|
29036 |
"should_log": false,
|
29037 |
"should_save": true,
|
29038 |
-
"should_training_stop":
|
29039 |
},
|
29040 |
"attributes": {}
|
29041 |
}
|
29042 |
},
|
29043 |
-
"total_flos":
|
29044 |
"train_batch_size": 1,
|
29045 |
"trial_name": null,
|
29046 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 0.04092838987708092,
|
3 |
"best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
|
4 |
+
"epoch": 1.9999081192000245,
|
5 |
"eval_steps": 50,
|
6 |
+
"global_step": 16324,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
29021 |
"eval_steps_per_second": 0.787,
|
29022 |
"num_input_tokens_seen": 109112120,
|
29023 |
"step": 16300
|
29024 |
+
},
|
29025 |
+
{
|
29026 |
+
"epoch": 1.9975804722673118,
|
29027 |
+
"grad_norm": 2.005094271805205,
|
29028 |
+
"learning_rate": 3.7041694531114723e-10,
|
29029 |
+
"loss": 0.2376,
|
29030 |
+
"num_input_tokens_seen": 109145312,
|
29031 |
+
"step": 16305
|
29032 |
+
},
|
29033 |
+
{
|
29034 |
+
"epoch": 1.9981930109338153,
|
29035 |
+
"grad_norm": 1.4867493885349998,
|
29036 |
+
"learning_rate": 2.0111291483271288e-10,
|
29037 |
+
"loss": 0.2031,
|
29038 |
+
"num_input_tokens_seen": 109179312,
|
29039 |
+
"step": 16310
|
29040 |
+
},
|
29041 |
+
{
|
29042 |
+
"epoch": 1.9988055496003185,
|
29043 |
+
"grad_norm": 1.584898833540703,
|
29044 |
+
"learning_rate": 8.311302300278989e-11,
|
29045 |
+
"loss": 0.2135,
|
29046 |
+
"num_input_tokens_seen": 109212488,
|
29047 |
+
"step": 16315
|
29048 |
+
},
|
29049 |
+
{
|
29050 |
+
"epoch": 1.9994180882668218,
|
29051 |
+
"grad_norm": 1.9175958869363128,
|
29052 |
+
"learning_rate": 1.6417390907852438e-11,
|
29053 |
+
"loss": 0.2092,
|
29054 |
+
"num_input_tokens_seen": 109246064,
|
29055 |
+
"step": 16320
|
29056 |
}
|
29057 |
],
|
29058 |
"logging_steps": 5,
|
29059 |
"max_steps": 16324,
|
29060 |
+
"num_input_tokens_seen": 109273408,
|
29061 |
"num_train_epochs": 2,
|
29062 |
"save_steps": 50,
|
29063 |
"stateful_callbacks": {
|
|
|
29067 |
"should_evaluate": false,
|
29068 |
"should_log": false,
|
29069 |
"should_save": true,
|
29070 |
+
"should_training_stop": true
|
29071 |
},
|
29072 |
"attributes": {}
|
29073 |
}
|
29074 |
},
|
29075 |
+
"total_flos": 6853640571322368.0,
|
29076 |
"train_batch_size": 1,
|
29077 |
"trial_name": null,
|
29078 |
"trial_params": null
|