ben81828 commited on
Commit
74b1d76
1 Parent(s): fd3b7e8

Training in progress, step 16324, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e159368efe95459cbd6040fa012982b90b37b01e930578243495dac68e165708
3
  size 29034840
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6dd27678e81e52b77b203ee21948888d64924194d564a7da5da8608977a649ef
3
  size 29034840
last-checkpoint/global_step16324/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:faca1e311a8daac12fe43436ced3bf489902c39e393afcaf02c7abbef8f24360
3
+ size 43429616
last-checkpoint/global_step16324/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36ab456294a7d399bad30d81580434a28f6c7386aaedd027de42b53aeb7f853a
3
+ size 43429616
last-checkpoint/global_step16324/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c7922d760d5b1a91c02ebc1013ba311be44ed60980d5477a0e251e4483539249
3
+ size 43429616
last-checkpoint/global_step16324/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d857b446e3a6357b3b0df18eb42ff50f717dac378e49b32526d9bb5046c4f8a
3
+ size 43429616
last-checkpoint/global_step16324/zero_pp_rank_0_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:734487e0fe33955a906e0aa5485a809829d38b8e91a508a471fab410c8a16aa9
3
+ size 637299
last-checkpoint/global_step16324/zero_pp_rank_1_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4464de89a444833e29b3f20fe88cc45c3dc57e65cc1aa037ff5265410f7d659
3
+ size 637171
last-checkpoint/global_step16324/zero_pp_rank_2_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c4b3acdcc17ad92ef9613e4eb74aae6343516021b167e2173523df0a2ec08df
3
+ size 637171
last-checkpoint/global_step16324/zero_pp_rank_3_mp_rank_00_model_states.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:196792206c368bd09992379cd8104b1403f3697d59854c5e2a8b5d1f8f9e27a7
3
+ size 637171
last-checkpoint/latest CHANGED
@@ -1 +1 @@
1
- global_step16300
 
1
+ global_step16324
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18f0da524ba3c270b44cd4990f1f30630c32c971dc011d598daeb646e4af0dcb
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fd84006555d36e2389fd01b027fd18ebefe9c980286dac28012c182eb4d23e3
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.04092838987708092,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
4
- "epoch": 1.9969679336008086,
5
  "eval_steps": 50,
6
- "global_step": 16300,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -29021,11 +29021,43 @@
29021
  "eval_steps_per_second": 0.787,
29022
  "num_input_tokens_seen": 109112120,
29023
  "step": 16300
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
29024
  }
29025
  ],
29026
  "logging_steps": 5,
29027
  "max_steps": 16324,
29028
- "num_input_tokens_seen": 109112120,
29029
  "num_train_epochs": 2,
29030
  "save_steps": 50,
29031
  "stateful_callbacks": {
@@ -29035,12 +29067,12 @@
29035
  "should_evaluate": false,
29036
  "should_log": false,
29037
  "should_save": true,
29038
- "should_training_stop": false
29039
  },
29040
  "attributes": {}
29041
  }
29042
  },
29043
- "total_flos": 6843538279890944.0,
29044
  "train_batch_size": 1,
29045
  "trial_name": null,
29046
  "trial_params": null
 
1
  {
2
  "best_metric": 0.04092838987708092,
3
  "best_model_checkpoint": "saves/CADICA_qwenvl_detect_classify_augmented/lora/sft/checkpoint-9050",
4
+ "epoch": 1.9999081192000245,
5
  "eval_steps": 50,
6
+ "global_step": 16324,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
29021
  "eval_steps_per_second": 0.787,
29022
  "num_input_tokens_seen": 109112120,
29023
  "step": 16300
29024
+ },
29025
+ {
29026
+ "epoch": 1.9975804722673118,
29027
+ "grad_norm": 2.005094271805205,
29028
+ "learning_rate": 3.7041694531114723e-10,
29029
+ "loss": 0.2376,
29030
+ "num_input_tokens_seen": 109145312,
29031
+ "step": 16305
29032
+ },
29033
+ {
29034
+ "epoch": 1.9981930109338153,
29035
+ "grad_norm": 1.4867493885349998,
29036
+ "learning_rate": 2.0111291483271288e-10,
29037
+ "loss": 0.2031,
29038
+ "num_input_tokens_seen": 109179312,
29039
+ "step": 16310
29040
+ },
29041
+ {
29042
+ "epoch": 1.9988055496003185,
29043
+ "grad_norm": 1.584898833540703,
29044
+ "learning_rate": 8.311302300278989e-11,
29045
+ "loss": 0.2135,
29046
+ "num_input_tokens_seen": 109212488,
29047
+ "step": 16315
29048
+ },
29049
+ {
29050
+ "epoch": 1.9994180882668218,
29051
+ "grad_norm": 1.9175958869363128,
29052
+ "learning_rate": 1.6417390907852438e-11,
29053
+ "loss": 0.2092,
29054
+ "num_input_tokens_seen": 109246064,
29055
+ "step": 16320
29056
  }
29057
  ],
29058
  "logging_steps": 5,
29059
  "max_steps": 16324,
29060
+ "num_input_tokens_seen": 109273408,
29061
  "num_train_epochs": 2,
29062
  "save_steps": 50,
29063
  "stateful_callbacks": {
 
29067
  "should_evaluate": false,
29068
  "should_log": false,
29069
  "should_save": true,
29070
+ "should_training_stop": true
29071
  },
29072
  "attributes": {}
29073
  }
29074
  },
29075
+ "total_flos": 6853640571322368.0,
29076
  "train_batch_size": 1,
29077
  "trial_name": null,
29078
  "trial_params": null