Starred commited on
Commit
5a4cb0d
·
verified ·
1 Parent(s): 59d095e

Training in progress, step 500, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5fba4ce126571fed34ee25b1d50c97224679bff4e28f648cb5b533214ae0c65b
3
  size 84972248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0b7426f77545243fe44b61bdf327bbe8386aaecd03d9b377cde6bb8b482e8a3a
3
  size 84972248
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:861c7e64af22126d86807af652070b1a4fb616b03f2a0b5124b65afecf26f325
3
- size 43434149
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76fc2fa6250a96154629a5641458d43a79a6deaf4569bdd3d1051fc60b41bdab
3
+ size 43434405
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1b8adeed8ef1e4e1fb02821b08a58ce4cad29d44855cb22be2f764747fad931
3
  size 14917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27978b0c0a2d70a6d9e24fba0f2ed928bba9b47aecb6cc92523c33dd4a9bc29f
3
  size 14917
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d930f1f012f62572a539c80416eb7211c7e334068f916c4231ec05e7070b3a22
3
  size 14917
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d4883ea9e9e4170ef036b08f27772e01a500be9d13e3c9492d90b6357d7e8c6
3
  size 14917
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f5b5392788f8c9ab10debba0348bb95011a14915ac18e4f0cb3246c0d97d61a7
3
  size 1465
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46a5dde96232ac87d1fa1ed9715479941761512093c62e0c951b01e5c58be3a3
3
  size 1465
last-checkpoint/trainer_state.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "best_global_step": 250,
3
- "best_metric": 0.6187728047370911,
4
- "best_model_checkpoint": "/kaggle/working/obsidian_critic_qwen35_t4x2_unsloth/runs/obsidian_critic_full_epoch/checkpoint-250",
5
- "epoch": 0.11085859985588382,
6
  "eval_steps": 125,
7
- "global_step": 250,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
@@ -106,6 +106,92 @@
106
  "tokens_per_second": 363.7051920895653,
107
  "tokens_per_step": 1785.536,
108
  "total_tokens_seen": 446384
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
109
  }
110
  ],
111
  "logging_steps": 50,
@@ -134,7 +220,7 @@
134
  "attributes": {}
135
  }
136
  },
137
- "total_flos": 1.8027914150805504e+16,
138
  "train_batch_size": 1,
139
  "trial_name": null,
140
  "trial_params": null
 
1
  {
2
+ "best_global_step": 500,
3
+ "best_metric": 0.40328726172447205,
4
+ "best_model_checkpoint": "/kaggle/working/obsidian_critic_qwen35_t4x2_unsloth/runs/obsidian_critic_full_epoch/checkpoint-500",
5
+ "epoch": 0.22171719971176765,
6
  "eval_steps": 125,
7
+ "global_step": 500,
8
  "is_hyper_param_search": false,
9
  "is_local_process_zero": true,
10
  "is_world_process_zero": true,
 
106
  "tokens_per_second": 363.7051920895653,
107
  "tokens_per_step": 1785.536,
108
  "total_tokens_seen": 446384
109
+ },
110
+ {
111
+ "epoch": 0.1330303198270606,
112
+ "grad_norm": 0.5571497082710266,
113
+ "last_batch_tokens": 329,
114
+ "learning_rate": 9.584967947244769e-05,
115
+ "loss": 0.5449295806884765,
116
+ "lr": 9.582179859078793e-05,
117
+ "step": 300,
118
+ "tokens_per_second": 81.02823836816424,
119
+ "tokens_per_step": 1724.5733333333333,
120
+ "total_tokens_seen": 517372
121
+ },
122
+ {
123
+ "epoch": 0.15520203979823735,
124
+ "grad_norm": 0.7961392998695374,
125
+ "last_batch_tokens": 165,
126
+ "learning_rate": 9.434740857432105e-05,
127
+ "loss": 0.46938041687011717,
128
+ "lr": 9.431513518232342e-05,
129
+ "step": 350,
130
+ "tokens_per_second": 89.52354398651325,
131
+ "tokens_per_step": 1704.1371428571429,
132
+ "total_tokens_seen": 596448
133
+ },
134
+ {
135
+ "epoch": 0.16628789978382574,
136
+ "eval_loss": 0.4863806366920471,
137
+ "eval_runtime": 87.0251,
138
+ "eval_samples_per_second": 4.171,
139
+ "eval_steps_per_second": 2.091,
140
+ "last_batch_tokens": 172,
141
+ "lr": 9.348041345533653e-05,
142
+ "step": 375,
143
+ "tokens_per_second": 135.38091364115044,
144
+ "tokens_per_step": 1784.712,
145
+ "total_tokens_seen": 669267
146
+ },
147
+ {
148
+ "epoch": 0.1773737597694141,
149
+ "grad_norm": 0.7586395144462585,
150
+ "last_batch_tokens": 351,
151
+ "learning_rate": 9.262927340344295e-05,
152
+ "loss": 0.4675440216064453,
153
+ "lr": 9.259276459421655e-05,
154
+ "step": 400,
155
+ "tokens_per_second": 81.3096016563381,
156
+ "tokens_per_step": 1764.9875,
157
+ "total_tokens_seen": 705995
158
+ },
159
+ {
160
+ "epoch": 0.19954547974059086,
161
+ "grad_norm": 0.7313582897186279,
162
+ "last_batch_tokens": 369,
163
+ "learning_rate": 9.070363710911735e-05,
164
+ "loss": 0.3964078140258789,
165
+ "lr": 9.066307059197612e-05,
166
+ "step": 450,
167
+ "tokens_per_second": 87.86278133239196,
168
+ "tokens_per_step": 1744.9444444444443,
169
+ "total_tokens_seen": 785225
170
+ },
171
+ {
172
+ "epoch": 0.22171719971176765,
173
+ "grad_norm": 0.5969849228858948,
174
+ "last_batch_tokens": 193,
175
+ "learning_rate": 8.857987286762718e-05,
176
+ "loss": 0.3672472381591797,
177
+ "lr": 8.853544610307675e-05,
178
+ "step": 500,
179
+ "tokens_per_second": 87.74574317837812,
180
+ "tokens_per_step": 1729.026,
181
+ "total_tokens_seen": 864513
182
+ },
183
+ {
184
+ "epoch": 0.22171719971176765,
185
+ "eval_loss": 0.40328726172447205,
186
+ "eval_runtime": 87.1124,
187
+ "eval_samples_per_second": 4.167,
188
+ "eval_steps_per_second": 2.089,
189
+ "last_batch_tokens": 172,
190
+ "lr": 8.853544610307675e-05,
191
+ "step": 500,
192
+ "tokens_per_second": 368.8907701487212,
193
+ "tokens_per_step": 1793.302,
194
+ "total_tokens_seen": 896651
195
  }
196
  ],
197
  "logging_steps": 50,
 
220
  "attributes": {}
221
  }
222
  },
223
+ "total_flos": 3.66545497793495e+16,
224
  "train_batch_size": 1,
225
  "trial_name": null,
226
  "trial_params": null