fats-fme commited on
Commit
08ab164
1 Parent(s): 3db7e4c

Training in progress, step 48, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:16a0fa8ff10e2b47c91f216fca7ae09ce7d67db73074f8e58c5084d72a1497e5
3
  size 63592
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dc6f62f3a4fb689c7ba8051e72971e6b48b8d501ffe4ec2fa67f6313b0ba71ec
3
  size 63592
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:68c036e462fc01b009e149797ee73941675bd15cf7f6ca35412a8358dceba4e3
3
  size 136814
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bf715e2a13c84448a3ff9e9018304d4beccc087840b59ab29bbea95f48c66616
3
  size 136814
last-checkpoint/rng_state_0.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6f435e2cf5af084ff87f38ac364aa32f5f9250d67a5fd3cb62ee9bb2e9e48da7
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f0769ac9f6e5a775e3172601984970ae5aeb6570d2e59f38a58d84854e22367
3
  size 14512
last-checkpoint/rng_state_1.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dfdc610d6111b3c5b2c5f58f7b4bc723b1377af735ffdde336defa263ece09d
3
  size 14512
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4b69f085dcd7d3c6e07acfbd354d973bf19c16d6283ab3f9fe7eafe320fa66a1
3
  size 14512
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5a4ecc45073a35262869a0d9392090bbde50163f9775b63db54f8daf68eadbe7
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2997d603631ca8d1209b6aeca1ac3d249bd00d50d014e5ffdd28c52cc649ef27
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.512,
5
  "eval_steps": 16,
6
- "global_step": 32,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -255,6 +255,126 @@
255
  "eval_samples_per_second": 131.72,
256
  "eval_steps_per_second": 33.551,
257
  "step": 32
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
258
  }
259
  ],
260
  "logging_steps": 1,
@@ -274,7 +394,7 @@
274
  "attributes": {}
275
  }
276
  },
277
- "total_flos": 958314577920.0,
278
  "train_batch_size": 2,
279
  "trial_name": null,
280
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.768,
5
  "eval_steps": 16,
6
+ "global_step": 48,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
255
  "eval_samples_per_second": 131.72,
256
  "eval_steps_per_second": 33.551,
257
  "step": 32
258
+ },
259
+ {
260
+ "epoch": 0.528,
261
+ "grad_norm": 1.6445367336273193,
262
+ "learning_rate": 0.0001563320058063622,
263
+ "loss": 88.6368,
264
+ "step": 33
265
+ },
266
+ {
267
+ "epoch": 0.544,
268
+ "grad_norm": 1.6764920949935913,
269
+ "learning_rate": 0.00015000000000000001,
270
+ "loss": 88.6352,
271
+ "step": 34
272
+ },
273
+ {
274
+ "epoch": 0.56,
275
+ "grad_norm": 1.6223012208938599,
276
+ "learning_rate": 0.00014338837391175582,
277
+ "loss": 88.6562,
278
+ "step": 35
279
+ },
280
+ {
281
+ "epoch": 0.576,
282
+ "grad_norm": 1.6213066577911377,
283
+ "learning_rate": 0.00013653410243663952,
284
+ "loss": 88.6551,
285
+ "step": 36
286
+ },
287
+ {
288
+ "epoch": 0.592,
289
+ "grad_norm": 1.6623308658599854,
290
+ "learning_rate": 0.00012947551744109043,
291
+ "loss": 88.6895,
292
+ "step": 37
293
+ },
294
+ {
295
+ "epoch": 0.608,
296
+ "grad_norm": 1.8002474308013916,
297
+ "learning_rate": 0.00012225209339563145,
298
+ "loss": 88.674,
299
+ "step": 38
300
+ },
301
+ {
302
+ "epoch": 0.624,
303
+ "grad_norm": 1.8450833559036255,
304
+ "learning_rate": 0.00011490422661761744,
305
+ "loss": 88.5731,
306
+ "step": 39
307
+ },
308
+ {
309
+ "epoch": 0.64,
310
+ "grad_norm": 1.9053994417190552,
311
+ "learning_rate": 0.00010747300935864243,
312
+ "loss": 88.6383,
313
+ "step": 40
314
+ },
315
+ {
316
+ "epoch": 0.656,
317
+ "grad_norm": 1.9040968418121338,
318
+ "learning_rate": 0.0001,
319
+ "loss": 88.6602,
320
+ "step": 41
321
+ },
322
+ {
323
+ "epoch": 0.672,
324
+ "grad_norm": 2.163938522338867,
325
+ "learning_rate": 9.252699064135758e-05,
326
+ "loss": 88.5704,
327
+ "step": 42
328
+ },
329
+ {
330
+ "epoch": 0.688,
331
+ "grad_norm": 2.156372547149658,
332
+ "learning_rate": 8.509577338238255e-05,
333
+ "loss": 88.4837,
334
+ "step": 43
335
+ },
336
+ {
337
+ "epoch": 0.704,
338
+ "grad_norm": 2.1990439891815186,
339
+ "learning_rate": 7.774790660436858e-05,
340
+ "loss": 88.6662,
341
+ "step": 44
342
+ },
343
+ {
344
+ "epoch": 0.72,
345
+ "grad_norm": 2.474072217941284,
346
+ "learning_rate": 7.052448255890957e-05,
347
+ "loss": 88.5809,
348
+ "step": 45
349
+ },
350
+ {
351
+ "epoch": 0.736,
352
+ "grad_norm": 2.7647547721862793,
353
+ "learning_rate": 6.34658975633605e-05,
354
+ "loss": 88.6076,
355
+ "step": 46
356
+ },
357
+ {
358
+ "epoch": 0.752,
359
+ "grad_norm": 2.5142226219177246,
360
+ "learning_rate": 5.6611626088244194e-05,
361
+ "loss": 88.7863,
362
+ "step": 47
363
+ },
364
+ {
365
+ "epoch": 0.768,
366
+ "grad_norm": 1.5332900285720825,
367
+ "learning_rate": 5.000000000000002e-05,
368
+ "loss": 88.6006,
369
+ "step": 48
370
+ },
371
+ {
372
+ "epoch": 0.768,
373
+ "eval_loss": 11.07020378112793,
374
+ "eval_runtime": 0.612,
375
+ "eval_samples_per_second": 173.197,
376
+ "eval_steps_per_second": 44.116,
377
+ "step": 48
378
  }
379
  ],
380
  "logging_steps": 1,
 
394
  "attributes": {}
395
  }
396
  },
397
+ "total_flos": 1437471866880.0,
398
  "train_batch_size": 2,
399
  "trial_name": null,
400
  "trial_params": null