masatochi commited on
Commit
8ee7fdc
1 Parent(s): df0f6e5

Training in progress, step 55, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:adac11fe9b13bf1b76ad183dcdd7e81553c07712d98750b3a009b0fde4a8f61b
3
  size 59827904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1cd5aeb2351bf3f7f393d5da945d90195388ecf503901c6048484c28d16beed
3
  size 59827904
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f08485c8ba652d542301fa3d085478be8bed85616b2bcb26abf83e95f4faea3
3
  size 30875540
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:32143bc13d61a27624fdfea7555a00420c4d16ddde6073b23f26708af84d89fc
3
  size 30875540
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:887c396fe906e451edd5dc1fa4bc857e822a44e9109e8deb2d1acfa642c3f206
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a595415dd381301168ca7a830d46eddf8b42b4c9fd8235b4b7e89468bc79b8d1
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c6c6ef871b14e9c2b6e6517244b08d50c3b172209dbf2c7f2f35e5b70f657975
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:61487aeef4449a4dec50f3ae9ec76bca52908878863009a40746a21c237f51ad
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 0.024451372333272204,
5
  "eval_steps": 34,
6
- "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -373,6 +373,41 @@
373
  "learning_rate": 0.00019324722294043558,
374
  "loss": 10.5181,
375
  "step": 50
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
376
  }
377
  ],
378
  "logging_steps": 1,
@@ -392,7 +427,7 @@
392
  "attributes": {}
393
  }
394
  },
395
- "total_flos": 1.102239837978624e+17,
396
  "train_batch_size": 3,
397
  "trial_name": null,
398
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 0.026896509566599426,
5
  "eval_steps": 34,
6
+ "global_step": 55,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
373
  "learning_rate": 0.00019324722294043558,
374
  "loss": 10.5181,
375
  "step": 50
376
+ },
377
+ {
378
+ "epoch": 0.024940399779937648,
379
+ "grad_norm": Infinity,
380
+ "learning_rate": 0.00019256376597815564,
381
+ "loss": 9.8439,
382
+ "step": 51
383
+ },
384
+ {
385
+ "epoch": 0.02542942722660309,
386
+ "grad_norm": Infinity,
387
+ "learning_rate": 0.00019184869857459232,
388
+ "loss": 9.1904,
389
+ "step": 52
390
+ },
391
+ {
392
+ "epoch": 0.02591845467326854,
393
+ "grad_norm": Infinity,
394
+ "learning_rate": 0.00019110226492460885,
395
+ "loss": 8.9795,
396
+ "step": 53
397
+ },
398
+ {
399
+ "epoch": 0.026407482119933982,
400
+ "grad_norm": 1.3233301388714312e+18,
401
+ "learning_rate": 0.0001903247199346129,
402
+ "loss": 8.6349,
403
+ "step": 54
404
+ },
405
+ {
406
+ "epoch": 0.026896509566599426,
407
+ "grad_norm": Infinity,
408
+ "learning_rate": 0.00018951632913550626,
409
+ "loss": 9.4657,
410
+ "step": 55
411
  }
412
  ],
413
  "logging_steps": 1,
 
427
  "attributes": {}
428
  }
429
  },
430
+ "total_flos": 1.2124638217764864e+17,
431
  "train_batch_size": 3,
432
  "trial_name": null,
433
  "trial_params": null