PEFT
Safetensors
Generated from Trainer
paulrichmond commited on
Commit
d655747
·
verified ·
1 Parent(s): 81996fe

Training in progress, step 18444, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:efbdd287397a30ac9703304868f153af7eaa1f55050da0ca359fb8ab1568a5be
3
  size 18899856
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f62f3f80bdbe4ec415de0be3ae4ac3a7be429dba42b12f2a5da98039040e1c8
3
  size 18899856
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5c785e7b3916fb588f1e30bf4f86e8bd20b41f7798731e7d8eeb9de48a7c5d91
3
  size 37911546
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bae5a9fac15bbc68944eb74383a2a1c3c77311ab6098f221a8b3e3a99c150606
3
  size 37911546
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f4f77c1509e996519a6483f61bf0c10948dac58bd7f9064c1970977169b2ba1
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a8530fcb35807d7008ff35c5159c8264cd9ca963aa9eb088c64b4bb159785bd
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:82ab3f4c71b15efcf419ae699bcc09c12d9d69fda628dcac52fdb8059760dd34
3
  size 1000
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:112b0e3f3850ccaa762f9eb894c6c56a98cff86a4eab4f8d883ff8a96dbc8748
3
  size 1000
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 3.9973975276512688,
5
  "eval_steps": 2048,
6
- "global_step": 18432,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -129211,6 +129211,90 @@
129211
  "eval_test_samples_per_second": 14.923,
129212
  "eval_test_steps_per_second": 0.933,
129213
  "step": 18432
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
129214
  }
129215
  ],
129216
  "logging_steps": 1,
@@ -129225,12 +129309,12 @@
129225
  "should_evaluate": false,
129226
  "should_log": false,
129227
  "should_save": true,
129228
- "should_training_stop": false
129229
  },
129230
  "attributes": {}
129231
  }
129232
  },
129233
- "total_flos": 4.350735014776799e+18,
129234
  "train_batch_size": 16,
129235
  "trial_name": null,
129236
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 4.0,
5
  "eval_steps": 2048,
6
+ "global_step": 18444,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
129211
  "eval_test_samples_per_second": 14.923,
129212
  "eval_test_steps_per_second": 0.933,
129213
  "step": 18432
129214
+ },
129215
+ {
129216
+ "epoch": 3.9976144003469964,
129217
+ "grad_norm": 0.8142043948173523,
129218
+ "learning_rate": 3.0003218233616098e-06,
129219
+ "loss": 2.0155,
129220
+ "step": 18433
129221
+ },
129222
+ {
129223
+ "epoch": 3.997831273042724,
129224
+ "grad_norm": 0.721076488494873,
129225
+ "learning_rate": 3.000265969737023e-06,
129226
+ "loss": 2.0117,
129227
+ "step": 18434
129228
+ },
129229
+ {
129230
+ "epoch": 3.9980481457384514,
129231
+ "grad_norm": 0.7382825613021851,
129232
+ "learning_rate": 3.000215435499207e-06,
129233
+ "loss": 2.0278,
129234
+ "step": 18435
129235
+ },
129236
+ {
129237
+ "epoch": 3.998265018434179,
129238
+ "grad_norm": 0.7999047040939331,
129239
+ "learning_rate": 3.000170220649976e-06,
129240
+ "loss": 2.1662,
129241
+ "step": 18436
129242
+ },
129243
+ {
129244
+ "epoch": 3.9984818911299067,
129245
+ "grad_norm": 0.735248327255249,
129246
+ "learning_rate": 3.000130325190978e-06,
129247
+ "loss": 2.172,
129248
+ "step": 18437
129249
+ },
129250
+ {
129251
+ "epoch": 3.9986987638256344,
129252
+ "grad_norm": 0.6671957969665527,
129253
+ "learning_rate": 3.0000957491236143e-06,
129254
+ "loss": 2.1481,
129255
+ "step": 18438
129256
+ },
129257
+ {
129258
+ "epoch": 3.998915636521362,
129259
+ "grad_norm": 0.9371853470802307,
129260
+ "learning_rate": 3.000066492449139e-06,
129261
+ "loss": 2.0251,
129262
+ "step": 18439
129263
+ },
129264
+ {
129265
+ "epoch": 3.9991325092170893,
129266
+ "grad_norm": 0.696266233921051,
129267
+ "learning_rate": 3.00004255516859e-06,
129268
+ "loss": 1.9831,
129269
+ "step": 18440
129270
+ },
129271
+ {
129272
+ "epoch": 3.9993493819128174,
129273
+ "grad_norm": 0.6927635669708252,
129274
+ "learning_rate": 3.0000239372828404e-06,
129275
+ "loss": 2.0868,
129276
+ "step": 18441
129277
+ },
129278
+ {
129279
+ "epoch": 3.9995662546085446,
129280
+ "grad_norm": 0.7484452724456787,
129281
+ "learning_rate": 3.0000106387925348e-06,
129282
+ "loss": 2.0038,
129283
+ "step": 18442
129284
+ },
129285
+ {
129286
+ "epoch": 3.9997831273042723,
129287
+ "grad_norm": 0.7934520244598389,
129288
+ "learning_rate": 3.00000265969815e-06,
129289
+ "loss": 2.1349,
129290
+ "step": 18443
129291
+ },
129292
+ {
129293
+ "epoch": 4.0,
129294
+ "grad_norm": 1.1140004396438599,
129295
+ "learning_rate": 2.9999999999999997e-06,
129296
+ "loss": 2.1212,
129297
+ "step": 18444
129298
  }
129299
  ],
129300
  "logging_steps": 1,
 
129309
  "should_evaluate": false,
129310
  "should_log": false,
129311
  "should_save": true,
129312
+ "should_training_stop": true
129313
  },
129314
  "attributes": {}
129315
  }
129316
  },
129317
+ "total_flos": 4.3534139748762255e+18,
129318
  "train_batch_size": 16,
129319
  "trial_name": null,
129320
  "trial_params": null