besimray commited on
Commit
a1cc42a
·
verified ·
1 Parent(s): a3b4e8f

Training in progress, step 160, checkpoint

Browse files
last-checkpoint/adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea24a06d4297cb7bc9a35aaa42d564a5666190bf2fe714ed3d34247746608f5f
3
  size 1579384
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:706e09d4bf9933ddb5fabda6b8c5f9d74d24c10bf155ccc13184acd93e4d1ab4
3
  size 1579384
last-checkpoint/optimizer.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1cf8d4457b46c5d8fa05b2f35ee8ddf4c1df71c1017622a65d418ecc7b2a0e9
3
  size 857274
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5766353332fb305f49bebe1e2e0337a74ff7d9408989efba803ce4256f7fe826
3
  size 857274
last-checkpoint/rng_state.pth CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:056a3c074142756e34db3a1d83ef56a2230b3ca87d2eb93ba1fc46523c99e84f
3
  size 14244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa5a5d1a33e5ca36cbafb461e09204198d05c7ff6494a37033d5d14176625399
3
  size 14244
last-checkpoint/scheduler.pt CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b8fa6a7a6709edeb55cdf23229934c07be2d8aae0e4056fbdb6ff2482d0eb3d3
3
  size 1064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ac2f052986474c47aa92ee09f81c986cf5cd3c01bde827e5d887e85b2bbda4c2
3
  size 1064
last-checkpoint/trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.2816241383552551,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-130",
4
- "epoch": 0.074064930255524,
5
  "eval_steps": 10,
6
- "global_step": 150,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -1185,6 +1185,84 @@
1185
  "eval_samples_per_second": 28.018,
1186
  "eval_steps_per_second": 28.018,
1187
  "step": 150
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1188
  }
1189
  ],
1190
  "logging_steps": 1,
@@ -1199,7 +1277,7 @@
1199
  "early_stopping_threshold": 0.0
1200
  },
1201
  "attributes": {
1202
- "early_stopping_patience_counter": 2
1203
  }
1204
  },
1205
  "TrainerControl": {
@@ -1208,12 +1286,12 @@
1208
  "should_evaluate": false,
1209
  "should_log": false,
1210
  "should_save": true,
1211
- "should_training_stop": false
1212
  },
1213
  "attributes": {}
1214
  }
1215
  },
1216
- "total_flos": 118289439129600.0,
1217
  "train_batch_size": 1,
1218
  "trial_name": null,
1219
  "trial_params": null
 
1
  {
2
  "best_metric": 0.2816241383552551,
3
  "best_model_checkpoint": "miner_id_24/checkpoint-130",
4
+ "epoch": 0.07900259227255894,
5
  "eval_steps": 10,
6
+ "global_step": 160,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
1185
  "eval_samples_per_second": 28.018,
1186
  "eval_steps_per_second": 28.018,
1187
  "step": 150
1188
+ },
1189
+ {
1190
+ "epoch": 0.0745586964572275,
1191
+ "grad_norm": 216.89869689941406,
1192
+ "learning_rate": 0.00016184643766056317,
1193
+ "loss": 1.9158,
1194
+ "step": 151
1195
+ },
1196
+ {
1197
+ "epoch": 0.075052462658931,
1198
+ "grad_norm": 2457.674072265625,
1199
+ "learning_rate": 0.00016134135287043669,
1200
+ "loss": 2.4605,
1201
+ "step": 152
1202
+ },
1203
+ {
1204
+ "epoch": 0.07554622886063449,
1205
+ "grad_norm": 1994.4697265625,
1206
+ "learning_rate": 0.00016083374657755134,
1207
+ "loss": 4.3665,
1208
+ "step": 153
1209
+ },
1210
+ {
1211
+ "epoch": 0.07603999506233798,
1212
+ "grad_norm": 812.1641845703125,
1213
+ "learning_rate": 0.00016032363964761363,
1214
+ "loss": 0.5252,
1215
+ "step": 154
1216
+ },
1217
+ {
1218
+ "epoch": 0.07653376126404147,
1219
+ "grad_norm": 978.52294921875,
1220
+ "learning_rate": 0.00015981105304912162,
1221
+ "loss": 2.1287,
1222
+ "step": 155
1223
+ },
1224
+ {
1225
+ "epoch": 0.07702752746574497,
1226
+ "grad_norm": 10419.509765625,
1227
+ "learning_rate": 0.00015929600785250257,
1228
+ "loss": 0.9229,
1229
+ "step": 156
1230
+ },
1231
+ {
1232
+ "epoch": 0.07752129366744846,
1233
+ "grad_norm": 1868.2894287109375,
1234
+ "learning_rate": 0.00015877852522924732,
1235
+ "loss": 2.6291,
1236
+ "step": 157
1237
+ },
1238
+ {
1239
+ "epoch": 0.07801505986915196,
1240
+ "grad_norm": 595.7179565429688,
1241
+ "learning_rate": 0.0001582586264510396,
1242
+ "loss": 0.3959,
1243
+ "step": 158
1244
+ },
1245
+ {
1246
+ "epoch": 0.07850882607085545,
1247
+ "grad_norm": 5246.302734375,
1248
+ "learning_rate": 0.00015773633288888197,
1249
+ "loss": 0.9382,
1250
+ "step": 159
1251
+ },
1252
+ {
1253
+ "epoch": 0.07900259227255894,
1254
+ "grad_norm": 2367.507080078125,
1255
+ "learning_rate": 0.00015721166601221698,
1256
+ "loss": 0.6265,
1257
+ "step": 160
1258
+ },
1259
+ {
1260
+ "epoch": 0.07900259227255894,
1261
+ "eval_loss": 0.3629942238330841,
1262
+ "eval_runtime": 15.208,
1263
+ "eval_samples_per_second": 28.077,
1264
+ "eval_steps_per_second": 28.077,
1265
+ "step": 160
1266
  }
1267
  ],
1268
  "logging_steps": 1,
 
1277
  "early_stopping_threshold": 0.0
1278
  },
1279
  "attributes": {
1280
+ "early_stopping_patience_counter": 3
1281
  }
1282
  },
1283
  "TrainerControl": {
 
1286
  "should_evaluate": false,
1287
  "should_log": false,
1288
  "should_save": true,
1289
+ "should_training_stop": true
1290
  },
1291
  "attributes": {}
1292
  }
1293
  },
1294
+ "total_flos": 126175401738240.0,
1295
  "train_batch_size": 1,
1296
  "trial_name": null,
1297
  "trial_params": null