Training in progress, epoch 74, checkpoint
Browse files
last-checkpoint/model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 166496880
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:421852674ee7e5fb884fedde823129400a2a77e2ce9fdcda7248dbe602295e87
|
3 |
size 166496880
|
last-checkpoint/optimizer.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 330495866
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:885256dbd271d4acdcdc5dd1fe21689d1142647c5cd71cec6d3533802ebff448
|
3 |
size 330495866
|
last-checkpoint/rng_state.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 14244
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6000c37f90dc3b850dac3042834ee9dddd756f53ca02164e22124efd3c87f850
|
3 |
size 14244
|
last-checkpoint/scheduler.pt
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:52801c60befa357564462eaa30eb8290d188ee442c2f8c95fd10a63dc3d264df
|
3 |
size 1064
|
last-checkpoint/trainer_state.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
"best_metric": 0.31184816360473633,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-35000",
|
4 |
-
"epoch":
|
5 |
"eval_steps": 500,
|
6 |
-
"global_step":
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
@@ -10417,6 +10417,151 @@
|
|
10417 |
"eval_samples_per_second": 7.119,
|
10418 |
"eval_steps_per_second": 0.925,
|
10419 |
"step": 36500
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
10420 |
}
|
10421 |
],
|
10422 |
"logging_steps": 30,
|
@@ -10436,7 +10581,7 @@
|
|
10436 |
"attributes": {}
|
10437 |
}
|
10438 |
},
|
10439 |
-
"total_flos": 1.
|
10440 |
"train_batch_size": 2,
|
10441 |
"trial_name": null,
|
10442 |
"trial_params": null
|
|
|
1 |
{
|
2 |
"best_metric": 0.31184816360473633,
|
3 |
"best_model_checkpoint": "chickens-composite-201616161616-150-epochs-wo-transform/checkpoint-35000",
|
4 |
+
"epoch": 74.0,
|
5 |
"eval_steps": 500,
|
6 |
+
"global_step": 37000,
|
7 |
"is_hyper_param_search": false,
|
8 |
"is_local_process_zero": true,
|
9 |
"is_world_process_zero": true,
|
|
|
10417 |
"eval_samples_per_second": 7.119,
|
10418 |
"eval_steps_per_second": 0.925,
|
10419 |
"step": 36500
|
10420 |
+
},
|
10421 |
+
{
|
10422 |
+
"epoch": 73.02,
|
10423 |
+
"grad_norm": 61.354408264160156,
|
10424 |
+
"learning_rate": 5.207285692376427e-06,
|
10425 |
+
"loss": 0.2531,
|
10426 |
+
"step": 36510
|
10427 |
+
},
|
10428 |
+
{
|
10429 |
+
"epoch": 73.08,
|
10430 |
+
"grad_norm": 135.92481994628906,
|
10431 |
+
"learning_rate": 5.201007746814767e-06,
|
10432 |
+
"loss": 0.2641,
|
10433 |
+
"step": 36540
|
10434 |
+
},
|
10435 |
+
{
|
10436 |
+
"epoch": 73.14,
|
10437 |
+
"grad_norm": 42.83466339111328,
|
10438 |
+
"learning_rate": 5.194729483834438e-06,
|
10439 |
+
"loss": 0.2872,
|
10440 |
+
"step": 36570
|
10441 |
+
},
|
10442 |
+
{
|
10443 |
+
"epoch": 73.2,
|
10444 |
+
"grad_norm": 34.50326156616211,
|
10445 |
+
"learning_rate": 5.188450913349674e-06,
|
10446 |
+
"loss": 0.2668,
|
10447 |
+
"step": 36600
|
10448 |
+
},
|
10449 |
+
{
|
10450 |
+
"epoch": 73.26,
|
10451 |
+
"grad_norm": 67.49690246582031,
|
10452 |
+
"learning_rate": 5.1821720452751945e-06,
|
10453 |
+
"loss": 0.2556,
|
10454 |
+
"step": 36630
|
10455 |
+
},
|
10456 |
+
{
|
10457 |
+
"epoch": 73.32,
|
10458 |
+
"grad_norm": 62.13235092163086,
|
10459 |
+
"learning_rate": 5.175892889526189e-06,
|
10460 |
+
"loss": 0.2524,
|
10461 |
+
"step": 36660
|
10462 |
+
},
|
10463 |
+
{
|
10464 |
+
"epoch": 73.38,
|
10465 |
+
"grad_norm": 39.14817428588867,
|
10466 |
+
"learning_rate": 5.1696134560183045e-06,
|
10467 |
+
"loss": 0.2292,
|
10468 |
+
"step": 36690
|
10469 |
+
},
|
10470 |
+
{
|
10471 |
+
"epoch": 73.44,
|
10472 |
+
"grad_norm": 62.339149475097656,
|
10473 |
+
"learning_rate": 5.16333375466762e-06,
|
10474 |
+
"loss": 0.2844,
|
10475 |
+
"step": 36720
|
10476 |
+
},
|
10477 |
+
{
|
10478 |
+
"epoch": 73.5,
|
10479 |
+
"grad_norm": 92.51763153076172,
|
10480 |
+
"learning_rate": 5.157053795390642e-06,
|
10481 |
+
"loss": 0.2974,
|
10482 |
+
"step": 36750
|
10483 |
+
},
|
10484 |
+
{
|
10485 |
+
"epoch": 73.56,
|
10486 |
+
"grad_norm": 57.300418853759766,
|
10487 |
+
"learning_rate": 5.150773588104284e-06,
|
10488 |
+
"loss": 0.2336,
|
10489 |
+
"step": 36780
|
10490 |
+
},
|
10491 |
+
{
|
10492 |
+
"epoch": 73.62,
|
10493 |
+
"grad_norm": 58.37376403808594,
|
10494 |
+
"learning_rate": 5.144493142725851e-06,
|
10495 |
+
"loss": 0.255,
|
10496 |
+
"step": 36810
|
10497 |
+
},
|
10498 |
+
{
|
10499 |
+
"epoch": 73.68,
|
10500 |
+
"grad_norm": 54.429656982421875,
|
10501 |
+
"learning_rate": 5.138212469173022e-06,
|
10502 |
+
"loss": 0.3182,
|
10503 |
+
"step": 36840
|
10504 |
+
},
|
10505 |
+
{
|
10506 |
+
"epoch": 73.74,
|
10507 |
+
"grad_norm": 142.04763793945312,
|
10508 |
+
"learning_rate": 5.13193157736384e-06,
|
10509 |
+
"loss": 0.2789,
|
10510 |
+
"step": 36870
|
10511 |
+
},
|
10512 |
+
{
|
10513 |
+
"epoch": 73.8,
|
10514 |
+
"grad_norm": 66.44054412841797,
|
10515 |
+
"learning_rate": 5.1256504772166885e-06,
|
10516 |
+
"loss": 0.2505,
|
10517 |
+
"step": 36900
|
10518 |
+
},
|
10519 |
+
{
|
10520 |
+
"epoch": 73.86,
|
10521 |
+
"grad_norm": 67.2595443725586,
|
10522 |
+
"learning_rate": 5.119369178650282e-06,
|
10523 |
+
"loss": 0.2399,
|
10524 |
+
"step": 36930
|
10525 |
+
},
|
10526 |
+
{
|
10527 |
+
"epoch": 73.92,
|
10528 |
+
"grad_norm": 67.150146484375,
|
10529 |
+
"learning_rate": 5.1130876915836495e-06,
|
10530 |
+
"loss": 0.2854,
|
10531 |
+
"step": 36960
|
10532 |
+
},
|
10533 |
+
{
|
10534 |
+
"epoch": 73.98,
|
10535 |
+
"grad_norm": 74.2684326171875,
|
10536 |
+
"learning_rate": 5.1068060259361155e-06,
|
10537 |
+
"loss": 0.2582,
|
10538 |
+
"step": 36990
|
10539 |
+
},
|
10540 |
+
{
|
10541 |
+
"epoch": 74.0,
|
10542 |
+
"eval_loss": 0.31870606541633606,
|
10543 |
+
"eval_map": 0.7877,
|
10544 |
+
"eval_map_50": 0.9519,
|
10545 |
+
"eval_map_75": 0.9094,
|
10546 |
+
"eval_map_chicken": 0.7903,
|
10547 |
+
"eval_map_duck": 0.7278,
|
10548 |
+
"eval_map_large": 0.796,
|
10549 |
+
"eval_map_medium": 0.7885,
|
10550 |
+
"eval_map_plant": 0.8449,
|
10551 |
+
"eval_map_small": 0.3513,
|
10552 |
+
"eval_mar_1": 0.3134,
|
10553 |
+
"eval_mar_10": 0.8268,
|
10554 |
+
"eval_mar_100": 0.829,
|
10555 |
+
"eval_mar_100_chicken": 0.8341,
|
10556 |
+
"eval_mar_100_duck": 0.7742,
|
10557 |
+
"eval_mar_100_plant": 0.8788,
|
10558 |
+
"eval_mar_large": 0.8335,
|
10559 |
+
"eval_mar_medium": 0.8315,
|
10560 |
+
"eval_mar_small": 0.3976,
|
10561 |
+
"eval_runtime": 14.3727,
|
10562 |
+
"eval_samples_per_second": 6.958,
|
10563 |
+
"eval_steps_per_second": 0.904,
|
10564 |
+
"step": 37000
|
10565 |
}
|
10566 |
],
|
10567 |
"logging_steps": 30,
|
|
|
10581 |
"attributes": {}
|
10582 |
}
|
10583 |
},
|
10584 |
+
"total_flos": 1.27288551776256e+19,
|
10585 |
"train_batch_size": 2,
|
10586 |
"trial_name": null,
|
10587 |
"trial_params": null
|