|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 126.66666666666667, |
|
"global_step": 3800, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 1.67, |
|
"learning_rate": 4.9542483660130725e-05, |
|
"loss": 4.085, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 3.33, |
|
"learning_rate": 4.888888888888889e-05, |
|
"loss": 2.4616, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 5.0, |
|
"learning_rate": 4.823529411764706e-05, |
|
"loss": 1.4793, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 6.67, |
|
"learning_rate": 4.7581699346405226e-05, |
|
"loss": 0.7873, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 8.33, |
|
"learning_rate": 4.69281045751634e-05, |
|
"loss": 0.4859, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 10.0, |
|
"learning_rate": 4.6274509803921576e-05, |
|
"loss": 0.3261, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 11.67, |
|
"learning_rate": 4.562091503267974e-05, |
|
"loss": 0.2257, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 13.33, |
|
"learning_rate": 4.496732026143791e-05, |
|
"loss": 0.1711, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 15.0, |
|
"learning_rate": 4.431372549019608e-05, |
|
"loss": 0.126, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 16.67, |
|
"learning_rate": 4.366013071895425e-05, |
|
"loss": 0.0819, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 18.33, |
|
"learning_rate": 4.300653594771242e-05, |
|
"loss": 0.0699, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 20.0, |
|
"learning_rate": 4.235294117647059e-05, |
|
"loss": 0.0568, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 21.67, |
|
"learning_rate": 4.169934640522876e-05, |
|
"loss": 0.0473, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 23.33, |
|
"learning_rate": 4.104575163398693e-05, |
|
"loss": 0.0428, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 25.0, |
|
"learning_rate": 4.03921568627451e-05, |
|
"loss": 0.0382, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 26.67, |
|
"learning_rate": 3.973856209150327e-05, |
|
"loss": 0.033, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 28.33, |
|
"learning_rate": 3.908496732026144e-05, |
|
"loss": 0.0306, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 30.0, |
|
"learning_rate": 3.8431372549019614e-05, |
|
"loss": 0.0272, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 31.67, |
|
"learning_rate": 3.777777777777778e-05, |
|
"loss": 0.0225, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 33.33, |
|
"learning_rate": 3.712418300653595e-05, |
|
"loss": 0.0211, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 35.0, |
|
"learning_rate": 3.6470588235294114e-05, |
|
"loss": 0.0196, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 36.67, |
|
"learning_rate": 3.581699346405229e-05, |
|
"loss": 0.0167, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 38.33, |
|
"learning_rate": 3.5163398692810464e-05, |
|
"loss": 0.0158, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 40.0, |
|
"learning_rate": 3.450980392156863e-05, |
|
"loss": 0.0168, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 41.67, |
|
"learning_rate": 3.38562091503268e-05, |
|
"loss": 0.0101, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 43.33, |
|
"learning_rate": 3.3202614379084965e-05, |
|
"loss": 0.0128, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 45.0, |
|
"learning_rate": 3.254901960784314e-05, |
|
"loss": 0.0077, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 46.67, |
|
"learning_rate": 3.189542483660131e-05, |
|
"loss": 0.0065, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 48.33, |
|
"learning_rate": 3.124183006535948e-05, |
|
"loss": 0.0053, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 50.0, |
|
"learning_rate": 3.058823529411765e-05, |
|
"loss": 0.0064, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 51.67, |
|
"learning_rate": 2.9934640522875816e-05, |
|
"loss": 0.0059, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 53.33, |
|
"learning_rate": 2.928104575163399e-05, |
|
"loss": 0.0044, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 55.0, |
|
"learning_rate": 2.8627450980392155e-05, |
|
"loss": 0.0057, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 56.67, |
|
"learning_rate": 2.7973856209150327e-05, |
|
"loss": 0.0043, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 58.33, |
|
"learning_rate": 2.7320261437908495e-05, |
|
"loss": 0.0042, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 60.0, |
|
"learning_rate": 2.6666666666666667e-05, |
|
"loss": 0.0024, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 61.67, |
|
"learning_rate": 2.601307189542484e-05, |
|
"loss": 0.003, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 63.33, |
|
"learning_rate": 2.5359477124183006e-05, |
|
"loss": 0.0037, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 65.0, |
|
"learning_rate": 2.4705882352941178e-05, |
|
"loss": 0.0035, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 66.67, |
|
"learning_rate": 2.405228758169935e-05, |
|
"loss": 0.0019, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 68.33, |
|
"learning_rate": 2.3398692810457517e-05, |
|
"loss": 0.0027, |
|
"step": 2050 |
|
}, |
|
{ |
|
"epoch": 70.0, |
|
"learning_rate": 2.2745098039215685e-05, |
|
"loss": 0.003, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 71.67, |
|
"learning_rate": 2.2091503267973857e-05, |
|
"loss": 0.0028, |
|
"step": 2150 |
|
}, |
|
{ |
|
"epoch": 73.33, |
|
"learning_rate": 2.1437908496732025e-05, |
|
"loss": 0.0014, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 75.0, |
|
"learning_rate": 2.0784313725490197e-05, |
|
"loss": 0.0011, |
|
"step": 2250 |
|
}, |
|
{ |
|
"epoch": 76.67, |
|
"learning_rate": 2.0130718954248368e-05, |
|
"loss": 0.0017, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 78.33, |
|
"learning_rate": 1.9477124183006536e-05, |
|
"loss": 0.0007, |
|
"step": 2350 |
|
}, |
|
{ |
|
"epoch": 80.0, |
|
"learning_rate": 1.8823529411764708e-05, |
|
"loss": 0.0006, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 81.67, |
|
"learning_rate": 1.8169934640522876e-05, |
|
"loss": 0.0005, |
|
"step": 2450 |
|
}, |
|
{ |
|
"epoch": 83.33, |
|
"learning_rate": 1.7516339869281044e-05, |
|
"loss": 0.0005, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 85.0, |
|
"learning_rate": 1.686274509803922e-05, |
|
"loss": 0.0005, |
|
"step": 2550 |
|
}, |
|
{ |
|
"epoch": 86.67, |
|
"learning_rate": 1.6209150326797387e-05, |
|
"loss": 0.0005, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 88.33, |
|
"learning_rate": 1.5555555555555555e-05, |
|
"loss": 0.0005, |
|
"step": 2650 |
|
}, |
|
{ |
|
"epoch": 90.0, |
|
"learning_rate": 1.4901960784313726e-05, |
|
"loss": 0.0006, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 91.67, |
|
"learning_rate": 1.4248366013071896e-05, |
|
"loss": 0.0017, |
|
"step": 2750 |
|
}, |
|
{ |
|
"epoch": 93.33, |
|
"learning_rate": 1.3594771241830064e-05, |
|
"loss": 0.0009, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 95.0, |
|
"learning_rate": 1.2941176470588238e-05, |
|
"loss": 0.0008, |
|
"step": 2850 |
|
}, |
|
{ |
|
"epoch": 96.67, |
|
"learning_rate": 1.2287581699346406e-05, |
|
"loss": 0.0005, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 98.33, |
|
"learning_rate": 1.1633986928104575e-05, |
|
"loss": 0.0004, |
|
"step": 2950 |
|
}, |
|
{ |
|
"epoch": 100.0, |
|
"learning_rate": 1.0980392156862745e-05, |
|
"loss": 0.0005, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 101.67, |
|
"learning_rate": 1.0326797385620915e-05, |
|
"loss": 0.0004, |
|
"step": 3050 |
|
}, |
|
{ |
|
"epoch": 103.33, |
|
"learning_rate": 9.673202614379087e-06, |
|
"loss": 0.0004, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 105.0, |
|
"learning_rate": 9.019607843137255e-06, |
|
"loss": 0.0004, |
|
"step": 3150 |
|
}, |
|
{ |
|
"epoch": 106.67, |
|
"learning_rate": 8.366013071895424e-06, |
|
"loss": 0.0004, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 108.33, |
|
"learning_rate": 7.712418300653596e-06, |
|
"loss": 0.0004, |
|
"step": 3250 |
|
}, |
|
{ |
|
"epoch": 110.0, |
|
"learning_rate": 7.058823529411765e-06, |
|
"loss": 0.0003, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 111.67, |
|
"learning_rate": 6.405228758169935e-06, |
|
"loss": 0.0003, |
|
"step": 3350 |
|
}, |
|
{ |
|
"epoch": 113.33, |
|
"learning_rate": 5.7516339869281045e-06, |
|
"loss": 0.0003, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 115.0, |
|
"learning_rate": 5.098039215686274e-06, |
|
"loss": 0.0004, |
|
"step": 3450 |
|
}, |
|
{ |
|
"epoch": 116.67, |
|
"learning_rate": 4.444444444444445e-06, |
|
"loss": 0.0004, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 118.33, |
|
"learning_rate": 3.7908496732026144e-06, |
|
"loss": 0.0003, |
|
"step": 3550 |
|
}, |
|
{ |
|
"epoch": 120.0, |
|
"learning_rate": 3.1372549019607846e-06, |
|
"loss": 0.0003, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 121.67, |
|
"learning_rate": 2.4836601307189544e-06, |
|
"loss": 0.0003, |
|
"step": 3650 |
|
}, |
|
{ |
|
"epoch": 123.33, |
|
"learning_rate": 1.8300653594771244e-06, |
|
"loss": 0.0003, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 125.0, |
|
"learning_rate": 1.1764705882352942e-06, |
|
"loss": 0.0003, |
|
"step": 3750 |
|
}, |
|
{ |
|
"epoch": 126.67, |
|
"learning_rate": 5.22875816993464e-07, |
|
"loss": 0.0003, |
|
"step": 3800 |
|
} |
|
], |
|
"max_steps": 3840, |
|
"num_train_epochs": 128, |
|
"total_flos": 29937460838400000, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|