imdatta0's picture
End of training
8b1b653 verified
raw
history blame
27.8 kB
{
"best_metric": 1.967491626739502,
"best_model_checkpoint": "/home/datta0/models/lora_final/Mistral-7B-v0.3_pct_ortho_r32/checkpoint-8",
"epoch": 0.9981059842836993,
"eval_steps": 8,
"global_step": 387,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0025790852307072333,
"grad_norm": 7.501750946044922,
"learning_rate": 1.25e-05,
"loss": 2.2239,
"step": 1
},
{
"epoch": 0.010316340922828933,
"grad_norm": 4.365421295166016,
"learning_rate": 5e-05,
"loss": 2.1072,
"step": 4
},
{
"epoch": 0.020632681845657867,
"grad_norm": 3.2742128372192383,
"learning_rate": 0.0001,
"loss": 1.9673,
"step": 8
},
{
"epoch": 0.020632681845657867,
"eval_loss": 1.967491626739502,
"eval_runtime": 94.9785,
"eval_samples_per_second": 2.59,
"eval_steps_per_second": 2.59,
"step": 8
},
{
"epoch": 0.0309490227684868,
"grad_norm": 3.315385580062866,
"learning_rate": 9.997251843068762e-05,
"loss": 1.9892,
"step": 12
},
{
"epoch": 0.04126536369131573,
"grad_norm": 3.2009124755859375,
"learning_rate": 9.989010393221656e-05,
"loss": 1.982,
"step": 16
},
{
"epoch": 0.04126536369131573,
"eval_loss": 1.976182460784912,
"eval_runtime": 122.0257,
"eval_samples_per_second": 2.016,
"eval_steps_per_second": 2.016,
"step": 16
},
{
"epoch": 0.05158170461414467,
"grad_norm": 2.4668922424316406,
"learning_rate": 9.97528470997769e-05,
"loss": 1.931,
"step": 20
},
{
"epoch": 0.0618980455369736,
"grad_norm": 2.5818123817443848,
"learning_rate": 9.956089881469482e-05,
"loss": 1.9563,
"step": 24
},
{
"epoch": 0.0618980455369736,
"eval_loss": 1.9745111465454102,
"eval_runtime": 82.5917,
"eval_samples_per_second": 2.979,
"eval_steps_per_second": 2.979,
"step": 24
},
{
"epoch": 0.07221438645980253,
"grad_norm": 2.7342212200164795,
"learning_rate": 9.931447007857432e-05,
"loss": 2.003,
"step": 28
},
{
"epoch": 0.08253072738263147,
"grad_norm": 2.6568496227264404,
"learning_rate": 9.901383178135113e-05,
"loss": 1.957,
"step": 32
},
{
"epoch": 0.08253072738263147,
"eval_loss": 1.97785484790802,
"eval_runtime": 110.4681,
"eval_samples_per_second": 2.227,
"eval_steps_per_second": 2.227,
"step": 32
},
{
"epoch": 0.09284706830546041,
"grad_norm": 3.633096218109131,
"learning_rate": 9.865931440351337e-05,
"loss": 1.9643,
"step": 36
},
{
"epoch": 0.10316340922828934,
"grad_norm": 2.6414403915405273,
"learning_rate": 9.825130765281668e-05,
"loss": 2.0248,
"step": 40
},
{
"epoch": 0.10316340922828934,
"eval_loss": 1.9851678609848022,
"eval_runtime": 85.4255,
"eval_samples_per_second": 2.88,
"eval_steps_per_second": 2.88,
"step": 40
},
{
"epoch": 0.11347975015111827,
"grad_norm": 4.06710958480835,
"learning_rate": 9.779026003589304e-05,
"loss": 1.9751,
"step": 44
},
{
"epoch": 0.1237960910739472,
"grad_norm": 2.688786506652832,
"learning_rate": 9.727667836522407e-05,
"loss": 1.9753,
"step": 48
},
{
"epoch": 0.1237960910739472,
"eval_loss": 1.9988040924072266,
"eval_runtime": 83.5761,
"eval_samples_per_second": 2.943,
"eval_steps_per_second": 2.943,
"step": 48
},
{
"epoch": 0.13411243199677614,
"grad_norm": 2.983027696609497,
"learning_rate": 9.6711127202021e-05,
"loss": 2.0274,
"step": 52
},
{
"epoch": 0.14442877291960507,
"grad_norm": 2.963914394378662,
"learning_rate": 9.609422823562345e-05,
"loss": 1.9752,
"step": 56
},
{
"epoch": 0.14442877291960507,
"eval_loss": 1.997394323348999,
"eval_runtime": 109.8713,
"eval_samples_per_second": 2.239,
"eval_steps_per_second": 2.239,
"step": 56
},
{
"epoch": 0.154745113842434,
"grad_norm": 2.4576425552368164,
"learning_rate": 9.542665960009959e-05,
"loss": 1.9912,
"step": 60
},
{
"epoch": 0.16506145476526293,
"grad_norm": 2.9133598804473877,
"learning_rate": 9.470915512879852e-05,
"loss": 2.0253,
"step": 64
},
{
"epoch": 0.16506145476526293,
"eval_loss": 1.994693398475647,
"eval_runtime": 85.1875,
"eval_samples_per_second": 2.888,
"eval_steps_per_second": 2.888,
"step": 64
},
{
"epoch": 0.1753777956880919,
"grad_norm": 2.7807934284210205,
"learning_rate": 9.394250354767467e-05,
"loss": 2.0054,
"step": 68
},
{
"epoch": 0.18569413661092082,
"grad_norm": 2.75816011428833,
"learning_rate": 9.312754760827061e-05,
"loss": 2.0073,
"step": 72
},
{
"epoch": 0.18569413661092082,
"eval_loss": 1.9871970415115356,
"eval_runtime": 116.486,
"eval_samples_per_second": 2.112,
"eval_steps_per_second": 2.112,
"step": 72
},
{
"epoch": 0.19601047753374976,
"grad_norm": 2.501340389251709,
"learning_rate": 9.226518316131176e-05,
"loss": 2.013,
"step": 76
},
{
"epoch": 0.2063268184565787,
"grad_norm": 3.0418202877044678,
"learning_rate": 9.1356358171931e-05,
"loss": 1.9826,
"step": 80
},
{
"epoch": 0.2063268184565787,
"eval_loss": 1.9953396320343018,
"eval_runtime": 79.6585,
"eval_samples_per_second": 3.088,
"eval_steps_per_second": 3.088,
"step": 80
},
{
"epoch": 0.21664315937940762,
"grad_norm": 2.696978807449341,
"learning_rate": 9.040207167760586e-05,
"loss": 1.9693,
"step": 84
},
{
"epoch": 0.22695950030223655,
"grad_norm": 2.604274034500122,
"learning_rate": 8.940337268995385e-05,
"loss": 1.9907,
"step": 88
},
{
"epoch": 0.22695950030223655,
"eval_loss": 2.0014798641204834,
"eval_runtime": 119.4751,
"eval_samples_per_second": 2.059,
"eval_steps_per_second": 2.059,
"step": 88
},
{
"epoch": 0.23727584122506548,
"grad_norm": 2.6495442390441895,
"learning_rate": 8.836135904159302e-05,
"loss": 2.0053,
"step": 92
},
{
"epoch": 0.2475921821478944,
"grad_norm": 2.7287392616271973,
"learning_rate": 8.727717617933544e-05,
"loss": 1.9795,
"step": 96
},
{
"epoch": 0.2475921821478944,
"eval_loss": 1.9950815439224243,
"eval_runtime": 82.9311,
"eval_samples_per_second": 2.966,
"eval_steps_per_second": 2.966,
"step": 96
},
{
"epoch": 0.25790852307072337,
"grad_norm": 2.6722488403320312,
"learning_rate": 8.615201590504017e-05,
"loss": 2.033,
"step": 100
},
{
"epoch": 0.2682248639935523,
"grad_norm": 3.8678576946258545,
"learning_rate": 8.498711506550983e-05,
"loss": 1.9882,
"step": 104
},
{
"epoch": 0.2682248639935523,
"eval_loss": 2.0020015239715576,
"eval_runtime": 79.5681,
"eval_samples_per_second": 3.092,
"eval_steps_per_second": 3.092,
"step": 104
},
{
"epoch": 0.27854120491638124,
"grad_norm": 3.1447510719299316,
"learning_rate": 8.378375419287099e-05,
"loss": 2.0347,
"step": 108
},
{
"epoch": 0.28885754583921014,
"grad_norm": 2.7415413856506348,
"learning_rate": 8.25432560969328e-05,
"loss": 1.9896,
"step": 112
},
{
"epoch": 0.28885754583921014,
"eval_loss": 1.9962644577026367,
"eval_runtime": 119.6182,
"eval_samples_per_second": 2.057,
"eval_steps_per_second": 2.057,
"step": 112
},
{
"epoch": 0.2991738867620391,
"grad_norm": 2.7823774814605713,
"learning_rate": 8.126698441107146e-05,
"loss": 1.9733,
"step": 116
},
{
"epoch": 0.309490227684868,
"grad_norm": 2.875882387161255,
"learning_rate": 7.995634209323886e-05,
"loss": 2.0177,
"step": 120
},
{
"epoch": 0.309490227684868,
"eval_loss": 2.0145890712738037,
"eval_runtime": 82.837,
"eval_samples_per_second": 2.97,
"eval_steps_per_second": 2.97,
"step": 120
},
{
"epoch": 0.31980656860769696,
"grad_norm": 2.600053548812866,
"learning_rate": 7.861276988374302e-05,
"loss": 2.0148,
"step": 124
},
{
"epoch": 0.33012290953052587,
"grad_norm": 3.0531198978424072,
"learning_rate": 7.723774472149601e-05,
"loss": 2.0131,
"step": 128
},
{
"epoch": 0.33012290953052587,
"eval_loss": 2.0012505054473877,
"eval_runtime": 112.1719,
"eval_samples_per_second": 2.193,
"eval_steps_per_second": 2.193,
"step": 128
},
{
"epoch": 0.3404392504533548,
"grad_norm": 2.9197065830230713,
"learning_rate": 7.583277812046993e-05,
"loss": 2.0205,
"step": 132
},
{
"epoch": 0.3507555913761838,
"grad_norm": 2.766697406768799,
"learning_rate": 7.439941450814591e-05,
"loss": 2.0384,
"step": 136
},
{
"epoch": 0.3507555913761838,
"eval_loss": 2.001709222793579,
"eval_runtime": 84.1224,
"eval_samples_per_second": 2.924,
"eval_steps_per_second": 2.924,
"step": 136
},
{
"epoch": 0.3610719322990127,
"grad_norm": 3.0844602584838867,
"learning_rate": 7.293922952778239e-05,
"loss": 2.0319,
"step": 140
},
{
"epoch": 0.37138827322184165,
"grad_norm": 3.19889760017395,
"learning_rate": 7.145382830636924e-05,
"loss": 2.0587,
"step": 144
},
{
"epoch": 0.37138827322184165,
"eval_loss": 2.001880168914795,
"eval_runtime": 114.9668,
"eval_samples_per_second": 2.14,
"eval_steps_per_second": 2.14,
"step": 144
},
{
"epoch": 0.38170461414467055,
"grad_norm": 3.044001817703247,
"learning_rate": 6.994484369017143e-05,
"loss": 2.0052,
"step": 148
},
{
"epoch": 0.3920209550674995,
"grad_norm": 3.1708483695983887,
"learning_rate": 6.841393444980177e-05,
"loss": 1.9998,
"step": 152
},
{
"epoch": 0.3920209550674995,
"eval_loss": 1.996483325958252,
"eval_runtime": 111.8248,
"eval_samples_per_second": 2.2,
"eval_steps_per_second": 2.2,
"step": 152
},
{
"epoch": 0.4023372959903284,
"grad_norm": 2.9223573207855225,
"learning_rate": 6.686278345679625e-05,
"loss": 1.9855,
"step": 156
},
{
"epoch": 0.4126536369131574,
"grad_norm": 2.8330798149108887,
"learning_rate": 6.529309583369605e-05,
"loss": 1.9729,
"step": 160
},
{
"epoch": 0.4126536369131574,
"eval_loss": 1.9904886484146118,
"eval_runtime": 85.3147,
"eval_samples_per_second": 2.883,
"eval_steps_per_second": 2.883,
"step": 160
},
{
"epoch": 0.4229699778359863,
"grad_norm": 2.665085792541504,
"learning_rate": 6.370659707966967e-05,
"loss": 1.994,
"step": 164
},
{
"epoch": 0.43328631875881524,
"grad_norm": 3.091874122619629,
"learning_rate": 6.2105031173736e-05,
"loss": 2.0339,
"step": 168
},
{
"epoch": 0.43328631875881524,
"eval_loss": 2.0232627391815186,
"eval_runtime": 113.1394,
"eval_samples_per_second": 2.174,
"eval_steps_per_second": 2.174,
"step": 168
},
{
"epoch": 0.44360265968164414,
"grad_norm": 3.6090798377990723,
"learning_rate": 6.049015865767318e-05,
"loss": 1.9818,
"step": 172
},
{
"epoch": 0.4539190006044731,
"grad_norm": 3.000669479370117,
"learning_rate": 5.88637547007204e-05,
"loss": 2.0029,
"step": 176
},
{
"epoch": 0.4539190006044731,
"eval_loss": 1.9972038269042969,
"eval_runtime": 80.5833,
"eval_samples_per_second": 3.053,
"eval_steps_per_second": 3.053,
"step": 176
},
{
"epoch": 0.46423534152730206,
"grad_norm": 2.6875836849212646,
"learning_rate": 5.722760714820057e-05,
"loss": 2.0263,
"step": 180
},
{
"epoch": 0.47455168245013096,
"grad_norm": 3.1060383319854736,
"learning_rate": 5.5583514556208514e-05,
"loss": 1.997,
"step": 184
},
{
"epoch": 0.47455168245013096,
"eval_loss": 1.997556209564209,
"eval_runtime": 116.9074,
"eval_samples_per_second": 2.104,
"eval_steps_per_second": 2.104,
"step": 184
},
{
"epoch": 0.4848680233729599,
"grad_norm": 2.5845789909362793,
"learning_rate": 5.393328421452514e-05,
"loss": 1.9811,
"step": 188
},
{
"epoch": 0.4951843642957888,
"grad_norm": 2.890994071960449,
"learning_rate": 5.2278730159931076e-05,
"loss": 1.9808,
"step": 192
},
{
"epoch": 0.4951843642957888,
"eval_loss": 2.000748872756958,
"eval_runtime": 83.3047,
"eval_samples_per_second": 2.953,
"eval_steps_per_second": 2.953,
"step": 192
},
{
"epoch": 0.5055007052186178,
"grad_norm": 3.2054591178894043,
"learning_rate": 5.062167118210367e-05,
"loss": 2.0514,
"step": 196
},
{
"epoch": 0.5158170461414467,
"grad_norm": 2.911627769470215,
"learning_rate": 4.896392882428901e-05,
"loss": 2.0169,
"step": 200
},
{
"epoch": 0.5158170461414467,
"eval_loss": 1.9872047901153564,
"eval_runtime": 78.3344,
"eval_samples_per_second": 3.14,
"eval_steps_per_second": 3.14,
"step": 200
},
{
"epoch": 0.5261333870642756,
"grad_norm": 3.094301223754883,
"learning_rate": 4.730732538094749e-05,
"loss": 1.9815,
"step": 204
},
{
"epoch": 0.5364497279871046,
"grad_norm": 2.6194097995758057,
"learning_rate": 4.565368189457313e-05,
"loss": 1.9605,
"step": 208
},
{
"epoch": 0.5364497279871046,
"eval_loss": 1.9974828958511353,
"eval_runtime": 116.5117,
"eval_samples_per_second": 2.111,
"eval_steps_per_second": 2.111,
"step": 208
},
{
"epoch": 0.5467660689099335,
"grad_norm": 3.0435950756073,
"learning_rate": 4.400481615388948e-05,
"loss": 1.9643,
"step": 212
},
{
"epoch": 0.5570824098327625,
"grad_norm": 2.7957849502563477,
"learning_rate": 4.236254069562213e-05,
"loss": 2.0195,
"step": 216
},
{
"epoch": 0.5570824098327625,
"eval_loss": 1.9963051080703735,
"eval_runtime": 83.9136,
"eval_samples_per_second": 2.932,
"eval_steps_per_second": 2.932,
"step": 216
},
{
"epoch": 0.5673987507555914,
"grad_norm": 2.5869903564453125,
"learning_rate": 4.0728660812044536e-05,
"loss": 2.0336,
"step": 220
},
{
"epoch": 0.5777150916784203,
"grad_norm": 2.6142055988311768,
"learning_rate": 3.910497256648742e-05,
"loss": 1.9619,
"step": 224
},
{
"epoch": 0.5777150916784203,
"eval_loss": 1.9877939224243164,
"eval_runtime": 113.1041,
"eval_samples_per_second": 2.175,
"eval_steps_per_second": 2.175,
"step": 224
},
{
"epoch": 0.5880314326012492,
"grad_norm": 2.57497501373291,
"learning_rate": 3.749326081899329e-05,
"loss": 1.9923,
"step": 228
},
{
"epoch": 0.5983477735240782,
"grad_norm": 2.97292160987854,
"learning_rate": 3.589529726428615e-05,
"loss": 1.9361,
"step": 232
},
{
"epoch": 0.5983477735240782,
"eval_loss": 2.004483699798584,
"eval_runtime": 85.1956,
"eval_samples_per_second": 2.887,
"eval_steps_per_second": 2.887,
"step": 232
},
{
"epoch": 0.6086641144469072,
"grad_norm": 2.6015868186950684,
"learning_rate": 3.431283848421347e-05,
"loss": 2.033,
"step": 236
},
{
"epoch": 0.618980455369736,
"grad_norm": 2.7763192653656006,
"learning_rate": 3.274762401680124e-05,
"loss": 1.9932,
"step": 240
},
{
"epoch": 0.618980455369736,
"eval_loss": 1.9814573526382446,
"eval_runtime": 114.536,
"eval_samples_per_second": 2.148,
"eval_steps_per_second": 2.148,
"step": 240
},
{
"epoch": 0.629296796292565,
"grad_norm": 2.6632089614868164,
"learning_rate": 3.120137444404442e-05,
"loss": 1.9622,
"step": 244
},
{
"epoch": 0.6396131372153939,
"grad_norm": 2.701653480529785,
"learning_rate": 2.9675789500535328e-05,
"loss": 1.9519,
"step": 248
},
{
"epoch": 0.6396131372153939,
"eval_loss": 1.9895679950714111,
"eval_runtime": 79.8573,
"eval_samples_per_second": 3.08,
"eval_steps_per_second": 3.08,
"step": 248
},
{
"epoch": 0.6499294781382229,
"grad_norm": 2.8519248962402344,
"learning_rate": 2.8172546205008683e-05,
"loss": 2.0187,
"step": 252
},
{
"epoch": 0.6602458190610517,
"grad_norm": 2.4840056896209717,
"learning_rate": 2.6693297016857188e-05,
"loss": 1.9843,
"step": 256
},
{
"epoch": 0.6602458190610517,
"eval_loss": 1.9900814294815063,
"eval_runtime": 89.9348,
"eval_samples_per_second": 2.735,
"eval_steps_per_second": 2.735,
"step": 256
},
{
"epoch": 0.6705621599838807,
"grad_norm": 2.6297292709350586,
"learning_rate": 2.523966801964468e-05,
"loss": 1.985,
"step": 260
},
{
"epoch": 0.6808785009067096,
"grad_norm": 2.6009626388549805,
"learning_rate": 2.3813257133612827e-05,
"loss": 1.963,
"step": 264
},
{
"epoch": 0.6808785009067096,
"eval_loss": 1.9819916486740112,
"eval_runtime": 117.6742,
"eval_samples_per_second": 2.091,
"eval_steps_per_second": 2.091,
"step": 264
},
{
"epoch": 0.6911948418295386,
"grad_norm": 2.4555246829986572,
"learning_rate": 2.2415632359146856e-05,
"loss": 1.9888,
"step": 268
},
{
"epoch": 0.7015111827523676,
"grad_norm": 2.845592975616455,
"learning_rate": 2.104833005313131e-05,
"loss": 1.9376,
"step": 272
},
{
"epoch": 0.7015111827523676,
"eval_loss": 1.979328989982605,
"eval_runtime": 83.6262,
"eval_samples_per_second": 2.942,
"eval_steps_per_second": 2.942,
"step": 272
},
{
"epoch": 0.7118275236751964,
"grad_norm": 2.663501262664795,
"learning_rate": 1.971285324008994e-05,
"loss": 2.0383,
"step": 276
},
{
"epoch": 0.7221438645980254,
"grad_norm": 2.5632455348968506,
"learning_rate": 1.84106699599668e-05,
"loss": 1.9876,
"step": 280
},
{
"epoch": 0.7221438645980254,
"eval_loss": 1.9884700775146484,
"eval_runtime": 126.7545,
"eval_samples_per_second": 1.941,
"eval_steps_per_second": 1.941,
"step": 280
},
{
"epoch": 0.7324602055208543,
"grad_norm": 2.6025450229644775,
"learning_rate": 1.7143211654364762e-05,
"loss": 2.0093,
"step": 284
},
{
"epoch": 0.7427765464436833,
"grad_norm": 3.391676187515259,
"learning_rate": 1.5911871593014837e-05,
"loss": 2.0157,
"step": 288
},
{
"epoch": 0.7427765464436833,
"eval_loss": 1.9833872318267822,
"eval_runtime": 84.5923,
"eval_samples_per_second": 2.908,
"eval_steps_per_second": 2.908,
"step": 288
},
{
"epoch": 0.7530928873665121,
"grad_norm": 2.5039122104644775,
"learning_rate": 1.4718003342206722e-05,
"loss": 1.9728,
"step": 292
},
{
"epoch": 0.7634092282893411,
"grad_norm": 3.065073013305664,
"learning_rate": 1.3562919276863844e-05,
"loss": 2.011,
"step": 296
},
{
"epoch": 0.7634092282893411,
"eval_loss": 1.9842520952224731,
"eval_runtime": 81.8194,
"eval_samples_per_second": 3.007,
"eval_steps_per_second": 3.007,
"step": 296
},
{
"epoch": 0.7737255692121701,
"grad_norm": 2.7710351943969727,
"learning_rate": 1.2447889137898293e-05,
"loss": 2.0308,
"step": 300
},
{
"epoch": 0.784041910134999,
"grad_norm": 3.067017078399658,
"learning_rate": 1.1374138636432053e-05,
"loss": 2.0179,
"step": 304
},
{
"epoch": 0.784041910134999,
"eval_loss": 1.9779284000396729,
"eval_runtime": 128.2963,
"eval_samples_per_second": 1.917,
"eval_steps_per_second": 1.917,
"step": 304
},
{
"epoch": 0.794358251057828,
"grad_norm": 2.3697543144226074,
"learning_rate": 1.0342848106418368e-05,
"loss": 1.9998,
"step": 308
},
{
"epoch": 0.8046745919806568,
"grad_norm": 2.8660995960235596,
"learning_rate": 9.35515120714447e-06,
"loss": 1.9693,
"step": 312
},
{
"epoch": 0.8046745919806568,
"eval_loss": 1.978676676750183,
"eval_runtime": 85.106,
"eval_samples_per_second": 2.891,
"eval_steps_per_second": 2.891,
"step": 312
},
{
"epoch": 0.8149909329034858,
"grad_norm": 2.3432037830352783,
"learning_rate": 8.41213367704224e-06,
"loss": 2.0149,
"step": 316
},
{
"epoch": 0.8253072738263147,
"grad_norm": 2.753044605255127,
"learning_rate": 7.51483214017637e-06,
"loss": 1.9632,
"step": 320
},
{
"epoch": 0.8253072738263147,
"eval_loss": 1.982351541519165,
"eval_runtime": 114.3934,
"eval_samples_per_second": 2.15,
"eval_steps_per_second": 2.15,
"step": 320
},
{
"epoch": 0.8356236147491437,
"grad_norm": 2.3161306381225586,
"learning_rate": 6.664232966721995e-06,
"loss": 1.9832,
"step": 324
},
{
"epoch": 0.8459399556719726,
"grad_norm": 2.218827486038208,
"learning_rate": 5.8612711886848196e-06,
"loss": 1.9367,
"step": 328
},
{
"epoch": 0.8459399556719726,
"eval_loss": 1.9775891304016113,
"eval_runtime": 90.4335,
"eval_samples_per_second": 2.72,
"eval_steps_per_second": 2.72,
"step": 328
},
{
"epoch": 0.8562562965948015,
"grad_norm": 2.4218974113464355,
"learning_rate": 5.106829472055202e-06,
"loss": 1.9481,
"step": 332
},
{
"epoch": 0.8665726375176305,
"grad_norm": 2.6613683700561523,
"learning_rate": 4.401737146526219e-06,
"loss": 1.9824,
"step": 336
},
{
"epoch": 0.8665726375176305,
"eval_loss": 1.9729845523834229,
"eval_runtime": 119.6431,
"eval_samples_per_second": 2.056,
"eval_steps_per_second": 2.056,
"step": 336
},
{
"epoch": 0.8768889784404594,
"grad_norm": 2.633181095123291,
"learning_rate": 3.7467692938425057e-06,
"loss": 1.9396,
"step": 340
},
{
"epoch": 0.8872053193632883,
"grad_norm": 2.728327989578247,
"learning_rate": 3.142645895781715e-06,
"loss": 1.9911,
"step": 344
},
{
"epoch": 0.8872053193632883,
"eval_loss": 1.9718983173370361,
"eval_runtime": 82.9362,
"eval_samples_per_second": 2.966,
"eval_steps_per_second": 2.966,
"step": 344
},
{
"epoch": 0.8975216602861172,
"grad_norm": 2.4041829109191895,
"learning_rate": 2.5900310427053044e-06,
"loss": 1.9742,
"step": 348
},
{
"epoch": 0.9078380012089462,
"grad_norm": 2.4489846229553223,
"learning_rate": 2.089532203548794e-06,
"loss": 2.0075,
"step": 352
},
{
"epoch": 0.9078380012089462,
"eval_loss": 1.9730160236358643,
"eval_runtime": 91.0108,
"eval_samples_per_second": 2.703,
"eval_steps_per_second": 2.703,
"step": 352
},
{
"epoch": 0.9181543421317752,
"grad_norm": 2.5976321697235107,
"learning_rate": 1.6416995580537664e-06,
"loss": 1.9874,
"step": 356
},
{
"epoch": 0.9284706830546041,
"grad_norm": 3.244720458984375,
"learning_rate": 1.247025391975698e-06,
"loss": 1.9809,
"step": 360
},
{
"epoch": 0.9284706830546041,
"eval_loss": 1.9729856252670288,
"eval_runtime": 118.5331,
"eval_samples_per_second": 2.075,
"eval_steps_per_second": 2.075,
"step": 360
},
{
"epoch": 0.938787023977433,
"grad_norm": 2.2416179180145264,
"learning_rate": 9.059435559326257e-07,
"loss": 2.0036,
"step": 364
},
{
"epoch": 0.9491033649002619,
"grad_norm": 2.4049975872039795,
"learning_rate": 6.188289884893062e-07,
"loss": 1.9971,
"step": 368
},
{
"epoch": 0.9491033649002619,
"eval_loss": 1.9721975326538086,
"eval_runtime": 83.8998,
"eval_samples_per_second": 2.932,
"eval_steps_per_second": 2.932,
"step": 368
},
{
"epoch": 0.9594197058230909,
"grad_norm": 2.46026611328125,
"learning_rate": 3.8599730400115107e-07,
"loss": 1.9341,
"step": 372
},
{
"epoch": 0.9697360467459198,
"grad_norm": 2.334559917449951,
"learning_rate": 2.0770444567118075e-07,
"loss": 1.9913,
"step": 376
},
{
"epoch": 0.9697360467459198,
"eval_loss": 1.9720317125320435,
"eval_runtime": 127.7811,
"eval_samples_per_second": 1.925,
"eval_steps_per_second": 1.925,
"step": 376
},
{
"epoch": 0.9800523876687487,
"grad_norm": 2.3480026721954346,
"learning_rate": 8.414640420116305e-08,
"loss": 1.9948,
"step": 380
},
{
"epoch": 0.9903687285915777,
"grad_norm": 2.4215712547302246,
"learning_rate": 1.5459002346324135e-08,
"loss": 1.916,
"step": 384
},
{
"epoch": 0.9903687285915777,
"eval_loss": 1.972062587738037,
"eval_runtime": 84.2065,
"eval_samples_per_second": 2.921,
"eval_steps_per_second": 2.921,
"step": 384
}
],
"logging_steps": 4,
"max_steps": 387,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 8,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 4.5374998322774016e+17,
"train_batch_size": 1,
"trial_name": null,
"trial_params": null
}