Video-CoM-SFT / trainer_state.json
mmaaz60's picture
Upload folder using huggingface_hub
092e723 verified
{
"best_global_step": null,
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9995676610462603,
"eval_steps": 500,
"global_step": 578,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0017293558149589277,
"grad_norm": 14.538986412992182,
"learning_rate": 0.0,
"loss": 1.6313,
"step": 1
},
{
"epoch": 0.0034587116299178555,
"grad_norm": 14.326584895058424,
"learning_rate": 5.555555555555555e-07,
"loss": 1.5796,
"step": 2
},
{
"epoch": 0.005188067444876783,
"grad_norm": 14.303713845230615,
"learning_rate": 1.111111111111111e-06,
"loss": 1.6146,
"step": 3
},
{
"epoch": 0.006917423259835711,
"grad_norm": 14.270354087741836,
"learning_rate": 1.6666666666666667e-06,
"loss": 1.601,
"step": 4
},
{
"epoch": 0.00864677907479464,
"grad_norm": 13.056520461824615,
"learning_rate": 2.222222222222222e-06,
"loss": 1.5589,
"step": 5
},
{
"epoch": 0.010376134889753566,
"grad_norm": 10.770002519875254,
"learning_rate": 2.7777777777777783e-06,
"loss": 1.4667,
"step": 6
},
{
"epoch": 0.012105490704712495,
"grad_norm": 7.95324664209596,
"learning_rate": 3.3333333333333333e-06,
"loss": 1.2753,
"step": 7
},
{
"epoch": 0.013834846519671422,
"grad_norm": 7.532832624580702,
"learning_rate": 3.88888888888889e-06,
"loss": 1.2647,
"step": 8
},
{
"epoch": 0.01556420233463035,
"grad_norm": 5.2620835279629645,
"learning_rate": 4.444444444444444e-06,
"loss": 1.1498,
"step": 9
},
{
"epoch": 0.01729355814958928,
"grad_norm": 5.262371985501062,
"learning_rate": 5e-06,
"loss": 1.0606,
"step": 10
},
{
"epoch": 0.019022913964548204,
"grad_norm": 4.75552495213558,
"learning_rate": 5.555555555555557e-06,
"loss": 1.039,
"step": 11
},
{
"epoch": 0.020752269779507133,
"grad_norm": 4.109629839647021,
"learning_rate": 6.111111111111112e-06,
"loss": 0.9838,
"step": 12
},
{
"epoch": 0.02248162559446606,
"grad_norm": 3.2895830782411433,
"learning_rate": 6.666666666666667e-06,
"loss": 0.9485,
"step": 13
},
{
"epoch": 0.02421098140942499,
"grad_norm": 3.4515341118841554,
"learning_rate": 7.222222222222223e-06,
"loss": 0.8929,
"step": 14
},
{
"epoch": 0.02594033722438392,
"grad_norm": 3.5675009192020917,
"learning_rate": 7.77777777777778e-06,
"loss": 0.8761,
"step": 15
},
{
"epoch": 0.027669693039342844,
"grad_norm": 3.1238221221969846,
"learning_rate": 8.333333333333334e-06,
"loss": 0.8579,
"step": 16
},
{
"epoch": 0.029399048854301772,
"grad_norm": 2.730324810433143,
"learning_rate": 8.888888888888888e-06,
"loss": 0.825,
"step": 17
},
{
"epoch": 0.0311284046692607,
"grad_norm": 2.142543702234471,
"learning_rate": 9.444444444444445e-06,
"loss": 0.8275,
"step": 18
},
{
"epoch": 0.03285776048421963,
"grad_norm": 2.2407490739694618,
"learning_rate": 1e-05,
"loss": 0.798,
"step": 19
},
{
"epoch": 0.03458711629917856,
"grad_norm": 2.084673650668219,
"learning_rate": 9.999921320324328e-06,
"loss": 0.7791,
"step": 20
},
{
"epoch": 0.03631647211413749,
"grad_norm": 2.2050652721508817,
"learning_rate": 9.999685283773504e-06,
"loss": 0.801,
"step": 21
},
{
"epoch": 0.03804582792909641,
"grad_norm": 1.9949942374964489,
"learning_rate": 9.999291897776043e-06,
"loss": 0.7673,
"step": 22
},
{
"epoch": 0.03977518374405534,
"grad_norm": 1.9222200778544485,
"learning_rate": 9.998741174712534e-06,
"loss": 0.7546,
"step": 23
},
{
"epoch": 0.041504539559014265,
"grad_norm": 2.085559689480793,
"learning_rate": 9.998033131915266e-06,
"loss": 0.7733,
"step": 24
},
{
"epoch": 0.043233895373973194,
"grad_norm": 1.731301007495722,
"learning_rate": 9.997167791667668e-06,
"loss": 0.7406,
"step": 25
},
{
"epoch": 0.04496325118893212,
"grad_norm": 1.9090014318422763,
"learning_rate": 9.996145181203616e-06,
"loss": 0.7478,
"step": 26
},
{
"epoch": 0.04669260700389105,
"grad_norm": 1.6716462018458187,
"learning_rate": 9.994965332706574e-06,
"loss": 0.7293,
"step": 27
},
{
"epoch": 0.04842196281884998,
"grad_norm": 1.641744923224004,
"learning_rate": 9.993628283308582e-06,
"loss": 0.7157,
"step": 28
},
{
"epoch": 0.05015131863380891,
"grad_norm": 1.5639464733995954,
"learning_rate": 9.992134075089085e-06,
"loss": 0.7337,
"step": 29
},
{
"epoch": 0.05188067444876784,
"grad_norm": 1.5670396520195686,
"learning_rate": 9.990482755073607e-06,
"loss": 0.6944,
"step": 30
},
{
"epoch": 0.05361003026372676,
"grad_norm": 1.5501858401764228,
"learning_rate": 9.98867437523228e-06,
"loss": 0.7382,
"step": 31
},
{
"epoch": 0.05533938607868569,
"grad_norm": 1.5475730980812552,
"learning_rate": 9.986708992478202e-06,
"loss": 0.6836,
"step": 32
},
{
"epoch": 0.057068741893644616,
"grad_norm": 1.5368459860394041,
"learning_rate": 9.984586668665641e-06,
"loss": 0.7309,
"step": 33
},
{
"epoch": 0.058798097708603544,
"grad_norm": 1.573756977286856,
"learning_rate": 9.982307470588097e-06,
"loss": 0.6921,
"step": 34
},
{
"epoch": 0.06052745352356247,
"grad_norm": 1.4598653582851544,
"learning_rate": 9.979871469976197e-06,
"loss": 0.6824,
"step": 35
},
{
"epoch": 0.0622568093385214,
"grad_norm": 1.3407147196823341,
"learning_rate": 9.977278743495434e-06,
"loss": 0.695,
"step": 36
},
{
"epoch": 0.06398616515348032,
"grad_norm": 1.3427046109275345,
"learning_rate": 9.974529372743762e-06,
"loss": 0.6849,
"step": 37
},
{
"epoch": 0.06571552096843926,
"grad_norm": 1.385016607008538,
"learning_rate": 9.97162344424902e-06,
"loss": 0.704,
"step": 38
},
{
"epoch": 0.06744487678339818,
"grad_norm": 1.3624925480173673,
"learning_rate": 9.968561049466214e-06,
"loss": 0.6648,
"step": 39
},
{
"epoch": 0.06917423259835712,
"grad_norm": 1.6070208118878249,
"learning_rate": 9.965342284774633e-06,
"loss": 0.7004,
"step": 40
},
{
"epoch": 0.07090358841331604,
"grad_norm": 1.4348520811083416,
"learning_rate": 9.961967251474823e-06,
"loss": 0.6625,
"step": 41
},
{
"epoch": 0.07263294422827497,
"grad_norm": 1.3921519261612385,
"learning_rate": 9.958436055785391e-06,
"loss": 0.6709,
"step": 42
},
{
"epoch": 0.0743623000432339,
"grad_norm": 1.3214257024283271,
"learning_rate": 9.954748808839675e-06,
"loss": 0.6526,
"step": 43
},
{
"epoch": 0.07609165585819282,
"grad_norm": 1.3703861516781,
"learning_rate": 9.950905626682229e-06,
"loss": 0.6485,
"step": 44
},
{
"epoch": 0.07782101167315175,
"grad_norm": 1.3610898546796686,
"learning_rate": 9.946906630265184e-06,
"loss": 0.6509,
"step": 45
},
{
"epoch": 0.07955036748811067,
"grad_norm": 1.4666639612285957,
"learning_rate": 9.942751945444437e-06,
"loss": 0.6575,
"step": 46
},
{
"epoch": 0.08127972330306961,
"grad_norm": 1.2396616063563637,
"learning_rate": 9.938441702975689e-06,
"loss": 0.6395,
"step": 47
},
{
"epoch": 0.08300907911802853,
"grad_norm": 1.4092903953844345,
"learning_rate": 9.933976038510334e-06,
"loss": 0.6554,
"step": 48
},
{
"epoch": 0.08473843493298747,
"grad_norm": 1.7100974038421335,
"learning_rate": 9.92935509259118e-06,
"loss": 0.6528,
"step": 49
},
{
"epoch": 0.08646779074794639,
"grad_norm": 1.4242700514926885,
"learning_rate": 9.924579010648042e-06,
"loss": 0.6486,
"step": 50
},
{
"epoch": 0.08819714656290532,
"grad_norm": 1.5209362334431213,
"learning_rate": 9.91964794299315e-06,
"loss": 0.6773,
"step": 51
},
{
"epoch": 0.08992650237786425,
"grad_norm": 1.4185285804137782,
"learning_rate": 9.914562044816424e-06,
"loss": 0.6488,
"step": 52
},
{
"epoch": 0.09165585819282317,
"grad_norm": 1.3378741936637544,
"learning_rate": 9.909321476180594e-06,
"loss": 0.6517,
"step": 53
},
{
"epoch": 0.0933852140077821,
"grad_norm": 1.3516713188326783,
"learning_rate": 9.903926402016153e-06,
"loss": 0.6396,
"step": 54
},
{
"epoch": 0.09511456982274102,
"grad_norm": 1.5410717485635828,
"learning_rate": 9.898376992116179e-06,
"loss": 0.6447,
"step": 55
},
{
"epoch": 0.09684392563769996,
"grad_norm": 1.5098400169635777,
"learning_rate": 9.892673421130979e-06,
"loss": 0.6342,
"step": 56
},
{
"epoch": 0.09857328145265888,
"grad_norm": 1.4229655753415995,
"learning_rate": 9.886815868562596e-06,
"loss": 0.623,
"step": 57
},
{
"epoch": 0.10030263726761782,
"grad_norm": 1.3977945588526572,
"learning_rate": 9.88080451875917e-06,
"loss": 0.6397,
"step": 58
},
{
"epoch": 0.10203199308257674,
"grad_norm": 1.2877794451825768,
"learning_rate": 9.874639560909118e-06,
"loss": 0.6697,
"step": 59
},
{
"epoch": 0.10376134889753567,
"grad_norm": 1.4091354007797192,
"learning_rate": 9.868321189035196e-06,
"loss": 0.6359,
"step": 60
},
{
"epoch": 0.1054907047124946,
"grad_norm": 1.362969728293516,
"learning_rate": 9.861849601988384e-06,
"loss": 0.6418,
"step": 61
},
{
"epoch": 0.10722006052745352,
"grad_norm": 1.453550759696413,
"learning_rate": 9.855225003441629e-06,
"loss": 0.6469,
"step": 62
},
{
"epoch": 0.10894941634241245,
"grad_norm": 1.4051406418213799,
"learning_rate": 9.848447601883436e-06,
"loss": 0.6257,
"step": 63
},
{
"epoch": 0.11067877215737137,
"grad_norm": 1.3211083898505493,
"learning_rate": 9.841517610611309e-06,
"loss": 0.6396,
"step": 64
},
{
"epoch": 0.11240812797233031,
"grad_norm": 1.3740318481622646,
"learning_rate": 9.834435247725032e-06,
"loss": 0.6259,
"step": 65
},
{
"epoch": 0.11413748378728923,
"grad_norm": 1.484320978899618,
"learning_rate": 9.827200736119815e-06,
"loss": 0.6219,
"step": 66
},
{
"epoch": 0.11586683960224817,
"grad_norm": 1.3380643332618636,
"learning_rate": 9.819814303479268e-06,
"loss": 0.6192,
"step": 67
},
{
"epoch": 0.11759619541720709,
"grad_norm": 1.4153726476019632,
"learning_rate": 9.812276182268236e-06,
"loss": 0.617,
"step": 68
},
{
"epoch": 0.11932555123216602,
"grad_norm": 1.435579910730408,
"learning_rate": 9.804586609725499e-06,
"loss": 0.6082,
"step": 69
},
{
"epoch": 0.12105490704712495,
"grad_norm": 1.3611927229844167,
"learning_rate": 9.79674582785628e-06,
"loss": 0.6234,
"step": 70
},
{
"epoch": 0.12278426286208387,
"grad_norm": 1.4203295780177876,
"learning_rate": 9.788754083424654e-06,
"loss": 0.6333,
"step": 71
},
{
"epoch": 0.1245136186770428,
"grad_norm": 1.29595870212252,
"learning_rate": 9.78061162794576e-06,
"loss": 0.6267,
"step": 72
},
{
"epoch": 0.12624297449200172,
"grad_norm": 1.5861725748430446,
"learning_rate": 9.772318717677905e-06,
"loss": 0.6008,
"step": 73
},
{
"epoch": 0.12797233030696065,
"grad_norm": 1.3871875221331877,
"learning_rate": 9.763875613614482e-06,
"loss": 0.6099,
"step": 74
},
{
"epoch": 0.1297016861219196,
"grad_norm": 1.3701024745992456,
"learning_rate": 9.755282581475769e-06,
"loss": 0.6165,
"step": 75
},
{
"epoch": 0.13143104193687852,
"grad_norm": 1.3461257221248066,
"learning_rate": 9.746539891700558e-06,
"loss": 0.6329,
"step": 76
},
{
"epoch": 0.13316039775183744,
"grad_norm": 1.3814830311022952,
"learning_rate": 9.737647819437645e-06,
"loss": 0.6238,
"step": 77
},
{
"epoch": 0.13488975356679636,
"grad_norm": 1.3317024763543674,
"learning_rate": 9.728606644537177e-06,
"loss": 0.6436,
"step": 78
},
{
"epoch": 0.13661910938175528,
"grad_norm": 1.3559610943853933,
"learning_rate": 9.719416651541839e-06,
"loss": 0.6093,
"step": 79
},
{
"epoch": 0.13834846519671423,
"grad_norm": 1.3142606469258882,
"learning_rate": 9.710078129677895e-06,
"loss": 0.6265,
"step": 80
},
{
"epoch": 0.14007782101167315,
"grad_norm": 1.4252162273974214,
"learning_rate": 9.700591372846096e-06,
"loss": 0.62,
"step": 81
},
{
"epoch": 0.14180717682663208,
"grad_norm": 1.284483722976347,
"learning_rate": 9.690956679612422e-06,
"loss": 0.6387,
"step": 82
},
{
"epoch": 0.143536532641591,
"grad_norm": 1.4287308609026501,
"learning_rate": 9.681174353198687e-06,
"loss": 0.6155,
"step": 83
},
{
"epoch": 0.14526588845654995,
"grad_norm": 1.3659377539854851,
"learning_rate": 9.671244701472999e-06,
"loss": 0.5813,
"step": 84
},
{
"epoch": 0.14699524427150887,
"grad_norm": 1.3254239612936913,
"learning_rate": 9.661168036940071e-06,
"loss": 0.607,
"step": 85
},
{
"epoch": 0.1487246000864678,
"grad_norm": 1.3325242903071612,
"learning_rate": 9.650944676731383e-06,
"loss": 0.5811,
"step": 86
},
{
"epoch": 0.1504539559014267,
"grad_norm": 1.3721199226780008,
"learning_rate": 9.640574942595195e-06,
"loss": 0.6213,
"step": 87
},
{
"epoch": 0.15218331171638563,
"grad_norm": 1.3252154289940847,
"learning_rate": 9.63005916088644e-06,
"loss": 0.598,
"step": 88
},
{
"epoch": 0.15391266753134458,
"grad_norm": 1.341549271858989,
"learning_rate": 9.619397662556434e-06,
"loss": 0.6159,
"step": 89
},
{
"epoch": 0.1556420233463035,
"grad_norm": 1.306931095537237,
"learning_rate": 9.608590783142471e-06,
"loss": 0.5866,
"step": 90
},
{
"epoch": 0.15737137916126243,
"grad_norm": 1.3395574538440298,
"learning_rate": 9.597638862757255e-06,
"loss": 0.6132,
"step": 91
},
{
"epoch": 0.15910073497622135,
"grad_norm": 1.4247495939885102,
"learning_rate": 9.586542246078203e-06,
"loss": 0.6033,
"step": 92
},
{
"epoch": 0.1608300907911803,
"grad_norm": 1.4103034043856166,
"learning_rate": 9.5753012823366e-06,
"loss": 0.5958,
"step": 93
},
{
"epoch": 0.16255944660613922,
"grad_norm": 1.2779807076086858,
"learning_rate": 9.563916325306595e-06,
"loss": 0.5999,
"step": 94
},
{
"epoch": 0.16428880242109814,
"grad_norm": 1.3678860405045183,
"learning_rate": 9.552387733294081e-06,
"loss": 0.6047,
"step": 95
},
{
"epoch": 0.16601815823605706,
"grad_norm": 1.4177812307852564,
"learning_rate": 9.540715869125407e-06,
"loss": 0.6032,
"step": 96
},
{
"epoch": 0.16774751405101598,
"grad_norm": 1.3929063773953412,
"learning_rate": 9.528901100135971e-06,
"loss": 0.5914,
"step": 97
},
{
"epoch": 0.16947686986597493,
"grad_norm": 1.4452904668325175,
"learning_rate": 9.51694379815865e-06,
"loss": 0.591,
"step": 98
},
{
"epoch": 0.17120622568093385,
"grad_norm": 1.384558049529939,
"learning_rate": 9.504844339512096e-06,
"loss": 0.5935,
"step": 99
},
{
"epoch": 0.17293558149589278,
"grad_norm": 1.3744626820811583,
"learning_rate": 9.492603104988907e-06,
"loss": 0.6093,
"step": 100
},
{
"epoch": 0.1746649373108517,
"grad_norm": 1.4015343739030506,
"learning_rate": 9.480220479843627e-06,
"loss": 0.5864,
"step": 101
},
{
"epoch": 0.17639429312581065,
"grad_norm": 1.3434145487030242,
"learning_rate": 9.467696853780625e-06,
"loss": 0.5907,
"step": 102
},
{
"epoch": 0.17812364894076957,
"grad_norm": 1.3106599461212003,
"learning_rate": 9.45503262094184e-06,
"loss": 0.6035,
"step": 103
},
{
"epoch": 0.1798530047557285,
"grad_norm": 1.2960137141000876,
"learning_rate": 9.442228179894362e-06,
"loss": 0.58,
"step": 104
},
{
"epoch": 0.1815823605706874,
"grad_norm": 1.339610530818961,
"learning_rate": 9.4292839336179e-06,
"loss": 0.6019,
"step": 105
},
{
"epoch": 0.18331171638564633,
"grad_norm": 1.437718437019135,
"learning_rate": 9.416200289492092e-06,
"loss": 0.6098,
"step": 106
},
{
"epoch": 0.18504107220060528,
"grad_norm": 1.4523767804936614,
"learning_rate": 9.40297765928369e-06,
"loss": 0.6018,
"step": 107
},
{
"epoch": 0.1867704280155642,
"grad_norm": 1.2828185912784842,
"learning_rate": 9.389616459133597e-06,
"loss": 0.5764,
"step": 108
},
{
"epoch": 0.18849978383052313,
"grad_norm": 1.4256098510891433,
"learning_rate": 9.376117109543769e-06,
"loss": 0.6159,
"step": 109
},
{
"epoch": 0.19022913964548205,
"grad_norm": 1.3501440943806944,
"learning_rate": 9.362480035363987e-06,
"loss": 0.5793,
"step": 110
},
{
"epoch": 0.191958495460441,
"grad_norm": 1.3361244720373389,
"learning_rate": 9.348705665778479e-06,
"loss": 0.5712,
"step": 111
},
{
"epoch": 0.19368785127539992,
"grad_norm": 1.2782083486564007,
"learning_rate": 9.334794434292416e-06,
"loss": 0.5825,
"step": 112
},
{
"epoch": 0.19541720709035884,
"grad_norm": 1.2823908295358897,
"learning_rate": 9.320746778718274e-06,
"loss": 0.6332,
"step": 113
},
{
"epoch": 0.19714656290531776,
"grad_norm": 1.4254059120553437,
"learning_rate": 9.306563141162046e-06,
"loss": 0.6021,
"step": 114
},
{
"epoch": 0.19887591872027668,
"grad_norm": 1.3615562993020114,
"learning_rate": 9.292243968009332e-06,
"loss": 0.5919,
"step": 115
},
{
"epoch": 0.20060527453523563,
"grad_norm": 1.4003633053761828,
"learning_rate": 9.27778970991129e-06,
"loss": 0.6122,
"step": 116
},
{
"epoch": 0.20233463035019456,
"grad_norm": 1.4139482327306883,
"learning_rate": 9.263200821770462e-06,
"loss": 0.5731,
"step": 117
},
{
"epoch": 0.20406398616515348,
"grad_norm": 1.4647997165826316,
"learning_rate": 9.248477762726438e-06,
"loss": 0.5797,
"step": 118
},
{
"epoch": 0.2057933419801124,
"grad_norm": 1.3922473806972735,
"learning_rate": 9.233620996141421e-06,
"loss": 0.5598,
"step": 119
},
{
"epoch": 0.20752269779507135,
"grad_norm": 1.3498874982995888,
"learning_rate": 9.218630989585647e-06,
"loss": 0.5824,
"step": 120
},
{
"epoch": 0.20925205361003027,
"grad_norm": 1.4059303390475026,
"learning_rate": 9.203508214822652e-06,
"loss": 0.6209,
"step": 121
},
{
"epoch": 0.2109814094249892,
"grad_norm": 1.4908066800771167,
"learning_rate": 9.188253147794443e-06,
"loss": 0.5784,
"step": 122
},
{
"epoch": 0.2127107652399481,
"grad_norm": 1.3718507338485877,
"learning_rate": 9.172866268606514e-06,
"loss": 0.5962,
"step": 123
},
{
"epoch": 0.21444012105490703,
"grad_norm": 1.402469081391288,
"learning_rate": 9.157348061512728e-06,
"loss": 0.5929,
"step": 124
},
{
"epoch": 0.21616947686986598,
"grad_norm": 1.3206199086660828,
"learning_rate": 9.141699014900084e-06,
"loss": 0.5903,
"step": 125
},
{
"epoch": 0.2178988326848249,
"grad_norm": 1.3281498552638782,
"learning_rate": 9.125919621273348e-06,
"loss": 0.5946,
"step": 126
},
{
"epoch": 0.21962818849978383,
"grad_norm": 1.3209410288937773,
"learning_rate": 9.110010377239552e-06,
"loss": 0.5626,
"step": 127
},
{
"epoch": 0.22135754431474275,
"grad_norm": 1.3329308119803185,
"learning_rate": 9.093971783492354e-06,
"loss": 0.5918,
"step": 128
},
{
"epoch": 0.2230869001297017,
"grad_norm": 1.347489334146743,
"learning_rate": 9.077804344796302e-06,
"loss": 0.5878,
"step": 129
},
{
"epoch": 0.22481625594466062,
"grad_norm": 1.4072905998791443,
"learning_rate": 9.061508569970926e-06,
"loss": 0.5895,
"step": 130
},
{
"epoch": 0.22654561175961954,
"grad_norm": 1.4623429864651023,
"learning_rate": 9.045084971874738e-06,
"loss": 0.5935,
"step": 131
},
{
"epoch": 0.22827496757457846,
"grad_norm": 1.3789532216954916,
"learning_rate": 9.028534067389087e-06,
"loss": 0.5886,
"step": 132
},
{
"epoch": 0.23000432338953739,
"grad_norm": 1.3644289851875127,
"learning_rate": 9.011856377401891e-06,
"loss": 0.5682,
"step": 133
},
{
"epoch": 0.23173367920449633,
"grad_norm": 1.2442423854348275,
"learning_rate": 8.995052426791247e-06,
"loss": 0.5684,
"step": 134
},
{
"epoch": 0.23346303501945526,
"grad_norm": 1.3245488152280875,
"learning_rate": 8.978122744408905e-06,
"loss": 0.5643,
"step": 135
},
{
"epoch": 0.23519239083441418,
"grad_norm": 1.3934642875348537,
"learning_rate": 8.961067863063638e-06,
"loss": 0.5707,
"step": 136
},
{
"epoch": 0.2369217466493731,
"grad_norm": 1.3586108466135645,
"learning_rate": 8.943888319504456e-06,
"loss": 0.5936,
"step": 137
},
{
"epoch": 0.23865110246433205,
"grad_norm": 1.2787757325817652,
"learning_rate": 8.926584654403725e-06,
"loss": 0.5743,
"step": 138
},
{
"epoch": 0.24038045827929097,
"grad_norm": 1.2401520618532709,
"learning_rate": 8.90915741234015e-06,
"loss": 0.5738,
"step": 139
},
{
"epoch": 0.2421098140942499,
"grad_norm": 1.3715834375196405,
"learning_rate": 8.89160714178163e-06,
"loss": 0.594,
"step": 140
},
{
"epoch": 0.2438391699092088,
"grad_norm": 1.3920239192472779,
"learning_rate": 8.873934395068006e-06,
"loss": 0.5891,
"step": 141
},
{
"epoch": 0.24556852572416774,
"grad_norm": 1.280071307258292,
"learning_rate": 8.856139728393667e-06,
"loss": 0.586,
"step": 142
},
{
"epoch": 0.24729788153912668,
"grad_norm": 1.4348874555698314,
"learning_rate": 8.838223701790057e-06,
"loss": 0.6031,
"step": 143
},
{
"epoch": 0.2490272373540856,
"grad_norm": 1.2605512084319181,
"learning_rate": 8.820186879108038e-06,
"loss": 0.579,
"step": 144
},
{
"epoch": 0.25075659316904453,
"grad_norm": 1.3303160437037416,
"learning_rate": 8.802029828000157e-06,
"loss": 0.5848,
"step": 145
},
{
"epoch": 0.25248594898400345,
"grad_norm": 1.4582040782143946,
"learning_rate": 8.783753119902766e-06,
"loss": 0.5758,
"step": 146
},
{
"epoch": 0.25421530479896237,
"grad_norm": 1.322898639608662,
"learning_rate": 8.765357330018056e-06,
"loss": 0.5869,
"step": 147
},
{
"epoch": 0.2559446606139213,
"grad_norm": 1.384126692032235,
"learning_rate": 8.746843037295936e-06,
"loss": 0.5588,
"step": 148
},
{
"epoch": 0.2576740164288802,
"grad_norm": 1.3286178990420618,
"learning_rate": 8.728210824415829e-06,
"loss": 0.5688,
"step": 149
},
{
"epoch": 0.2594033722438392,
"grad_norm": 1.255847642552832,
"learning_rate": 8.70946127776832e-06,
"loss": 0.5436,
"step": 150
},
{
"epoch": 0.2611327280587981,
"grad_norm": 1.5446363806642398,
"learning_rate": 8.690594987436705e-06,
"loss": 0.5669,
"step": 151
},
{
"epoch": 0.26286208387375704,
"grad_norm": 1.3185596223777625,
"learning_rate": 8.671612547178428e-06,
"loss": 0.5652,
"step": 152
},
{
"epoch": 0.26459143968871596,
"grad_norm": 1.3558175344108856,
"learning_rate": 8.652514554406388e-06,
"loss": 0.5872,
"step": 153
},
{
"epoch": 0.2663207955036749,
"grad_norm": 1.4068740291750244,
"learning_rate": 8.633301610170136e-06,
"loss": 0.5683,
"step": 154
},
{
"epoch": 0.2680501513186338,
"grad_norm": 1.3133616699361563,
"learning_rate": 8.613974319136959e-06,
"loss": 0.5693,
"step": 155
},
{
"epoch": 0.2697795071335927,
"grad_norm": 1.2872738406840478,
"learning_rate": 8.594533289572852e-06,
"loss": 0.5787,
"step": 156
},
{
"epoch": 0.27150886294855164,
"grad_norm": 1.3034688709492834,
"learning_rate": 8.574979133323378e-06,
"loss": 0.5644,
"step": 157
},
{
"epoch": 0.27323821876351057,
"grad_norm": 1.2605674494333527,
"learning_rate": 8.555312465794402e-06,
"loss": 0.5758,
"step": 158
},
{
"epoch": 0.27496757457846954,
"grad_norm": 1.3396122302856128,
"learning_rate": 8.535533905932739e-06,
"loss": 0.6039,
"step": 159
},
{
"epoch": 0.27669693039342846,
"grad_norm": 1.2966490137544022,
"learning_rate": 8.515644076206652e-06,
"loss": 0.5645,
"step": 160
},
{
"epoch": 0.2784262862083874,
"grad_norm": 1.4311476322477343,
"learning_rate": 8.495643602586287e-06,
"loss": 0.5705,
"step": 161
},
{
"epoch": 0.2801556420233463,
"grad_norm": 1.3422910724323573,
"learning_rate": 8.475533114523954e-06,
"loss": 0.5656,
"step": 162
},
{
"epoch": 0.28188499783830523,
"grad_norm": 1.2772032094100403,
"learning_rate": 8.455313244934324e-06,
"loss": 0.5907,
"step": 163
},
{
"epoch": 0.28361435365326415,
"grad_norm": 1.3072283691035707,
"learning_rate": 8.43498463017451e-06,
"loss": 0.5696,
"step": 164
},
{
"epoch": 0.2853437094682231,
"grad_norm": 1.3046935804589632,
"learning_rate": 8.414547910024035e-06,
"loss": 0.5775,
"step": 165
},
{
"epoch": 0.287073065283182,
"grad_norm": 1.2976841761251101,
"learning_rate": 8.39400372766471e-06,
"loss": 0.5714,
"step": 166
},
{
"epoch": 0.2888024210981409,
"grad_norm": 1.3247225670276452,
"learning_rate": 8.373352729660373e-06,
"loss": 0.5527,
"step": 167
},
{
"epoch": 0.2905317769130999,
"grad_norm": 1.3304518084464831,
"learning_rate": 8.352595565936554e-06,
"loss": 0.5772,
"step": 168
},
{
"epoch": 0.2922611327280588,
"grad_norm": 1.3567530464099045,
"learning_rate": 8.331732889760021e-06,
"loss": 0.5724,
"step": 169
},
{
"epoch": 0.29399048854301774,
"grad_norm": 1.3864092503101992,
"learning_rate": 8.310765357718207e-06,
"loss": 0.5785,
"step": 170
},
{
"epoch": 0.29571984435797666,
"grad_norm": 1.3431891176626012,
"learning_rate": 8.289693629698564e-06,
"loss": 0.5907,
"step": 171
},
{
"epoch": 0.2974492001729356,
"grad_norm": 1.3269529964125137,
"learning_rate": 8.268518368867781e-06,
"loss": 0.5794,
"step": 172
},
{
"epoch": 0.2991785559878945,
"grad_norm": 1.3677645716695777,
"learning_rate": 8.247240241650918e-06,
"loss": 0.5889,
"step": 173
},
{
"epoch": 0.3009079118028534,
"grad_norm": 1.4344791877726926,
"learning_rate": 8.22585991771044e-06,
"loss": 0.5501,
"step": 174
},
{
"epoch": 0.30263726761781234,
"grad_norm": 1.3773781206385498,
"learning_rate": 8.204378069925121e-06,
"loss": 0.5435,
"step": 175
},
{
"epoch": 0.30436662343277127,
"grad_norm": 1.3415343046809804,
"learning_rate": 8.182795374368893e-06,
"loss": 0.5658,
"step": 176
},
{
"epoch": 0.30609597924773024,
"grad_norm": 1.3607519825023242,
"learning_rate": 8.16111251028955e-06,
"loss": 0.5726,
"step": 177
},
{
"epoch": 0.30782533506268916,
"grad_norm": 1.3372886670110973,
"learning_rate": 8.139330160087374e-06,
"loss": 0.5801,
"step": 178
},
{
"epoch": 0.3095546908776481,
"grad_norm": 1.45444571694518,
"learning_rate": 8.117449009293668e-06,
"loss": 0.5675,
"step": 179
},
{
"epoch": 0.311284046692607,
"grad_norm": 1.4232508965452562,
"learning_rate": 8.095469746549172e-06,
"loss": 0.5502,
"step": 180
},
{
"epoch": 0.31301340250756593,
"grad_norm": 1.3513827085112067,
"learning_rate": 8.073393063582386e-06,
"loss": 0.5855,
"step": 181
},
{
"epoch": 0.31474275832252485,
"grad_norm": 1.3052608382865691,
"learning_rate": 8.051219655187818e-06,
"loss": 0.5618,
"step": 182
},
{
"epoch": 0.3164721141374838,
"grad_norm": 1.306514474737617,
"learning_rate": 8.0289502192041e-06,
"loss": 0.5617,
"step": 183
},
{
"epoch": 0.3182014699524427,
"grad_norm": 1.2967196879064988,
"learning_rate": 8.00658545649203e-06,
"loss": 0.5976,
"step": 184
},
{
"epoch": 0.3199308257674016,
"grad_norm": 1.4576437100632567,
"learning_rate": 7.984126070912519e-06,
"loss": 0.5688,
"step": 185
},
{
"epoch": 0.3216601815823606,
"grad_norm": 1.3704399401870944,
"learning_rate": 7.961572769304437e-06,
"loss": 0.5631,
"step": 186
},
{
"epoch": 0.3233895373973195,
"grad_norm": 1.319375380678081,
"learning_rate": 7.938926261462366e-06,
"loss": 0.5498,
"step": 187
},
{
"epoch": 0.32511889321227844,
"grad_norm": 1.3529344720904868,
"learning_rate": 7.916187260114264e-06,
"loss": 0.5423,
"step": 188
},
{
"epoch": 0.32684824902723736,
"grad_norm": 1.3027982855324163,
"learning_rate": 7.89335648089903e-06,
"loss": 0.582,
"step": 189
},
{
"epoch": 0.3285776048421963,
"grad_norm": 1.4050936006586736,
"learning_rate": 7.870434642343984e-06,
"loss": 0.5636,
"step": 190
},
{
"epoch": 0.3303069606571552,
"grad_norm": 1.2865924125836552,
"learning_rate": 7.84742246584226e-06,
"loss": 0.5494,
"step": 191
},
{
"epoch": 0.3320363164721141,
"grad_norm": 1.3715611689044482,
"learning_rate": 7.82432067563009e-06,
"loss": 0.566,
"step": 192
},
{
"epoch": 0.33376567228707305,
"grad_norm": 1.3718978175645358,
"learning_rate": 7.801129998764014e-06,
"loss": 0.5657,
"step": 193
},
{
"epoch": 0.33549502810203197,
"grad_norm": 1.2502241262211298,
"learning_rate": 7.777851165098012e-06,
"loss": 0.5656,
"step": 194
},
{
"epoch": 0.33722438391699094,
"grad_norm": 1.3852373692132978,
"learning_rate": 7.754484907260513e-06,
"loss": 0.5658,
"step": 195
},
{
"epoch": 0.33895373973194987,
"grad_norm": 1.3703136522991974,
"learning_rate": 7.731031960631354e-06,
"loss": 0.5716,
"step": 196
},
{
"epoch": 0.3406830955469088,
"grad_norm": 1.2486585135623867,
"learning_rate": 7.70749306331863e-06,
"loss": 0.5651,
"step": 197
},
{
"epoch": 0.3424124513618677,
"grad_norm": 1.2601495096730775,
"learning_rate": 7.68386895613546e-06,
"loss": 0.5546,
"step": 198
},
{
"epoch": 0.34414180717682663,
"grad_norm": 1.2502364746593353,
"learning_rate": 7.660160382576683e-06,
"loss": 0.5483,
"step": 199
},
{
"epoch": 0.34587116299178555,
"grad_norm": 1.3477149756990934,
"learning_rate": 7.636368088795451e-06,
"loss": 0.5472,
"step": 200
},
{
"epoch": 0.3476005188067445,
"grad_norm": 1.2901176761284996,
"learning_rate": 7.612492823579744e-06,
"loss": 0.5421,
"step": 201
},
{
"epoch": 0.3493298746217034,
"grad_norm": 1.4658846157690446,
"learning_rate": 7.588535338328816e-06,
"loss": 0.5906,
"step": 202
},
{
"epoch": 0.3510592304366623,
"grad_norm": 1.2964558327514957,
"learning_rate": 7.564496387029532e-06,
"loss": 0.5772,
"step": 203
},
{
"epoch": 0.3527885862516213,
"grad_norm": 1.2170162880503597,
"learning_rate": 7.540376726232648e-06,
"loss": 0.5808,
"step": 204
},
{
"epoch": 0.3545179420665802,
"grad_norm": 1.3354147357012909,
"learning_rate": 7.516177115029002e-06,
"loss": 0.5496,
"step": 205
},
{
"epoch": 0.35624729788153914,
"grad_norm": 1.3770402354419473,
"learning_rate": 7.491898315025615e-06,
"loss": 0.5718,
"step": 206
},
{
"epoch": 0.35797665369649806,
"grad_norm": 1.4004748528087516,
"learning_rate": 7.467541090321735e-06,
"loss": 0.5696,
"step": 207
},
{
"epoch": 0.359706009511457,
"grad_norm": 1.3103570302853653,
"learning_rate": 7.443106207484776e-06,
"loss": 0.5817,
"step": 208
},
{
"epoch": 0.3614353653264159,
"grad_norm": 1.4334519797119796,
"learning_rate": 7.4185944355261996e-06,
"loss": 0.5549,
"step": 209
},
{
"epoch": 0.3631647211413748,
"grad_norm": 1.397522237730983,
"learning_rate": 7.3940065458773146e-06,
"loss": 0.5436,
"step": 210
},
{
"epoch": 0.36489407695633375,
"grad_norm": 1.2965013000965513,
"learning_rate": 7.369343312364994e-06,
"loss": 0.5648,
"step": 211
},
{
"epoch": 0.36662343277129267,
"grad_norm": 1.3115268564578664,
"learning_rate": 7.344605511187322e-06,
"loss": 0.5663,
"step": 212
},
{
"epoch": 0.36835278858625164,
"grad_norm": 1.2857253578787209,
"learning_rate": 7.319793920889171e-06,
"loss": 0.5465,
"step": 213
},
{
"epoch": 0.37008214440121057,
"grad_norm": 1.2254130773211296,
"learning_rate": 7.294909322337689e-06,
"loss": 0.5332,
"step": 214
},
{
"epoch": 0.3718115002161695,
"grad_norm": 1.2505477215695469,
"learning_rate": 7.269952498697734e-06,
"loss": 0.5519,
"step": 215
},
{
"epoch": 0.3735408560311284,
"grad_norm": 1.2309151575793142,
"learning_rate": 7.244924235407224e-06,
"loss": 0.556,
"step": 216
},
{
"epoch": 0.37527021184608733,
"grad_norm": 1.2736490892998513,
"learning_rate": 7.219825320152411e-06,
"loss": 0.5753,
"step": 217
},
{
"epoch": 0.37699956766104625,
"grad_norm": 1.188067487294118,
"learning_rate": 7.194656542843103e-06,
"loss": 0.5457,
"step": 218
},
{
"epoch": 0.3787289234760052,
"grad_norm": 1.3573649976256121,
"learning_rate": 7.169418695587791e-06,
"loss": 0.5317,
"step": 219
},
{
"epoch": 0.3804582792909641,
"grad_norm": 1.3104765075971507,
"learning_rate": 7.1441125726687336e-06,
"loss": 0.5416,
"step": 220
},
{
"epoch": 0.382187635105923,
"grad_norm": 1.2524746976217112,
"learning_rate": 7.118738970516944e-06,
"loss": 0.5231,
"step": 221
},
{
"epoch": 0.383916990920882,
"grad_norm": 1.2912855193020514,
"learning_rate": 7.093298687687141e-06,
"loss": 0.577,
"step": 222
},
{
"epoch": 0.3856463467358409,
"grad_norm": 1.277812278463467,
"learning_rate": 7.067792524832604e-06,
"loss": 0.5627,
"step": 223
},
{
"epoch": 0.38737570255079984,
"grad_norm": 1.3047580618945316,
"learning_rate": 7.042221284679982e-06,
"loss": 0.5385,
"step": 224
},
{
"epoch": 0.38910505836575876,
"grad_norm": 1.1904223056651926,
"learning_rate": 7.016585772004026e-06,
"loss": 0.5504,
"step": 225
},
{
"epoch": 0.3908344141807177,
"grad_norm": 1.2398301118878279,
"learning_rate": 6.990886793602268e-06,
"loss": 0.5463,
"step": 226
},
{
"epoch": 0.3925637699956766,
"grad_norm": 1.2249280934558917,
"learning_rate": 6.965125158269619e-06,
"loss": 0.5524,
"step": 227
},
{
"epoch": 0.3942931258106355,
"grad_norm": 1.3151361816315885,
"learning_rate": 6.939301676772927e-06,
"loss": 0.5643,
"step": 228
},
{
"epoch": 0.39602248162559445,
"grad_norm": 1.4155270180747608,
"learning_rate": 6.913417161825449e-06,
"loss": 0.5396,
"step": 229
},
{
"epoch": 0.39775183744055337,
"grad_norm": 1.3122619817083747,
"learning_rate": 6.887472428061285e-06,
"loss": 0.5533,
"step": 230
},
{
"epoch": 0.39948119325551235,
"grad_norm": 1.2806770091938722,
"learning_rate": 6.8614682920097265e-06,
"loss": 0.5403,
"step": 231
},
{
"epoch": 0.40121054907047127,
"grad_norm": 1.3470173683287574,
"learning_rate": 6.835405572069572e-06,
"loss": 0.5395,
"step": 232
},
{
"epoch": 0.4029399048854302,
"grad_norm": 1.3345925151371414,
"learning_rate": 6.809285088483361e-06,
"loss": 0.5488,
"step": 233
},
{
"epoch": 0.4046692607003891,
"grad_norm": 1.364191950346444,
"learning_rate": 6.783107663311566e-06,
"loss": 0.5597,
"step": 234
},
{
"epoch": 0.40639861651534803,
"grad_norm": 1.229870751967945,
"learning_rate": 6.7568741204067145e-06,
"loss": 0.5532,
"step": 235
},
{
"epoch": 0.40812797233030695,
"grad_norm": 1.2511626652001453,
"learning_rate": 6.730585285387465e-06,
"loss": 0.5599,
"step": 236
},
{
"epoch": 0.4098573281452659,
"grad_norm": 1.1963629945054264,
"learning_rate": 6.704241985612625e-06,
"loss": 0.5662,
"step": 237
},
{
"epoch": 0.4115866839602248,
"grad_norm": 1.28804770356803,
"learning_rate": 6.6778450501551065e-06,
"loss": 0.5801,
"step": 238
},
{
"epoch": 0.4133160397751837,
"grad_norm": 1.2590013900847816,
"learning_rate": 6.651395309775837e-06,
"loss": 0.5644,
"step": 239
},
{
"epoch": 0.4150453955901427,
"grad_norm": 1.2551864064194094,
"learning_rate": 6.6248935968976135e-06,
"loss": 0.5394,
"step": 240
},
{
"epoch": 0.4167747514051016,
"grad_norm": 1.324809116558963,
"learning_rate": 6.598340745578908e-06,
"loss": 0.5698,
"step": 241
},
{
"epoch": 0.41850410722006054,
"grad_norm": 1.3179110819475957,
"learning_rate": 6.571737591487611e-06,
"loss": 0.5438,
"step": 242
},
{
"epoch": 0.42023346303501946,
"grad_norm": 1.2492716164423538,
"learning_rate": 6.545084971874738e-06,
"loss": 0.5534,
"step": 243
},
{
"epoch": 0.4219628188499784,
"grad_norm": 1.3793013860663719,
"learning_rate": 6.518383725548074e-06,
"loss": 0.5507,
"step": 244
},
{
"epoch": 0.4236921746649373,
"grad_norm": 1.2552437595758494,
"learning_rate": 6.491634692845781e-06,
"loss": 0.5283,
"step": 245
},
{
"epoch": 0.4254215304798962,
"grad_norm": 1.320285724366206,
"learning_rate": 6.464838715609945e-06,
"loss": 0.5381,
"step": 246
},
{
"epoch": 0.42715088629485515,
"grad_norm": 1.321244687463225,
"learning_rate": 6.437996637160086e-06,
"loss": 0.5632,
"step": 247
},
{
"epoch": 0.42888024210981407,
"grad_norm": 1.298650328971777,
"learning_rate": 6.411109302266616e-06,
"loss": 0.5436,
"step": 248
},
{
"epoch": 0.43060959792477305,
"grad_norm": 1.2466524194529058,
"learning_rate": 6.384177557124247e-06,
"loss": 0.536,
"step": 249
},
{
"epoch": 0.43233895373973197,
"grad_norm": 1.2659973914064022,
"learning_rate": 6.3572022493253715e-06,
"loss": 0.5666,
"step": 250
},
{
"epoch": 0.4340683095546909,
"grad_norm": 1.3125673782338234,
"learning_rate": 6.330184227833376e-06,
"loss": 0.5269,
"step": 251
},
{
"epoch": 0.4357976653696498,
"grad_norm": 1.2471832027915164,
"learning_rate": 6.303124342955928e-06,
"loss": 0.5277,
"step": 252
},
{
"epoch": 0.43752702118460873,
"grad_norm": 1.3130769407311655,
"learning_rate": 6.276023446318214e-06,
"loss": 0.5628,
"step": 253
},
{
"epoch": 0.43925637699956765,
"grad_norm": 1.2407127444236052,
"learning_rate": 6.248882390836135e-06,
"loss": 0.5583,
"step": 254
},
{
"epoch": 0.4409857328145266,
"grad_norm": 1.374714499529323,
"learning_rate": 6.2217020306894705e-06,
"loss": 0.5298,
"step": 255
},
{
"epoch": 0.4427150886294855,
"grad_norm": 1.4800838058998909,
"learning_rate": 6.194483221294989e-06,
"loss": 0.5469,
"step": 256
},
{
"epoch": 0.4444444444444444,
"grad_norm": 1.1817922709588438,
"learning_rate": 6.1672268192795285e-06,
"loss": 0.5425,
"step": 257
},
{
"epoch": 0.4461738002594034,
"grad_norm": 1.2263700782728304,
"learning_rate": 6.139933682453035e-06,
"loss": 0.5598,
"step": 258
},
{
"epoch": 0.4479031560743623,
"grad_norm": 1.3081432662421466,
"learning_rate": 6.112604669781572e-06,
"loss": 0.5451,
"step": 259
},
{
"epoch": 0.44963251188932124,
"grad_norm": 1.3868536914950849,
"learning_rate": 6.085240641360281e-06,
"loss": 0.5424,
"step": 260
},
{
"epoch": 0.45136186770428016,
"grad_norm": 1.2274166600226197,
"learning_rate": 6.057842458386315e-06,
"loss": 0.537,
"step": 261
},
{
"epoch": 0.4530912235192391,
"grad_norm": 1.272644249773423,
"learning_rate": 6.030410983131733e-06,
"loss": 0.56,
"step": 262
},
{
"epoch": 0.454820579334198,
"grad_norm": 1.2893096518593115,
"learning_rate": 6.002947078916365e-06,
"loss": 0.541,
"step": 263
},
{
"epoch": 0.4565499351491569,
"grad_norm": 1.2447414895886646,
"learning_rate": 5.975451610080643e-06,
"loss": 0.5439,
"step": 264
},
{
"epoch": 0.45827929096411585,
"grad_norm": 1.217222966677454,
"learning_rate": 5.947925441958393e-06,
"loss": 0.542,
"step": 265
},
{
"epoch": 0.46000864677907477,
"grad_norm": 1.2886825234554569,
"learning_rate": 5.920369440849609e-06,
"loss": 0.5543,
"step": 266
},
{
"epoch": 0.46173800259403375,
"grad_norm": 1.3251969210640597,
"learning_rate": 5.892784473993184e-06,
"loss": 0.5399,
"step": 267
},
{
"epoch": 0.46346735840899267,
"grad_norm": 1.2268752980398028,
"learning_rate": 5.865171409539614e-06,
"loss": 0.5467,
"step": 268
},
{
"epoch": 0.4651967142239516,
"grad_norm": 1.3068892898030797,
"learning_rate": 5.837531116523683e-06,
"loss": 0.5401,
"step": 269
},
{
"epoch": 0.4669260700389105,
"grad_norm": 1.2594383194818888,
"learning_rate": 5.809864464837105e-06,
"loss": 0.5584,
"step": 270
},
{
"epoch": 0.46865542585386943,
"grad_norm": 1.2235669424052997,
"learning_rate": 5.782172325201155e-06,
"loss": 0.5311,
"step": 271
},
{
"epoch": 0.47038478166882836,
"grad_norm": 1.3219449360910378,
"learning_rate": 5.754455569139258e-06,
"loss": 0.5299,
"step": 272
},
{
"epoch": 0.4721141374837873,
"grad_norm": 1.2887549822531428,
"learning_rate": 5.726715068949564e-06,
"loss": 0.5451,
"step": 273
},
{
"epoch": 0.4738434932987462,
"grad_norm": 1.2358733083826365,
"learning_rate": 5.698951697677498e-06,
"loss": 0.5379,
"step": 274
},
{
"epoch": 0.4755728491137051,
"grad_norm": 1.3090051655251154,
"learning_rate": 5.671166329088278e-06,
"loss": 0.5213,
"step": 275
},
{
"epoch": 0.4773022049286641,
"grad_norm": 1.3042917976396255,
"learning_rate": 5.643359837639419e-06,
"loss": 0.5235,
"step": 276
},
{
"epoch": 0.479031560743623,
"grad_norm": 1.298269280525944,
"learning_rate": 5.615533098453215e-06,
"loss": 0.5545,
"step": 277
},
{
"epoch": 0.48076091655858194,
"grad_norm": 1.2012345750779942,
"learning_rate": 5.587686987289189e-06,
"loss": 0.5544,
"step": 278
},
{
"epoch": 0.48249027237354086,
"grad_norm": 1.3007588033367934,
"learning_rate": 5.559822380516539e-06,
"loss": 0.5271,
"step": 279
},
{
"epoch": 0.4842196281884998,
"grad_norm": 1.3534538377292569,
"learning_rate": 5.531940155086557e-06,
"loss": 0.5688,
"step": 280
},
{
"epoch": 0.4859489840034587,
"grad_norm": 1.243334314478317,
"learning_rate": 5.504041188505022e-06,
"loss": 0.5473,
"step": 281
},
{
"epoch": 0.4876783398184176,
"grad_norm": 1.306645781202671,
"learning_rate": 5.476126358804594e-06,
"loss": 0.5422,
"step": 282
},
{
"epoch": 0.48940769563337655,
"grad_norm": 1.280897747736695,
"learning_rate": 5.448196544517168e-06,
"loss": 0.5516,
"step": 283
},
{
"epoch": 0.49113705144833547,
"grad_norm": 1.4064375573811547,
"learning_rate": 5.420252624646238e-06,
"loss": 0.5467,
"step": 284
},
{
"epoch": 0.49286640726329445,
"grad_norm": 1.4119469280433727,
"learning_rate": 5.392295478639226e-06,
"loss": 0.5712,
"step": 285
},
{
"epoch": 0.49459576307825337,
"grad_norm": 1.3832811952775954,
"learning_rate": 5.3643259863598015e-06,
"loss": 0.5311,
"step": 286
},
{
"epoch": 0.4963251188932123,
"grad_norm": 1.2793380120458782,
"learning_rate": 5.336345028060199e-06,
"loss": 0.5464,
"step": 287
},
{
"epoch": 0.4980544747081712,
"grad_norm": 1.325075817941503,
"learning_rate": 5.308353484353508e-06,
"loss": 0.5472,
"step": 288
},
{
"epoch": 0.49978383052313013,
"grad_norm": 1.323776420208779,
"learning_rate": 5.2803522361859596e-06,
"loss": 0.5162,
"step": 289
},
{
"epoch": 0.5015131863380891,
"grad_norm": 1.2724751142456727,
"learning_rate": 5.252342164809204e-06,
"loss": 0.5419,
"step": 290
},
{
"epoch": 0.503242542153048,
"grad_norm": 1.3155061940618022,
"learning_rate": 5.224324151752575e-06,
"loss": 0.5108,
"step": 291
},
{
"epoch": 0.5049718979680069,
"grad_norm": 1.325463563646962,
"learning_rate": 5.1962990787953436e-06,
"loss": 0.556,
"step": 292
},
{
"epoch": 0.5067012537829658,
"grad_norm": 1.3557284931566147,
"learning_rate": 5.168267827938971e-06,
"loss": 0.5205,
"step": 293
},
{
"epoch": 0.5084306095979247,
"grad_norm": 1.4690183640639547,
"learning_rate": 5.140231281379345e-06,
"loss": 0.5314,
"step": 294
},
{
"epoch": 0.5101599654128837,
"grad_norm": 1.3889196264518369,
"learning_rate": 5.112190321479026e-06,
"loss": 0.5421,
"step": 295
},
{
"epoch": 0.5118893212278426,
"grad_norm": 1.235942882014356,
"learning_rate": 5.084145830739462e-06,
"loss": 0.5311,
"step": 296
},
{
"epoch": 0.5136186770428015,
"grad_norm": 1.306512327093981,
"learning_rate": 5.05609869177323e-06,
"loss": 0.5273,
"step": 297
},
{
"epoch": 0.5153480328577604,
"grad_norm": 1.5917366093172325,
"learning_rate": 5.028049787276249e-06,
"loss": 0.5129,
"step": 298
},
{
"epoch": 0.5170773886727195,
"grad_norm": 1.40594572028163,
"learning_rate": 5e-06,
"loss": 0.5203,
"step": 299
},
{
"epoch": 0.5188067444876784,
"grad_norm": 1.262588520269215,
"learning_rate": 4.971950212723753e-06,
"loss": 0.5055,
"step": 300
},
{
"epoch": 0.5205361003026373,
"grad_norm": 1.4092188833304773,
"learning_rate": 4.943901308226771e-06,
"loss": 0.5434,
"step": 301
},
{
"epoch": 0.5222654561175962,
"grad_norm": 1.382690963218948,
"learning_rate": 4.91585416926054e-06,
"loss": 0.5449,
"step": 302
},
{
"epoch": 0.5239948119325551,
"grad_norm": 1.3495688952134626,
"learning_rate": 4.887809678520976e-06,
"loss": 0.5277,
"step": 303
},
{
"epoch": 0.5257241677475141,
"grad_norm": 1.2421088006172942,
"learning_rate": 4.859768718620656e-06,
"loss": 0.5393,
"step": 304
},
{
"epoch": 0.527453523562473,
"grad_norm": 1.2488445572106857,
"learning_rate": 4.831732172061032e-06,
"loss": 0.5225,
"step": 305
},
{
"epoch": 0.5291828793774319,
"grad_norm": 1.3844825536152099,
"learning_rate": 4.803700921204659e-06,
"loss": 0.5448,
"step": 306
},
{
"epoch": 0.5309122351923908,
"grad_norm": 1.2550069336336056,
"learning_rate": 4.775675848247427e-06,
"loss": 0.5326,
"step": 307
},
{
"epoch": 0.5326415910073498,
"grad_norm": 1.3042371941510982,
"learning_rate": 4.747657835190796e-06,
"loss": 0.5485,
"step": 308
},
{
"epoch": 0.5343709468223087,
"grad_norm": 1.439469833400696,
"learning_rate": 4.719647763814041e-06,
"loss": 0.5235,
"step": 309
},
{
"epoch": 0.5361003026372676,
"grad_norm": 1.258005734371576,
"learning_rate": 4.691646515646492e-06,
"loss": 0.5049,
"step": 310
},
{
"epoch": 0.5378296584522265,
"grad_norm": 1.2574136262278794,
"learning_rate": 4.663654971939802e-06,
"loss": 0.5306,
"step": 311
},
{
"epoch": 0.5395590142671854,
"grad_norm": 1.211366769644713,
"learning_rate": 4.6356740136402e-06,
"loss": 0.5158,
"step": 312
},
{
"epoch": 0.5412883700821444,
"grad_norm": 1.404490080536933,
"learning_rate": 4.6077045213607765e-06,
"loss": 0.5106,
"step": 313
},
{
"epoch": 0.5430177258971033,
"grad_norm": 1.3565090769351222,
"learning_rate": 4.579747375353763e-06,
"loss": 0.5115,
"step": 314
},
{
"epoch": 0.5447470817120622,
"grad_norm": 1.3745499639927738,
"learning_rate": 4.551803455482833e-06,
"loss": 0.5352,
"step": 315
},
{
"epoch": 0.5464764375270211,
"grad_norm": 1.3375963306559595,
"learning_rate": 4.5238736411954075e-06,
"loss": 0.5149,
"step": 316
},
{
"epoch": 0.5482057933419802,
"grad_norm": 1.2416690865187434,
"learning_rate": 4.4959588114949785e-06,
"loss": 0.5183,
"step": 317
},
{
"epoch": 0.5499351491569391,
"grad_norm": 1.2969100557148994,
"learning_rate": 4.468059844913444e-06,
"loss": 0.5305,
"step": 318
},
{
"epoch": 0.551664504971898,
"grad_norm": 1.3028745121039649,
"learning_rate": 4.4401776194834615e-06,
"loss": 0.53,
"step": 319
},
{
"epoch": 0.5533938607868569,
"grad_norm": 1.2529196452246425,
"learning_rate": 4.4123130127108125e-06,
"loss": 0.5631,
"step": 320
},
{
"epoch": 0.5551232166018158,
"grad_norm": 1.2914281020735483,
"learning_rate": 4.384466901546786e-06,
"loss": 0.5029,
"step": 321
},
{
"epoch": 0.5568525724167748,
"grad_norm": 1.3030244428041313,
"learning_rate": 4.356640162360582e-06,
"loss": 0.5547,
"step": 322
},
{
"epoch": 0.5585819282317337,
"grad_norm": 1.310975612444778,
"learning_rate": 4.3288336709117246e-06,
"loss": 0.5476,
"step": 323
},
{
"epoch": 0.5603112840466926,
"grad_norm": 1.3585233059927981,
"learning_rate": 4.3010483023225045e-06,
"loss": 0.5096,
"step": 324
},
{
"epoch": 0.5620406398616515,
"grad_norm": 1.3858533161552185,
"learning_rate": 4.273284931050438e-06,
"loss": 0.5432,
"step": 325
},
{
"epoch": 0.5637699956766105,
"grad_norm": 1.3005876771208573,
"learning_rate": 4.245544430860743e-06,
"loss": 0.5339,
"step": 326
},
{
"epoch": 0.5654993514915694,
"grad_norm": 1.3987689407306876,
"learning_rate": 4.217827674798845e-06,
"loss": 0.5096,
"step": 327
},
{
"epoch": 0.5672287073065283,
"grad_norm": 1.309864053992572,
"learning_rate": 4.190135535162894e-06,
"loss": 0.5074,
"step": 328
},
{
"epoch": 0.5689580631214872,
"grad_norm": 1.3268629259565734,
"learning_rate": 4.162468883476319e-06,
"loss": 0.5386,
"step": 329
},
{
"epoch": 0.5706874189364461,
"grad_norm": 1.3349966799384403,
"learning_rate": 4.1348285904603865e-06,
"loss": 0.5208,
"step": 330
},
{
"epoch": 0.5724167747514051,
"grad_norm": 1.2661602913195409,
"learning_rate": 4.107215526006818e-06,
"loss": 0.5107,
"step": 331
},
{
"epoch": 0.574146130566364,
"grad_norm": 1.2797920680965094,
"learning_rate": 4.079630559150392e-06,
"loss": 0.513,
"step": 332
},
{
"epoch": 0.5758754863813229,
"grad_norm": 1.3077982663455536,
"learning_rate": 4.052074558041608e-06,
"loss": 0.5154,
"step": 333
},
{
"epoch": 0.5776048421962818,
"grad_norm": 1.3760657710228763,
"learning_rate": 4.02454838991936e-06,
"loss": 0.5294,
"step": 334
},
{
"epoch": 0.5793341980112409,
"grad_norm": 1.184811772801038,
"learning_rate": 3.997052921083637e-06,
"loss": 0.4896,
"step": 335
},
{
"epoch": 0.5810635538261998,
"grad_norm": 1.4503676384261956,
"learning_rate": 3.969589016868269e-06,
"loss": 0.5131,
"step": 336
},
{
"epoch": 0.5827929096411587,
"grad_norm": 1.2572496861000189,
"learning_rate": 3.9421575416136866e-06,
"loss": 0.5356,
"step": 337
},
{
"epoch": 0.5845222654561176,
"grad_norm": 1.2168932626954965,
"learning_rate": 3.9147593586397195e-06,
"loss": 0.5012,
"step": 338
},
{
"epoch": 0.5862516212710766,
"grad_norm": 1.335579085108439,
"learning_rate": 3.887395330218429e-06,
"loss": 0.5073,
"step": 339
},
{
"epoch": 0.5879809770860355,
"grad_norm": 1.2148641410862266,
"learning_rate": 3.8600663175469675e-06,
"loss": 0.5216,
"step": 340
},
{
"epoch": 0.5897103329009944,
"grad_norm": 1.2477985045133986,
"learning_rate": 3.832773180720475e-06,
"loss": 0.509,
"step": 341
},
{
"epoch": 0.5914396887159533,
"grad_norm": 1.3416766428515066,
"learning_rate": 3.8055167787050134e-06,
"loss": 0.5163,
"step": 342
},
{
"epoch": 0.5931690445309122,
"grad_norm": 1.2682767103868227,
"learning_rate": 3.778297969310529e-06,
"loss": 0.527,
"step": 343
},
{
"epoch": 0.5948984003458712,
"grad_norm": 1.2414041985446462,
"learning_rate": 3.751117609163865e-06,
"loss": 0.5001,
"step": 344
},
{
"epoch": 0.5966277561608301,
"grad_norm": 1.375054008335257,
"learning_rate": 3.723976553681787e-06,
"loss": 0.5342,
"step": 345
},
{
"epoch": 0.598357111975789,
"grad_norm": 1.2480963236384548,
"learning_rate": 3.6968756570440735e-06,
"loss": 0.5028,
"step": 346
},
{
"epoch": 0.6000864677907479,
"grad_norm": 1.2159599196486677,
"learning_rate": 3.669815772166625e-06,
"loss": 0.5125,
"step": 347
},
{
"epoch": 0.6018158236057068,
"grad_norm": 1.4104148060749142,
"learning_rate": 3.6427977506746293e-06,
"loss": 0.5207,
"step": 348
},
{
"epoch": 0.6035451794206658,
"grad_norm": 1.1950066056646578,
"learning_rate": 3.6158224428757538e-06,
"loss": 0.4943,
"step": 349
},
{
"epoch": 0.6052745352356247,
"grad_norm": 1.147453355609145,
"learning_rate": 3.5888906977333858e-06,
"loss": 0.4992,
"step": 350
},
{
"epoch": 0.6070038910505836,
"grad_norm": 1.4493578720570028,
"learning_rate": 3.562003362839914e-06,
"loss": 0.5213,
"step": 351
},
{
"epoch": 0.6087332468655425,
"grad_norm": 1.2658733112800329,
"learning_rate": 3.5351612843900555e-06,
"loss": 0.5056,
"step": 352
},
{
"epoch": 0.6104626026805016,
"grad_norm": 1.1757360378271826,
"learning_rate": 3.50836530715422e-06,
"loss": 0.4856,
"step": 353
},
{
"epoch": 0.6121919584954605,
"grad_norm": 1.3055141846929021,
"learning_rate": 3.4816162744519266e-06,
"loss": 0.5317,
"step": 354
},
{
"epoch": 0.6139213143104194,
"grad_norm": 1.37730807773082,
"learning_rate": 3.4549150281252635e-06,
"loss": 0.5159,
"step": 355
},
{
"epoch": 0.6156506701253783,
"grad_norm": 1.1947534769598933,
"learning_rate": 3.42826240851239e-06,
"loss": 0.5214,
"step": 356
},
{
"epoch": 0.6173800259403373,
"grad_norm": 1.3226134189067087,
"learning_rate": 3.4016592544210937e-06,
"loss": 0.5264,
"step": 357
},
{
"epoch": 0.6191093817552962,
"grad_norm": 1.41629808960968,
"learning_rate": 3.375106403102389e-06,
"loss": 0.5379,
"step": 358
},
{
"epoch": 0.6208387375702551,
"grad_norm": 1.3266049262461792,
"learning_rate": 3.3486046902241663e-06,
"loss": 0.5017,
"step": 359
},
{
"epoch": 0.622568093385214,
"grad_norm": 1.3117139861674612,
"learning_rate": 3.322154949844897e-06,
"loss": 0.5066,
"step": 360
},
{
"epoch": 0.6242974492001729,
"grad_norm": 1.275223937661878,
"learning_rate": 3.295758014387375e-06,
"loss": 0.5107,
"step": 361
},
{
"epoch": 0.6260268050151319,
"grad_norm": 1.3609897446280972,
"learning_rate": 3.269414714612534e-06,
"loss": 0.5211,
"step": 362
},
{
"epoch": 0.6277561608300908,
"grad_norm": 1.4136951954746158,
"learning_rate": 3.2431258795932863e-06,
"loss": 0.5299,
"step": 363
},
{
"epoch": 0.6294855166450497,
"grad_norm": 1.3164541206983948,
"learning_rate": 3.216892336688435e-06,
"loss": 0.5251,
"step": 364
},
{
"epoch": 0.6312148724600086,
"grad_norm": 1.2729230339815254,
"learning_rate": 3.1907149115166403e-06,
"loss": 0.5052,
"step": 365
},
{
"epoch": 0.6329442282749675,
"grad_norm": 1.3720577760394352,
"learning_rate": 3.1645944279304296e-06,
"loss": 0.5142,
"step": 366
},
{
"epoch": 0.6346735840899265,
"grad_norm": 1.3838881292780676,
"learning_rate": 3.1385317079902743e-06,
"loss": 0.5301,
"step": 367
},
{
"epoch": 0.6364029399048854,
"grad_norm": 1.1823914759517473,
"learning_rate": 3.112527571938717e-06,
"loss": 0.5038,
"step": 368
},
{
"epoch": 0.6381322957198443,
"grad_norm": 1.4852396956399265,
"learning_rate": 3.0865828381745515e-06,
"loss": 0.5383,
"step": 369
},
{
"epoch": 0.6398616515348032,
"grad_norm": 1.3537270664670102,
"learning_rate": 3.060698323227075e-06,
"loss": 0.507,
"step": 370
},
{
"epoch": 0.6415910073497623,
"grad_norm": 1.313357691607948,
"learning_rate": 3.0348748417303826e-06,
"loss": 0.5076,
"step": 371
},
{
"epoch": 0.6433203631647212,
"grad_norm": 1.3130162160688268,
"learning_rate": 3.0091132063977344e-06,
"loss": 0.5267,
"step": 372
},
{
"epoch": 0.6450497189796801,
"grad_norm": 1.1920662735224292,
"learning_rate": 2.9834142279959754e-06,
"loss": 0.5155,
"step": 373
},
{
"epoch": 0.646779074794639,
"grad_norm": 1.2025632883759985,
"learning_rate": 2.95777871532002e-06,
"loss": 0.5096,
"step": 374
},
{
"epoch": 0.648508430609598,
"grad_norm": 1.3113109709173287,
"learning_rate": 2.932207475167398e-06,
"loss": 0.5002,
"step": 375
},
{
"epoch": 0.6502377864245569,
"grad_norm": 1.2461184315574925,
"learning_rate": 2.906701312312861e-06,
"loss": 0.5169,
"step": 376
},
{
"epoch": 0.6519671422395158,
"grad_norm": 1.2608068635952254,
"learning_rate": 2.8812610294830568e-06,
"loss": 0.5043,
"step": 377
},
{
"epoch": 0.6536964980544747,
"grad_norm": 1.322071977690957,
"learning_rate": 2.8558874273312677e-06,
"loss": 0.5039,
"step": 378
},
{
"epoch": 0.6554258538694336,
"grad_norm": 1.2333609191041428,
"learning_rate": 2.83058130441221e-06,
"loss": 0.5066,
"step": 379
},
{
"epoch": 0.6571552096843926,
"grad_norm": 1.2541631476872825,
"learning_rate": 2.805343457156898e-06,
"loss": 0.5191,
"step": 380
},
{
"epoch": 0.6588845654993515,
"grad_norm": 1.279925954820817,
"learning_rate": 2.7801746798475905e-06,
"loss": 0.4903,
"step": 381
},
{
"epoch": 0.6606139213143104,
"grad_norm": 1.2393630411242251,
"learning_rate": 2.755075764592777e-06,
"loss": 0.5007,
"step": 382
},
{
"epoch": 0.6623432771292693,
"grad_norm": 1.2113955627282111,
"learning_rate": 2.7300475013022666e-06,
"loss": 0.5187,
"step": 383
},
{
"epoch": 0.6640726329442282,
"grad_norm": 1.332875296702971,
"learning_rate": 2.705090677662311e-06,
"loss": 0.5116,
"step": 384
},
{
"epoch": 0.6658019887591872,
"grad_norm": 1.1776968966832972,
"learning_rate": 2.6802060791108304e-06,
"loss": 0.4994,
"step": 385
},
{
"epoch": 0.6675313445741461,
"grad_norm": 1.2110870121335493,
"learning_rate": 2.6553944888126772e-06,
"loss": 0.5184,
"step": 386
},
{
"epoch": 0.669260700389105,
"grad_norm": 1.3343623955630721,
"learning_rate": 2.6306566876350072e-06,
"loss": 0.5315,
"step": 387
},
{
"epoch": 0.6709900562040639,
"grad_norm": 1.238593700529844,
"learning_rate": 2.605993454122687e-06,
"loss": 0.5211,
"step": 388
},
{
"epoch": 0.672719412019023,
"grad_norm": 1.2623050551820758,
"learning_rate": 2.5814055644738013e-06,
"loss": 0.4985,
"step": 389
},
{
"epoch": 0.6744487678339819,
"grad_norm": 1.313293854036044,
"learning_rate": 2.5568937925152272e-06,
"loss": 0.524,
"step": 390
},
{
"epoch": 0.6761781236489408,
"grad_norm": 1.2517836743413833,
"learning_rate": 2.532458909678266e-06,
"loss": 0.5098,
"step": 391
},
{
"epoch": 0.6779074794638997,
"grad_norm": 1.1520446930274513,
"learning_rate": 2.508101684974387e-06,
"loss": 0.5015,
"step": 392
},
{
"epoch": 0.6796368352788587,
"grad_norm": 1.3005504514213648,
"learning_rate": 2.483822884971e-06,
"loss": 0.4803,
"step": 393
},
{
"epoch": 0.6813661910938176,
"grad_norm": 1.234117732958065,
"learning_rate": 2.4596232737673544e-06,
"loss": 0.5196,
"step": 394
},
{
"epoch": 0.6830955469087765,
"grad_norm": 1.1981906278111187,
"learning_rate": 2.43550361297047e-06,
"loss": 0.5059,
"step": 395
},
{
"epoch": 0.6848249027237354,
"grad_norm": 1.2997430500346672,
"learning_rate": 2.4114646616711844e-06,
"loss": 0.5008,
"step": 396
},
{
"epoch": 0.6865542585386943,
"grad_norm": 1.2614162559191038,
"learning_rate": 2.387507176420256e-06,
"loss": 0.502,
"step": 397
},
{
"epoch": 0.6882836143536533,
"grad_norm": 1.188230979407375,
"learning_rate": 2.3636319112045495e-06,
"loss": 0.5164,
"step": 398
},
{
"epoch": 0.6900129701686122,
"grad_norm": 1.334435091915391,
"learning_rate": 2.339839617423318e-06,
"loss": 0.5069,
"step": 399
},
{
"epoch": 0.6917423259835711,
"grad_norm": 1.2428049634796243,
"learning_rate": 2.31613104386454e-06,
"loss": 0.4934,
"step": 400
},
{
"epoch": 0.69347168179853,
"grad_norm": 1.283895066008161,
"learning_rate": 2.2925069366813718e-06,
"loss": 0.5075,
"step": 401
},
{
"epoch": 0.695201037613489,
"grad_norm": 1.276327425480895,
"learning_rate": 2.2689680393686457e-06,
"loss": 0.5138,
"step": 402
},
{
"epoch": 0.6969303934284479,
"grad_norm": 1.2630101628882442,
"learning_rate": 2.245515092739488e-06,
"loss": 0.5116,
"step": 403
},
{
"epoch": 0.6986597492434068,
"grad_norm": 1.1913216900412702,
"learning_rate": 2.2221488349019903e-06,
"loss": 0.5139,
"step": 404
},
{
"epoch": 0.7003891050583657,
"grad_norm": 1.3575890195381133,
"learning_rate": 2.1988700012359865e-06,
"loss": 0.4975,
"step": 405
},
{
"epoch": 0.7021184608733246,
"grad_norm": 1.205731465743301,
"learning_rate": 2.175679324369913e-06,
"loss": 0.5196,
"step": 406
},
{
"epoch": 0.7038478166882837,
"grad_norm": 1.2794680777002916,
"learning_rate": 2.1525775341577404e-06,
"loss": 0.4966,
"step": 407
},
{
"epoch": 0.7055771725032426,
"grad_norm": 1.2296036184996322,
"learning_rate": 2.1295653576560168e-06,
"loss": 0.4951,
"step": 408
},
{
"epoch": 0.7073065283182015,
"grad_norm": 1.2044790286403084,
"learning_rate": 2.1066435191009717e-06,
"loss": 0.4853,
"step": 409
},
{
"epoch": 0.7090358841331604,
"grad_norm": 1.2218236581464923,
"learning_rate": 2.083812739885738e-06,
"loss": 0.4934,
"step": 410
},
{
"epoch": 0.7107652399481194,
"grad_norm": 1.4009155350987372,
"learning_rate": 2.061073738537635e-06,
"loss": 0.4946,
"step": 411
},
{
"epoch": 0.7124945957630783,
"grad_norm": 1.1812811747354344,
"learning_rate": 2.038427230695565e-06,
"loss": 0.4951,
"step": 412
},
{
"epoch": 0.7142239515780372,
"grad_norm": 1.2191387291107492,
"learning_rate": 2.0158739290874822e-06,
"loss": 0.4986,
"step": 413
},
{
"epoch": 0.7159533073929961,
"grad_norm": 1.329295095070758,
"learning_rate": 1.9934145435079705e-06,
"loss": 0.519,
"step": 414
},
{
"epoch": 0.717682663207955,
"grad_norm": 1.2281812463093098,
"learning_rate": 1.971049780795901e-06,
"loss": 0.5497,
"step": 415
},
{
"epoch": 0.719412019022914,
"grad_norm": 1.179567921331548,
"learning_rate": 1.948780344812181e-06,
"loss": 0.4875,
"step": 416
},
{
"epoch": 0.7211413748378729,
"grad_norm": 1.3229803278267123,
"learning_rate": 1.9266069364176144e-06,
"loss": 0.513,
"step": 417
},
{
"epoch": 0.7228707306528318,
"grad_norm": 1.207231556517107,
"learning_rate": 1.9045302534508298e-06,
"loss": 0.5077,
"step": 418
},
{
"epoch": 0.7246000864677907,
"grad_norm": 1.239432725326481,
"learning_rate": 1.8825509907063328e-06,
"loss": 0.5081,
"step": 419
},
{
"epoch": 0.7263294422827496,
"grad_norm": 1.2768017670747227,
"learning_rate": 1.8606698399126261e-06,
"loss": 0.5047,
"step": 420
},
{
"epoch": 0.7280587980977086,
"grad_norm": 1.2443081672127343,
"learning_rate": 1.838887489710452e-06,
"loss": 0.5093,
"step": 421
},
{
"epoch": 0.7297881539126675,
"grad_norm": 1.3248575565441225,
"learning_rate": 1.8172046256311087e-06,
"loss": 0.5096,
"step": 422
},
{
"epoch": 0.7315175097276264,
"grad_norm": 1.2312772110980421,
"learning_rate": 1.7956219300748796e-06,
"loss": 0.4949,
"step": 423
},
{
"epoch": 0.7332468655425853,
"grad_norm": 1.2725397947766917,
"learning_rate": 1.7741400822895633e-06,
"loss": 0.4913,
"step": 424
},
{
"epoch": 0.7349762213575444,
"grad_norm": 1.2701112344747274,
"learning_rate": 1.7527597583490825e-06,
"loss": 0.5107,
"step": 425
},
{
"epoch": 0.7367055771725033,
"grad_norm": 1.22988252225655,
"learning_rate": 1.7314816311322219e-06,
"loss": 0.487,
"step": 426
},
{
"epoch": 0.7384349329874622,
"grad_norm": 1.23287252087033,
"learning_rate": 1.7103063703014372e-06,
"loss": 0.533,
"step": 427
},
{
"epoch": 0.7401642888024211,
"grad_norm": 1.184456040905248,
"learning_rate": 1.6892346422817945e-06,
"loss": 0.4993,
"step": 428
},
{
"epoch": 0.74189364461738,
"grad_norm": 1.220805415373837,
"learning_rate": 1.6682671102399806e-06,
"loss": 0.4865,
"step": 429
},
{
"epoch": 0.743623000432339,
"grad_norm": 1.1994409780411663,
"learning_rate": 1.647404434063447e-06,
"loss": 0.4992,
"step": 430
},
{
"epoch": 0.7453523562472979,
"grad_norm": 1.2631533307258702,
"learning_rate": 1.6266472703396286e-06,
"loss": 0.5236,
"step": 431
},
{
"epoch": 0.7470817120622568,
"grad_norm": 1.1929507809675688,
"learning_rate": 1.6059962723352912e-06,
"loss": 0.4921,
"step": 432
},
{
"epoch": 0.7488110678772157,
"grad_norm": 1.3115162066864332,
"learning_rate": 1.5854520899759656e-06,
"loss": 0.5123,
"step": 433
},
{
"epoch": 0.7505404236921747,
"grad_norm": 1.204903938948009,
"learning_rate": 1.5650153698254916e-06,
"loss": 0.4761,
"step": 434
},
{
"epoch": 0.7522697795071336,
"grad_norm": 1.3927511118217388,
"learning_rate": 1.544686755065677e-06,
"loss": 0.5209,
"step": 435
},
{
"epoch": 0.7539991353220925,
"grad_norm": 1.1622881435387544,
"learning_rate": 1.5244668854760459e-06,
"loss": 0.5054,
"step": 436
},
{
"epoch": 0.7557284911370514,
"grad_norm": 1.2474163957236213,
"learning_rate": 1.5043563974137132e-06,
"loss": 0.5037,
"step": 437
},
{
"epoch": 0.7574578469520103,
"grad_norm": 1.2604380979516712,
"learning_rate": 1.4843559237933474e-06,
"loss": 0.4891,
"step": 438
},
{
"epoch": 0.7591872027669693,
"grad_norm": 1.2530851990162881,
"learning_rate": 1.4644660940672628e-06,
"loss": 0.4903,
"step": 439
},
{
"epoch": 0.7609165585819282,
"grad_norm": 1.2300568194662433,
"learning_rate": 1.4446875342055988e-06,
"loss": 0.5059,
"step": 440
},
{
"epoch": 0.7626459143968871,
"grad_norm": 1.2621446251787094,
"learning_rate": 1.4250208666766235e-06,
"loss": 0.5084,
"step": 441
},
{
"epoch": 0.764375270211846,
"grad_norm": 1.1714965845789171,
"learning_rate": 1.4054667104271497e-06,
"loss": 0.4939,
"step": 442
},
{
"epoch": 0.7661046260268051,
"grad_norm": 1.150326828517619,
"learning_rate": 1.3860256808630429e-06,
"loss": 0.4998,
"step": 443
},
{
"epoch": 0.767833981841764,
"grad_norm": 1.4388728760063987,
"learning_rate": 1.3666983898298659e-06,
"loss": 0.506,
"step": 444
},
{
"epoch": 0.7695633376567229,
"grad_norm": 1.211792558367054,
"learning_rate": 1.3474854455936126e-06,
"loss": 0.4908,
"step": 445
},
{
"epoch": 0.7712926934716818,
"grad_norm": 1.1773193686251473,
"learning_rate": 1.3283874528215735e-06,
"loss": 0.4838,
"step": 446
},
{
"epoch": 0.7730220492866408,
"grad_norm": 1.2910457142537168,
"learning_rate": 1.3094050125632973e-06,
"loss": 0.5221,
"step": 447
},
{
"epoch": 0.7747514051015997,
"grad_norm": 1.2455298854301133,
"learning_rate": 1.2905387222316824e-06,
"loss": 0.5098,
"step": 448
},
{
"epoch": 0.7764807609165586,
"grad_norm": 1.2221925176741943,
"learning_rate": 1.2717891755841722e-06,
"loss": 0.4968,
"step": 449
},
{
"epoch": 0.7782101167315175,
"grad_norm": 1.2690867828669088,
"learning_rate": 1.2531569627040636e-06,
"loss": 0.497,
"step": 450
},
{
"epoch": 0.7799394725464764,
"grad_norm": 1.1942118517525946,
"learning_rate": 1.234642669981946e-06,
"loss": 0.524,
"step": 451
},
{
"epoch": 0.7816688283614354,
"grad_norm": 1.188395242204491,
"learning_rate": 1.2162468800972344e-06,
"loss": 0.4797,
"step": 452
},
{
"epoch": 0.7833981841763943,
"grad_norm": 1.239090182793657,
"learning_rate": 1.1979701719998454e-06,
"loss": 0.4852,
"step": 453
},
{
"epoch": 0.7851275399913532,
"grad_norm": 1.2443125589753226,
"learning_rate": 1.1798131208919628e-06,
"loss": 0.4869,
"step": 454
},
{
"epoch": 0.7868568958063121,
"grad_norm": 1.2094210133113774,
"learning_rate": 1.1617762982099446e-06,
"loss": 0.4977,
"step": 455
},
{
"epoch": 0.788586251621271,
"grad_norm": 1.2353201091140928,
"learning_rate": 1.143860271606333e-06,
"loss": 0.4867,
"step": 456
},
{
"epoch": 0.79031560743623,
"grad_norm": 1.2334222843362082,
"learning_rate": 1.1260656049319957e-06,
"loss": 0.4934,
"step": 457
},
{
"epoch": 0.7920449632511889,
"grad_norm": 0.9697687189599682,
"learning_rate": 1.1083928582183712e-06,
"loss": 0.5128,
"step": 458
},
{
"epoch": 0.7937743190661478,
"grad_norm": 0.9095178347207882,
"learning_rate": 1.0908425876598512e-06,
"loss": 0.5031,
"step": 459
},
{
"epoch": 0.7955036748811067,
"grad_norm": 0.9310196051126621,
"learning_rate": 1.0734153455962765e-06,
"loss": 0.5117,
"step": 460
},
{
"epoch": 0.7972330306960658,
"grad_norm": 0.9730539710942893,
"learning_rate": 1.0561116804955451e-06,
"loss": 0.5113,
"step": 461
},
{
"epoch": 0.7989623865110247,
"grad_norm": 0.894786526061026,
"learning_rate": 1.0389321369363636e-06,
"loss": 0.5018,
"step": 462
},
{
"epoch": 0.8006917423259836,
"grad_norm": 0.9674498638331304,
"learning_rate": 1.0218772555910955e-06,
"loss": 0.5042,
"step": 463
},
{
"epoch": 0.8024210981409425,
"grad_norm": 0.9438765948616291,
"learning_rate": 1.004947573208756e-06,
"loss": 0.5198,
"step": 464
},
{
"epoch": 0.8041504539559015,
"grad_norm": 0.9763518960839491,
"learning_rate": 9.881436225981107e-07,
"loss": 0.5058,
"step": 465
},
{
"epoch": 0.8058798097708604,
"grad_norm": 1.0177200696682724,
"learning_rate": 9.714659326109138e-07,
"loss": 0.5146,
"step": 466
},
{
"epoch": 0.8076091655858193,
"grad_norm": 0.9855108276847996,
"learning_rate": 9.549150281252633e-07,
"loss": 0.5047,
"step": 467
},
{
"epoch": 0.8093385214007782,
"grad_norm": 0.9820568110711639,
"learning_rate": 9.384914300290749e-07,
"loss": 0.5105,
"step": 468
},
{
"epoch": 0.8110678772157371,
"grad_norm": 0.9543734782446319,
"learning_rate": 9.221956552036992e-07,
"loss": 0.51,
"step": 469
},
{
"epoch": 0.8127972330306961,
"grad_norm": 1.0542711582531745,
"learning_rate": 9.060282165076462e-07,
"loss": 0.5025,
"step": 470
},
{
"epoch": 0.814526588845655,
"grad_norm": 0.8575237177652391,
"learning_rate": 8.899896227604509e-07,
"loss": 0.4907,
"step": 471
},
{
"epoch": 0.8162559446606139,
"grad_norm": 0.9537736768650427,
"learning_rate": 8.740803787266522e-07,
"loss": 0.502,
"step": 472
},
{
"epoch": 0.8179853004755728,
"grad_norm": 0.940840984310344,
"learning_rate": 8.58300985099918e-07,
"loss": 0.5079,
"step": 473
},
{
"epoch": 0.8197146562905318,
"grad_norm": 0.9855222060260419,
"learning_rate": 8.426519384872733e-07,
"loss": 0.5101,
"step": 474
},
{
"epoch": 0.8214440121054907,
"grad_norm": 1.0112701653651899,
"learning_rate": 8.271337313934869e-07,
"loss": 0.5122,
"step": 475
},
{
"epoch": 0.8231733679204496,
"grad_norm": 0.9202094573581726,
"learning_rate": 8.117468522055578e-07,
"loss": 0.5083,
"step": 476
},
{
"epoch": 0.8249027237354085,
"grad_norm": 0.9312213063787499,
"learning_rate": 7.964917851773496e-07,
"loss": 0.5058,
"step": 477
},
{
"epoch": 0.8266320795503674,
"grad_norm": 0.871339607306761,
"learning_rate": 7.813690104143557e-07,
"loss": 0.4966,
"step": 478
},
{
"epoch": 0.8283614353653265,
"grad_norm": 0.8812181772889855,
"learning_rate": 7.663790038585794e-07,
"loss": 0.5021,
"step": 479
},
{
"epoch": 0.8300907911802854,
"grad_norm": 0.9210803401855143,
"learning_rate": 7.515222372735648e-07,
"loss": 0.5084,
"step": 480
},
{
"epoch": 0.8318201469952443,
"grad_norm": 0.9105295776120536,
"learning_rate": 7.367991782295392e-07,
"loss": 0.5066,
"step": 481
},
{
"epoch": 0.8335495028102032,
"grad_norm": 0.9553285353198891,
"learning_rate": 7.222102900887102e-07,
"loss": 0.5137,
"step": 482
},
{
"epoch": 0.8352788586251622,
"grad_norm": 0.8726021151068502,
"learning_rate": 7.077560319906696e-07,
"loss": 0.5172,
"step": 483
},
{
"epoch": 0.8370082144401211,
"grad_norm": 0.907090839860992,
"learning_rate": 6.934368588379553e-07,
"loss": 0.507,
"step": 484
},
{
"epoch": 0.83873757025508,
"grad_norm": 0.9601616142220535,
"learning_rate": 6.792532212817271e-07,
"loss": 0.4932,
"step": 485
},
{
"epoch": 0.8404669260700389,
"grad_norm": 0.9038178937915357,
"learning_rate": 6.652055657075845e-07,
"loss": 0.5158,
"step": 486
},
{
"epoch": 0.8421962818849978,
"grad_norm": 0.8811166725350369,
"learning_rate": 6.512943342215234e-07,
"loss": 0.4991,
"step": 487
},
{
"epoch": 0.8439256376999568,
"grad_norm": 0.910468026382921,
"learning_rate": 6.375199646360142e-07,
"loss": 0.4942,
"step": 488
},
{
"epoch": 0.8456549935149157,
"grad_norm": 0.9415463030818443,
"learning_rate": 6.238828904562316e-07,
"loss": 0.4988,
"step": 489
},
{
"epoch": 0.8473843493298746,
"grad_norm": 0.9306562448394409,
"learning_rate": 6.103835408664032e-07,
"loss": 0.5218,
"step": 490
},
{
"epoch": 0.8491137051448335,
"grad_norm": 0.9306072629397343,
"learning_rate": 5.9702234071631e-07,
"loss": 0.5018,
"step": 491
},
{
"epoch": 0.8508430609597925,
"grad_norm": 0.9978545888527431,
"learning_rate": 5.83799710507909e-07,
"loss": 0.5015,
"step": 492
},
{
"epoch": 0.8525724167747514,
"grad_norm": 0.9281144531309704,
"learning_rate": 5.707160663821009e-07,
"loss": 0.5154,
"step": 493
},
{
"epoch": 0.8543017725897103,
"grad_norm": 0.9648920267544575,
"learning_rate": 5.577718201056392e-07,
"loss": 0.5024,
"step": 494
},
{
"epoch": 0.8560311284046692,
"grad_norm": 0.9518063416993736,
"learning_rate": 5.449673790581611e-07,
"loss": 0.5026,
"step": 495
},
{
"epoch": 0.8577604842196281,
"grad_norm": 0.9090864949437945,
"learning_rate": 5.323031462193757e-07,
"loss": 0.4975,
"step": 496
},
{
"epoch": 0.8594898400345872,
"grad_norm": 0.9027482960362793,
"learning_rate": 5.197795201563744e-07,
"loss": 0.5077,
"step": 497
},
{
"epoch": 0.8612191958495461,
"grad_norm": 0.9825856982352978,
"learning_rate": 5.073968950110941e-07,
"loss": 0.5175,
"step": 498
},
{
"epoch": 0.862948551664505,
"grad_norm": 1.0033645469552896,
"learning_rate": 4.951556604879049e-07,
"loss": 0.5103,
"step": 499
},
{
"epoch": 0.8646779074794639,
"grad_norm": 1.11450811759059,
"learning_rate": 4.830562018413532e-07,
"loss": 0.5135,
"step": 500
},
{
"epoch": 0.8664072632944229,
"grad_norm": 0.9286119478176571,
"learning_rate": 4.710988998640298e-07,
"loss": 0.4853,
"step": 501
},
{
"epoch": 0.8681366191093818,
"grad_norm": 0.9352342724589618,
"learning_rate": 4.5928413087459325e-07,
"loss": 0.5027,
"step": 502
},
{
"epoch": 0.8698659749243407,
"grad_norm": 0.9742580013974624,
"learning_rate": 4.4761226670592074e-07,
"loss": 0.5007,
"step": 503
},
{
"epoch": 0.8715953307392996,
"grad_norm": 0.9536093124243782,
"learning_rate": 4.3608367469340553e-07,
"loss": 0.4894,
"step": 504
},
{
"epoch": 0.8733246865542585,
"grad_norm": 0.9739686798504419,
"learning_rate": 4.2469871766340096e-07,
"loss": 0.5024,
"step": 505
},
{
"epoch": 0.8750540423692175,
"grad_norm": 0.9379743999219975,
"learning_rate": 4.134577539217965e-07,
"loss": 0.4924,
"step": 506
},
{
"epoch": 0.8767833981841764,
"grad_norm": 1.003766963981924,
"learning_rate": 4.0236113724274716e-07,
"loss": 0.4878,
"step": 507
},
{
"epoch": 0.8785127539991353,
"grad_norm": 0.9524223981180386,
"learning_rate": 3.9140921685753065e-07,
"loss": 0.5068,
"step": 508
},
{
"epoch": 0.8802421098140942,
"grad_norm": 0.8535490006491876,
"learning_rate": 3.8060233744356634e-07,
"loss": 0.4838,
"step": 509
},
{
"epoch": 0.8819714656290532,
"grad_norm": 0.9082342591220833,
"learning_rate": 3.6994083911356114e-07,
"loss": 0.5176,
"step": 510
},
{
"epoch": 0.8837008214440121,
"grad_norm": 0.9478217951483088,
"learning_rate": 3.5942505740480583e-07,
"loss": 0.4883,
"step": 511
},
{
"epoch": 0.885430177258971,
"grad_norm": 0.9157515033348981,
"learning_rate": 3.4905532326861946e-07,
"loss": 0.4856,
"step": 512
},
{
"epoch": 0.8871595330739299,
"grad_norm": 0.8748310499093214,
"learning_rate": 3.3883196305992906e-07,
"loss": 0.501,
"step": 513
},
{
"epoch": 0.8888888888888888,
"grad_norm": 0.9579844131883266,
"learning_rate": 3.287552985270015e-07,
"loss": 0.5008,
"step": 514
},
{
"epoch": 0.8906182447038478,
"grad_norm": 0.9866794087631354,
"learning_rate": 3.18825646801314e-07,
"loss": 0.5056,
"step": 515
},
{
"epoch": 0.8923476005188068,
"grad_norm": 0.9637431191512927,
"learning_rate": 3.0904332038757977e-07,
"loss": 0.5165,
"step": 516
},
{
"epoch": 0.8940769563337657,
"grad_norm": 0.9332323606841268,
"learning_rate": 2.9940862715390483e-07,
"loss": 0.5105,
"step": 517
},
{
"epoch": 0.8958063121487246,
"grad_norm": 1.009400230071864,
"learning_rate": 2.899218703221052e-07,
"loss": 0.5128,
"step": 518
},
{
"epoch": 0.8975356679636836,
"grad_norm": 0.8996826766855536,
"learning_rate": 2.8058334845816214e-07,
"loss": 0.5001,
"step": 519
},
{
"epoch": 0.8992650237786425,
"grad_norm": 0.8529312049842479,
"learning_rate": 2.7139335546282287e-07,
"loss": 0.4989,
"step": 520
},
{
"epoch": 0.9009943795936014,
"grad_norm": 0.9524051149908734,
"learning_rate": 2.6235218056235633e-07,
"loss": 0.5085,
"step": 521
},
{
"epoch": 0.9027237354085603,
"grad_norm": 0.9450098346365634,
"learning_rate": 2.534601082994437e-07,
"loss": 0.5041,
"step": 522
},
{
"epoch": 0.9044530912235192,
"grad_norm": 0.9688024639636753,
"learning_rate": 2.447174185242324e-07,
"loss": 0.5139,
"step": 523
},
{
"epoch": 0.9061824470384782,
"grad_norm": 0.9362455552180009,
"learning_rate": 2.3612438638551837e-07,
"loss": 0.4956,
"step": 524
},
{
"epoch": 0.9079118028534371,
"grad_norm": 0.8598614937743414,
"learning_rate": 2.276812823220964e-07,
"loss": 0.4875,
"step": 525
},
{
"epoch": 0.909641158668396,
"grad_norm": 0.909540077460528,
"learning_rate": 2.1938837205424002e-07,
"loss": 0.5117,
"step": 526
},
{
"epoch": 0.9113705144833549,
"grad_norm": 0.9096506960895155,
"learning_rate": 2.1124591657534776e-07,
"loss": 0.4955,
"step": 527
},
{
"epoch": 0.9130998702983139,
"grad_norm": 1.0142474014859533,
"learning_rate": 2.032541721437209e-07,
"loss": 0.507,
"step": 528
},
{
"epoch": 0.9148292261132728,
"grad_norm": 0.9131442997723532,
"learning_rate": 1.9541339027450256e-07,
"loss": 0.4895,
"step": 529
},
{
"epoch": 0.9165585819282317,
"grad_norm": 0.9853142920778307,
"learning_rate": 1.8772381773176417e-07,
"loss": 0.5211,
"step": 530
},
{
"epoch": 0.9182879377431906,
"grad_norm": 0.8974474768112111,
"learning_rate": 1.801856965207338e-07,
"loss": 0.4976,
"step": 531
},
{
"epoch": 0.9200172935581495,
"grad_norm": 0.9550109239119172,
"learning_rate": 1.7279926388018564e-07,
"loss": 0.5151,
"step": 532
},
{
"epoch": 0.9217466493731085,
"grad_norm": 0.9072672003349709,
"learning_rate": 1.6556475227496816e-07,
"loss": 0.516,
"step": 533
},
{
"epoch": 0.9234760051880675,
"grad_norm": 0.9399754149330305,
"learning_rate": 1.5848238938869332e-07,
"loss": 0.4909,
"step": 534
},
{
"epoch": 0.9252053610030264,
"grad_norm": 0.9437594133784243,
"learning_rate": 1.5155239811656562e-07,
"loss": 0.498,
"step": 535
},
{
"epoch": 0.9269347168179853,
"grad_norm": 0.9357434836371697,
"learning_rate": 1.4477499655837278e-07,
"loss": 0.4969,
"step": 536
},
{
"epoch": 0.9286640726329443,
"grad_norm": 0.8936101786239388,
"learning_rate": 1.3815039801161723e-07,
"loss": 0.4913,
"step": 537
},
{
"epoch": 0.9303934284479032,
"grad_norm": 0.9484054108177385,
"learning_rate": 1.3167881096480372e-07,
"loss": 0.4988,
"step": 538
},
{
"epoch": 0.9321227842628621,
"grad_norm": 0.9563796786443849,
"learning_rate": 1.253604390908819e-07,
"loss": 0.4925,
"step": 539
},
{
"epoch": 0.933852140077821,
"grad_norm": 0.8950613019917822,
"learning_rate": 1.191954812408308e-07,
"loss": 0.5052,
"step": 540
},
{
"epoch": 0.93558149589278,
"grad_norm": 0.9125279893925492,
"learning_rate": 1.1318413143740436e-07,
"loss": 0.5041,
"step": 541
},
{
"epoch": 0.9373108517077389,
"grad_norm": 0.9141300635124516,
"learning_rate": 1.0732657886902308e-07,
"loss": 0.5096,
"step": 542
},
{
"epoch": 0.9390402075226978,
"grad_norm": 0.9096167507969538,
"learning_rate": 1.0162300788382263e-07,
"loss": 0.4882,
"step": 543
},
{
"epoch": 0.9407695633376567,
"grad_norm": 0.9753044785021144,
"learning_rate": 9.607359798384785e-08,
"loss": 0.5103,
"step": 544
},
{
"epoch": 0.9424989191526156,
"grad_norm": 0.8747544695243777,
"learning_rate": 9.0678523819408e-08,
"loss": 0.504,
"step": 545
},
{
"epoch": 0.9442282749675746,
"grad_norm": 0.8448600329999801,
"learning_rate": 8.543795518357767e-08,
"loss": 0.5049,
"step": 546
},
{
"epoch": 0.9459576307825335,
"grad_norm": 0.9445491960031619,
"learning_rate": 8.035205700685167e-08,
"loss": 0.5033,
"step": 547
},
{
"epoch": 0.9476869865974924,
"grad_norm": 0.9108140075113013,
"learning_rate": 7.542098935195918e-08,
"loss": 0.5138,
"step": 548
},
{
"epoch": 0.9494163424124513,
"grad_norm": 0.9014111637252965,
"learning_rate": 7.064490740882057e-08,
"loss": 0.505,
"step": 549
},
{
"epoch": 0.9511456982274102,
"grad_norm": 0.8618772410756893,
"learning_rate": 6.602396148966795e-08,
"loss": 0.4939,
"step": 550
},
{
"epoch": 0.9528750540423692,
"grad_norm": 0.9034604742193665,
"learning_rate": 6.15582970243117e-08,
"loss": 0.4856,
"step": 551
},
{
"epoch": 0.9546044098573282,
"grad_norm": 0.8858115334722476,
"learning_rate": 5.7248054555563704e-08,
"loss": 0.4943,
"step": 552
},
{
"epoch": 0.9563337656722871,
"grad_norm": 0.9413567454097705,
"learning_rate": 5.3093369734816824e-08,
"loss": 0.4983,
"step": 553
},
{
"epoch": 0.958063121487246,
"grad_norm": 0.8631416544061653,
"learning_rate": 4.909437331777178e-08,
"loss": 0.4991,
"step": 554
},
{
"epoch": 0.959792477302205,
"grad_norm": 0.8972216946376435,
"learning_rate": 4.52511911603265e-08,
"loss": 0.4979,
"step": 555
},
{
"epoch": 0.9615218331171639,
"grad_norm": 0.9017046495091536,
"learning_rate": 4.15639442146093e-08,
"loss": 0.4968,
"step": 556
},
{
"epoch": 0.9632511889321228,
"grad_norm": 0.8883293939758925,
"learning_rate": 3.8032748525179684e-08,
"loss": 0.4959,
"step": 557
},
{
"epoch": 0.9649805447470817,
"grad_norm": 0.9234481550882554,
"learning_rate": 3.465771522536854e-08,
"loss": 0.5218,
"step": 558
},
{
"epoch": 0.9667099005620406,
"grad_norm": 0.914125086466903,
"learning_rate": 3.143895053378698e-08,
"loss": 0.5087,
"step": 559
},
{
"epoch": 0.9684392563769996,
"grad_norm": 0.8985893082923397,
"learning_rate": 2.837655575097964e-08,
"loss": 0.5053,
"step": 560
},
{
"epoch": 0.9701686121919585,
"grad_norm": 0.9385196676766113,
"learning_rate": 2.547062725623828e-08,
"loss": 0.5116,
"step": 561
},
{
"epoch": 0.9718979680069174,
"grad_norm": 0.9646239444487571,
"learning_rate": 2.2721256504567026e-08,
"loss": 0.5072,
"step": 562
},
{
"epoch": 0.9736273238218763,
"grad_norm": 0.8747663830607033,
"learning_rate": 2.012853002380466e-08,
"loss": 0.5022,
"step": 563
},
{
"epoch": 0.9753566796368353,
"grad_norm": 0.9334336705435727,
"learning_rate": 1.769252941190458e-08,
"loss": 0.5105,
"step": 564
},
{
"epoch": 0.9770860354517942,
"grad_norm": 0.9194588921972738,
"learning_rate": 1.541333133436018e-08,
"loss": 0.4985,
"step": 565
},
{
"epoch": 0.9788153912667531,
"grad_norm": 0.9520800525639294,
"learning_rate": 1.3291007521799015e-08,
"loss": 0.4977,
"step": 566
},
{
"epoch": 0.980544747081712,
"grad_norm": 0.9058630469142649,
"learning_rate": 1.132562476771959e-08,
"loss": 0.4738,
"step": 567
},
{
"epoch": 0.9822741028966709,
"grad_norm": 0.8879281633731727,
"learning_rate": 9.51724492639361e-09,
"loss": 0.5048,
"step": 568
},
{
"epoch": 0.9840034587116299,
"grad_norm": 0.8984029908537556,
"learning_rate": 7.865924910916977e-09,
"loss": 0.5024,
"step": 569
},
{
"epoch": 0.9857328145265889,
"grad_norm": 0.9121742433892018,
"learning_rate": 6.371716691419005e-09,
"loss": 0.5082,
"step": 570
},
{
"epoch": 0.9874621703415478,
"grad_norm": 0.9237443595386216,
"learning_rate": 5.034667293427053e-09,
"loss": 0.5097,
"step": 571
},
{
"epoch": 0.9891915261565067,
"grad_norm": 0.9117600237724628,
"learning_rate": 3.854818796385495e-09,
"loss": 0.5129,
"step": 572
},
{
"epoch": 0.9909208819714657,
"grad_norm": 0.9527043115093743,
"learning_rate": 2.8322083323334417e-09,
"loss": 0.5024,
"step": 573
},
{
"epoch": 0.9926502377864246,
"grad_norm": 0.9724377615202221,
"learning_rate": 1.9668680847356735e-09,
"loss": 0.4966,
"step": 574
},
{
"epoch": 0.9943795936013835,
"grad_norm": 0.9737523393148931,
"learning_rate": 1.2588252874673469e-09,
"loss": 0.4976,
"step": 575
},
{
"epoch": 0.9961089494163424,
"grad_norm": 0.8474552307580481,
"learning_rate": 7.081022239591173e-10,
"loss": 0.4907,
"step": 576
},
{
"epoch": 0.9978383052313013,
"grad_norm": 0.8918148859611189,
"learning_rate": 3.147162264971471e-10,
"loss": 0.4957,
"step": 577
},
{
"epoch": 0.9995676610462603,
"grad_norm": 0.8738364123206687,
"learning_rate": 7.867967567354306e-11,
"loss": 0.4892,
"step": 578
},
{
"epoch": 0.9995676610462603,
"step": 578,
"total_flos": 6.685487298542305e+17,
"train_loss": 0.5686832449324816,
"train_runtime": 23812.8313,
"train_samples_per_second": 7.141,
"train_steps_per_second": 0.024
}
],
"logging_steps": 1.0,
"max_steps": 578,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 50,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 6.685487298542305e+17,
"train_batch_size": 4,
"trial_name": null,
"trial_params": null
}