Text Generation
Transformers
PyTorch
Safetensors
Swedish
ctrl
Inference Endpoints
SweCTRL-Mini / trainer_state.json
dkalpakchi's picture
Uploaded model, tokenizer and the minimally necessary code
59392d8
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 28.5,
"global_step": 2786844,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.1,
"learning_rate": 0.020000000000000004,
"loss": 5.3687,
"step": 10000
},
{
"epoch": 0.2,
"learning_rate": 0.04000000000000001,
"loss": 4.0531,
"step": 20000
},
{
"epoch": 0.31,
"learning_rate": 0.04993567245443037,
"loss": 3.8149,
"step": 30000
},
{
"epoch": 0.41,
"learning_rate": 0.0498070173632911,
"loss": 3.6549,
"step": 40000
},
{
"epoch": 0.51,
"learning_rate": 0.04967836227215183,
"loss": 3.5544,
"step": 50000
},
{
"epoch": 0.61,
"learning_rate": 0.049549707181012564,
"loss": 3.4935,
"step": 60000
},
{
"epoch": 0.72,
"learning_rate": 0.0494210520898733,
"loss": 3.447,
"step": 70000
},
{
"epoch": 0.82,
"learning_rate": 0.049292396998734035,
"loss": 3.4134,
"step": 80000
},
{
"epoch": 0.92,
"learning_rate": 0.04916374190759477,
"loss": 3.3861,
"step": 90000
},
{
"epoch": 1.02,
"learning_rate": 0.049035086816455506,
"loss": 3.3523,
"step": 100000
},
{
"epoch": 1.12,
"learning_rate": 0.04890643172531624,
"loss": 3.3063,
"step": 110000
},
{
"epoch": 1.23,
"learning_rate": 0.04877777663417697,
"loss": 3.2983,
"step": 120000
},
{
"epoch": 1.33,
"learning_rate": 0.0486491215430377,
"loss": 3.2874,
"step": 130000
},
{
"epoch": 1.43,
"learning_rate": 0.04852046645189844,
"loss": 3.2785,
"step": 140000
},
{
"epoch": 1.53,
"learning_rate": 0.04839181136075917,
"loss": 3.2698,
"step": 150000
},
{
"epoch": 1.64,
"learning_rate": 0.048263156269619904,
"loss": 3.2599,
"step": 160000
},
{
"epoch": 1.74,
"learning_rate": 0.04813450117848064,
"loss": 3.2503,
"step": 170000
},
{
"epoch": 1.84,
"learning_rate": 0.048005846087341375,
"loss": 3.2432,
"step": 180000
},
{
"epoch": 1.94,
"learning_rate": 0.04787719099620211,
"loss": 3.2347,
"step": 190000
},
{
"epoch": 2.05,
"learning_rate": 0.04774853590506284,
"loss": 3.1812,
"step": 200000
},
{
"epoch": 2.15,
"learning_rate": 0.04761988081392357,
"loss": 3.1865,
"step": 210000
},
{
"epoch": 2.25,
"learning_rate": 0.0474912257227843,
"loss": 3.1873,
"step": 220000
},
{
"epoch": 2.35,
"learning_rate": 0.047362570631645035,
"loss": 3.1842,
"step": 230000
},
{
"epoch": 2.45,
"learning_rate": 0.047233915540505766,
"loss": 3.1824,
"step": 240000
},
{
"epoch": 2.56,
"learning_rate": 0.047105260449366505,
"loss": 3.1806,
"step": 250000
},
{
"epoch": 2.66,
"learning_rate": 0.04697660535822724,
"loss": 3.177,
"step": 260000
},
{
"epoch": 2.76,
"learning_rate": 0.04684795026708797,
"loss": 3.1741,
"step": 270000
},
{
"epoch": 2.86,
"learning_rate": 0.04671929517594871,
"loss": 3.1709,
"step": 280000
},
{
"epoch": 2.97,
"learning_rate": 0.04659064008480944,
"loss": 3.1682,
"step": 290000
},
{
"epoch": 3.07,
"learning_rate": 0.04646198499367017,
"loss": 3.1382,
"step": 300000
},
{
"epoch": 3.17,
"learning_rate": 0.046333329902530904,
"loss": 3.1283,
"step": 310000
},
{
"epoch": 3.27,
"learning_rate": 0.04620467481139164,
"loss": 3.1318,
"step": 320000
},
{
"epoch": 3.37,
"learning_rate": 0.046076019720252374,
"loss": 3.1319,
"step": 330000
},
{
"epoch": 3.48,
"learning_rate": 0.045947364629113106,
"loss": 3.1335,
"step": 340000
},
{
"epoch": 3.58,
"learning_rate": 0.04581870953797384,
"loss": 3.1326,
"step": 350000
},
{
"epoch": 3.68,
"learning_rate": 0.04569005444683458,
"loss": 3.1306,
"step": 360000
},
{
"epoch": 3.78,
"learning_rate": 0.04556139935569531,
"loss": 3.1289,
"step": 370000
},
{
"epoch": 3.89,
"learning_rate": 0.04543274426455604,
"loss": 3.1275,
"step": 380000
},
{
"epoch": 3.99,
"learning_rate": 0.04530408917341677,
"loss": 3.1259,
"step": 390000
},
{
"epoch": 4.09,
"learning_rate": 0.045175434082277505,
"loss": 3.0882,
"step": 400000
},
{
"epoch": 4.19,
"learning_rate": 0.04504677899113824,
"loss": 3.094,
"step": 410000
},
{
"epoch": 4.3,
"learning_rate": 0.04491812389999897,
"loss": 3.0958,
"step": 420000
},
{
"epoch": 4.4,
"learning_rate": 0.04478946880885971,
"loss": 3.0976,
"step": 430000
},
{
"epoch": 4.5,
"learning_rate": 0.04466081371772044,
"loss": 3.0996,
"step": 440000
},
{
"epoch": 4.6,
"learning_rate": 0.04453215862658117,
"loss": 3.099,
"step": 450000
},
{
"epoch": 4.7,
"learning_rate": 0.04440350353544191,
"loss": 3.1,
"step": 460000
},
{
"epoch": 4.81,
"learning_rate": 0.04427484844430264,
"loss": 3.0993,
"step": 470000
},
{
"epoch": 4.91,
"learning_rate": 0.044146193353163374,
"loss": 3.0985,
"step": 480000
},
{
"epoch": 5.01,
"learning_rate": 0.044017538262024106,
"loss": 3.094,
"step": 490000
},
{
"epoch": 5.11,
"learning_rate": 0.043888883170884845,
"loss": 3.0646,
"step": 500000
},
{
"epoch": 5.22,
"learning_rate": 0.043760228079745576,
"loss": 3.0694,
"step": 510000
},
{
"epoch": 5.32,
"learning_rate": 0.04363157298860631,
"loss": 3.0712,
"step": 520000
},
{
"epoch": 5.42,
"learning_rate": 0.04350291789746704,
"loss": 3.0736,
"step": 530000
},
{
"epoch": 5.52,
"learning_rate": 0.04337426280632778,
"loss": 3.0752,
"step": 540000
},
{
"epoch": 5.62,
"learning_rate": 0.04324560771518851,
"loss": 3.0759,
"step": 550000
},
{
"epoch": 5.73,
"learning_rate": 0.04311695262404924,
"loss": 3.0762,
"step": 560000
},
{
"epoch": 5.83,
"learning_rate": 0.042988297532909975,
"loss": 3.0756,
"step": 570000
},
{
"epoch": 5.93,
"learning_rate": 0.04285964244177071,
"loss": 3.076,
"step": 580000
},
{
"epoch": 6.03,
"learning_rate": 0.04273098735063144,
"loss": 3.0412,
"step": 590000
},
{
"epoch": 6.14,
"learning_rate": 0.04260233225949217,
"loss": 3.0464,
"step": 600000
},
{
"epoch": 6.24,
"learning_rate": 0.04247367716835291,
"loss": 3.05,
"step": 610000
},
{
"epoch": 6.34,
"learning_rate": 0.04234502207721364,
"loss": 3.0539,
"step": 620000
},
{
"epoch": 6.44,
"learning_rate": 0.04221636698607437,
"loss": 3.0554,
"step": 630000
},
{
"epoch": 6.55,
"learning_rate": 0.042087711894935105,
"loss": 3.0558,
"step": 640000
},
{
"epoch": 6.65,
"learning_rate": 0.041959056803795844,
"loss": 3.057,
"step": 650000
},
{
"epoch": 6.75,
"learning_rate": 0.041830401712656576,
"loss": 3.0584,
"step": 660000
},
{
"epoch": 6.85,
"learning_rate": 0.04170174662151731,
"loss": 3.0585,
"step": 670000
},
{
"epoch": 6.95,
"learning_rate": 0.04157309153037805,
"loss": 3.0593,
"step": 680000
},
{
"epoch": 7.06,
"learning_rate": 0.04144443643923878,
"loss": 3.0408,
"step": 690000
},
{
"epoch": 7.16,
"learning_rate": 0.04131578134809951,
"loss": 3.0325,
"step": 700000
},
{
"epoch": 7.26,
"learning_rate": 0.04118712625696024,
"loss": 3.035,
"step": 710000
},
{
"epoch": 7.36,
"learning_rate": 0.04105847116582098,
"loss": 3.0373,
"step": 720000
},
{
"epoch": 7.47,
"learning_rate": 0.04092981607468171,
"loss": 3.0405,
"step": 730000
},
{
"epoch": 7.57,
"learning_rate": 0.040801160983542445,
"loss": 3.0403,
"step": 740000
},
{
"epoch": 7.67,
"learning_rate": 0.04067250589240318,
"loss": 3.0431,
"step": 750000
},
{
"epoch": 7.77,
"learning_rate": 0.04054385080126391,
"loss": 3.0444,
"step": 760000
},
{
"epoch": 7.87,
"learning_rate": 0.04041519571012464,
"loss": 3.0445,
"step": 770000
},
{
"epoch": 7.98,
"learning_rate": 0.04028654061898537,
"loss": 3.0452,
"step": 780000
},
{
"epoch": 8.08,
"learning_rate": 0.04015788552784611,
"loss": 3.0217,
"step": 790000
},
{
"epoch": 8.18,
"learning_rate": 0.04002923043670684,
"loss": 3.02,
"step": 800000
},
{
"epoch": 8.28,
"learning_rate": 0.039900575345567575,
"loss": 3.0233,
"step": 810000
},
{
"epoch": 8.39,
"learning_rate": 0.03977192025442831,
"loss": 3.0259,
"step": 820000
},
{
"epoch": 8.49,
"learning_rate": 0.039643265163289046,
"loss": 3.0271,
"step": 830000
},
{
"epoch": 8.59,
"learning_rate": 0.03951461007214978,
"loss": 3.0121,
"step": 840000
},
{
"epoch": 8.69,
"learning_rate": 0.03938595498101051,
"loss": 3.0161,
"step": 850000
},
{
"epoch": 8.79,
"learning_rate": 0.03925729988987125,
"loss": 3.0195,
"step": 860000
},
{
"epoch": 8.9,
"learning_rate": 0.03912864479873198,
"loss": 3.021,
"step": 870000
},
{
"epoch": 9.0,
"learning_rate": 0.03899998970759271,
"loss": 3.0228,
"step": 880000
},
{
"epoch": 9.1,
"learning_rate": 0.038871334616453444,
"loss": 3.0073,
"step": 890000
},
{
"epoch": 9.2,
"learning_rate": 0.03874267952531418,
"loss": 3.0114,
"step": 900000
},
{
"epoch": 9.31,
"learning_rate": 0.038614024434174915,
"loss": 3.0151,
"step": 910000
},
{
"epoch": 9.41,
"learning_rate": 0.03848536934303565,
"loss": 3.0175,
"step": 920000
},
{
"epoch": 9.51,
"learning_rate": 0.03835671425189638,
"loss": 3.0193,
"step": 930000
},
{
"epoch": 9.61,
"learning_rate": 0.03822805916075711,
"loss": 3.0185,
"step": 940000
},
{
"epoch": 9.72,
"learning_rate": 0.03809940406961784,
"loss": 3.0228,
"step": 950000
},
{
"epoch": 9.82,
"learning_rate": 0.037970748978478575,
"loss": 3.0226,
"step": 960000
},
{
"epoch": 9.92,
"learning_rate": 0.03784209388733931,
"loss": 3.0227,
"step": 970000
},
{
"epoch": 10.02,
"learning_rate": 0.037713438796200045,
"loss": 3.017,
"step": 980000
},
{
"epoch": 10.12,
"learning_rate": 0.03758478370506078,
"loss": 2.9992,
"step": 990000
},
{
"epoch": 10.23,
"learning_rate": 0.03745612861392151,
"loss": 3.0007,
"step": 1000000
},
{
"epoch": 10.33,
"learning_rate": 0.03732747352278225,
"loss": 3.0047,
"step": 1010000
},
{
"epoch": 10.43,
"learning_rate": 0.03719881843164298,
"loss": 3.0075,
"step": 1020000
},
{
"epoch": 10.53,
"learning_rate": 0.03707016334050371,
"loss": 3.0095,
"step": 1030000
},
{
"epoch": 10.64,
"learning_rate": 0.03694150824936445,
"loss": 3.0101,
"step": 1040000
},
{
"epoch": 10.74,
"learning_rate": 0.03681285315822518,
"loss": 3.0123,
"step": 1050000
},
{
"epoch": 10.84,
"learning_rate": 0.036684198067085914,
"loss": 3.013,
"step": 1060000
},
{
"epoch": 10.94,
"learning_rate": 0.036555542975946646,
"loss": 3.0121,
"step": 1070000
},
{
"epoch": 11.04,
"learning_rate": 0.03642688788480738,
"loss": 2.9892,
"step": 1080000
},
{
"epoch": 11.15,
"learning_rate": 0.03629823279366811,
"loss": 2.9908,
"step": 1090000
},
{
"epoch": 11.25,
"learning_rate": 0.03616957770252884,
"loss": 2.9941,
"step": 1100000
},
{
"epoch": 11.35,
"learning_rate": 0.036040922611389574,
"loss": 2.9976,
"step": 1110000
},
{
"epoch": 11.45,
"learning_rate": 0.03591226752025031,
"loss": 2.9992,
"step": 1120000
},
{
"epoch": 11.56,
"learning_rate": 0.035783612429111045,
"loss": 3.0007,
"step": 1130000
},
{
"epoch": 11.66,
"learning_rate": 0.03565495733797178,
"loss": 3.0021,
"step": 1140000
},
{
"epoch": 11.76,
"learning_rate": 0.035526302246832515,
"loss": 3.0032,
"step": 1150000
},
{
"epoch": 11.86,
"learning_rate": 0.03539764715569325,
"loss": 3.0057,
"step": 1160000
},
{
"epoch": 11.97,
"learning_rate": 0.03526899206455398,
"loss": 3.0053,
"step": 1170000
},
{
"epoch": 12.07,
"learning_rate": 0.03514033697341471,
"loss": 2.9898,
"step": 1180000
},
{
"epoch": 12.17,
"learning_rate": 0.03501168188227545,
"loss": 2.9848,
"step": 1190000
},
{
"epoch": 12.27,
"learning_rate": 0.03488302679113618,
"loss": 2.9871,
"step": 1200000
},
{
"epoch": 12.37,
"learning_rate": 0.034754371699996914,
"loss": 2.9903,
"step": 1210000
},
{
"epoch": 12.48,
"learning_rate": 0.03462571660885765,
"loss": 2.9918,
"step": 1220000
},
{
"epoch": 12.58,
"learning_rate": 0.034497061517718385,
"loss": 2.9948,
"step": 1230000
},
{
"epoch": 12.68,
"learning_rate": 0.034368406426579116,
"loss": 2.9955,
"step": 1240000
},
{
"epoch": 12.78,
"learning_rate": 0.03423975133543985,
"loss": 2.9971,
"step": 1250000
},
{
"epoch": 12.89,
"learning_rate": 0.03411109624430058,
"loss": 2.9978,
"step": 1260000
},
{
"epoch": 12.99,
"learning_rate": 0.03398244115316131,
"loss": 2.9985,
"step": 1270000
},
{
"epoch": 13.09,
"learning_rate": 0.033853786062022044,
"loss": 2.9789,
"step": 1280000
},
{
"epoch": 13.19,
"learning_rate": 0.033725130970882776,
"loss": 2.9795,
"step": 1290000
},
{
"epoch": 13.29,
"learning_rate": 0.033596475879743515,
"loss": 2.9835,
"step": 1300000
},
{
"epoch": 13.4,
"learning_rate": 0.03346782078860425,
"loss": 2.9829,
"step": 1310000
},
{
"epoch": 13.5,
"learning_rate": 0.03333916569746498,
"loss": 2.9869,
"step": 1320000
},
{
"epoch": 13.6,
"learning_rate": 0.03321051060632572,
"loss": 2.9755,
"step": 1330000
},
{
"epoch": 13.7,
"learning_rate": 0.03308185551518645,
"loss": 2.978,
"step": 1340000
},
{
"epoch": 13.81,
"learning_rate": 0.03295320042404718,
"loss": 2.9811,
"step": 1350000
},
{
"epoch": 13.91,
"learning_rate": 0.03282454533290791,
"loss": 2.9824,
"step": 1360000
},
{
"epoch": 14.01,
"learning_rate": 0.03269589024176865,
"loss": 2.9832,
"step": 1370000
},
{
"epoch": 14.11,
"learning_rate": 0.032567235150629384,
"loss": 2.9726,
"step": 1380000
},
{
"epoch": 14.22,
"learning_rate": 0.032438580059490116,
"loss": 2.9762,
"step": 1390000
},
{
"epoch": 14.32,
"learning_rate": 0.03230992496835085,
"loss": 2.9786,
"step": 1400000
},
{
"epoch": 14.42,
"learning_rate": 0.03218126987721159,
"loss": 2.9804,
"step": 1410000
},
{
"epoch": 14.52,
"learning_rate": 0.03205261478607232,
"loss": 2.9821,
"step": 1420000
},
{
"epoch": 14.62,
"learning_rate": 0.03192395969493305,
"loss": 2.9825,
"step": 1430000
},
{
"epoch": 14.73,
"learning_rate": 0.03179530460379378,
"loss": 2.985,
"step": 1440000
},
{
"epoch": 14.83,
"learning_rate": 0.031666649512654514,
"loss": 2.9851,
"step": 1450000
},
{
"epoch": 14.93,
"learning_rate": 0.031537994421515246,
"loss": 2.9859,
"step": 1460000
},
{
"epoch": 15.03,
"learning_rate": 0.03140933933037598,
"loss": 2.9795,
"step": 1470000
},
{
"epoch": 15.14,
"learning_rate": 0.03128068423923672,
"loss": 2.9681,
"step": 1480000
},
{
"epoch": 15.24,
"learning_rate": 0.03115202914809745,
"loss": 2.9707,
"step": 1490000
},
{
"epoch": 15.34,
"learning_rate": 0.03102337405695818,
"loss": 2.9727,
"step": 1500000
},
{
"epoch": 15.44,
"learning_rate": 0.03089471896581892,
"loss": 2.9747,
"step": 1510000
},
{
"epoch": 15.54,
"learning_rate": 0.03076606387467965,
"loss": 2.9769,
"step": 1520000
},
{
"epoch": 15.65,
"learning_rate": 0.030637408783540383,
"loss": 2.9778,
"step": 1530000
},
{
"epoch": 15.75,
"learning_rate": 0.030508753692401115,
"loss": 2.9788,
"step": 1540000
},
{
"epoch": 15.85,
"learning_rate": 0.030380098601261854,
"loss": 2.9789,
"step": 1550000
},
{
"epoch": 15.95,
"learning_rate": 0.030251443510122586,
"loss": 2.9807,
"step": 1560000
},
{
"epoch": 16.06,
"learning_rate": 0.030122788418983318,
"loss": 2.9619,
"step": 1570000
},
{
"epoch": 16.16,
"learning_rate": 0.02999413332784405,
"loss": 2.9638,
"step": 1580000
},
{
"epoch": 16.26,
"learning_rate": 0.029865478236704785,
"loss": 2.9654,
"step": 1590000
},
{
"epoch": 16.36,
"learning_rate": 0.029736823145565517,
"loss": 2.9679,
"step": 1600000
},
{
"epoch": 16.46,
"learning_rate": 0.02960816805442625,
"loss": 2.9704,
"step": 1610000
},
{
"epoch": 16.57,
"learning_rate": 0.029479512963286988,
"loss": 2.9726,
"step": 1620000
},
{
"epoch": 16.67,
"learning_rate": 0.02935085787214772,
"loss": 2.9725,
"step": 1630000
},
{
"epoch": 16.77,
"learning_rate": 0.02922220278100845,
"loss": 2.9738,
"step": 1640000
},
{
"epoch": 16.87,
"learning_rate": 0.029093547689869183,
"loss": 2.9747,
"step": 1650000
},
{
"epoch": 16.98,
"learning_rate": 0.02896489259872992,
"loss": 2.9763,
"step": 1660000
},
{
"epoch": 17.08,
"learning_rate": 0.02883623750759065,
"loss": 2.9617,
"step": 1670000
},
{
"epoch": 17.18,
"learning_rate": 0.028707582416451383,
"loss": 2.9618,
"step": 1680000
},
{
"epoch": 17.28,
"learning_rate": 0.028578927325312115,
"loss": 2.9612,
"step": 1690000
},
{
"epoch": 17.39,
"learning_rate": 0.028450272234172853,
"loss": 2.9632,
"step": 1700000
},
{
"epoch": 17.49,
"learning_rate": 0.028321617143033585,
"loss": 2.9655,
"step": 1710000
},
{
"epoch": 17.59,
"learning_rate": 0.028192962051894317,
"loss": 2.9672,
"step": 1720000
},
{
"epoch": 17.69,
"learning_rate": 0.028064306960755056,
"loss": 2.9691,
"step": 1730000
},
{
"epoch": 17.79,
"learning_rate": 0.027935651869615788,
"loss": 2.9698,
"step": 1740000
},
{
"epoch": 17.9,
"learning_rate": 0.02780699677847652,
"loss": 2.9702,
"step": 1750000
},
{
"epoch": 18.0,
"learning_rate": 0.027678341687337252,
"loss": 2.9719,
"step": 1760000
},
{
"epoch": 18.1,
"learning_rate": 0.027549686596197987,
"loss": 2.9546,
"step": 1770000
},
{
"epoch": 18.2,
"learning_rate": 0.02742103150505872,
"loss": 2.9567,
"step": 1780000
},
{
"epoch": 18.31,
"learning_rate": 0.02729237641391945,
"loss": 2.9586,
"step": 1790000
},
{
"epoch": 18.41,
"learning_rate": 0.02716372132278019,
"loss": 2.9606,
"step": 1800000
},
{
"epoch": 18.51,
"learning_rate": 0.02703506623164092,
"loss": 2.9506,
"step": 1810000
},
{
"epoch": 18.61,
"learning_rate": 0.026906411140501654,
"loss": 2.9518,
"step": 1820000
},
{
"epoch": 18.71,
"learning_rate": 0.026777756049362385,
"loss": 2.9575,
"step": 1830000
},
{
"epoch": 18.82,
"learning_rate": 0.02664910095822312,
"loss": 2.9584,
"step": 1840000
},
{
"epoch": 18.92,
"learning_rate": 0.026520445867083853,
"loss": 2.9594,
"step": 1850000
},
{
"epoch": 19.02,
"learning_rate": 0.026391790775944585,
"loss": 2.9578,
"step": 1860000
},
{
"epoch": 19.12,
"learning_rate": 0.026263135684805317,
"loss": 2.9535,
"step": 1870000
},
{
"epoch": 19.23,
"learning_rate": 0.026134480593666055,
"loss": 2.9552,
"step": 1880000
},
{
"epoch": 19.33,
"learning_rate": 0.026005825502526787,
"loss": 2.9568,
"step": 1890000
},
{
"epoch": 19.43,
"learning_rate": 0.02587717041138752,
"loss": 2.9583,
"step": 1900000
},
{
"epoch": 19.53,
"learning_rate": 0.025748515320248258,
"loss": 2.9596,
"step": 1910000
},
{
"epoch": 19.64,
"learning_rate": 0.02561986022910899,
"loss": 2.9599,
"step": 1920000
},
{
"epoch": 19.74,
"learning_rate": 0.025491205137969722,
"loss": 2.9615,
"step": 1930000
},
{
"epoch": 19.84,
"learning_rate": 0.025362550046830454,
"loss": 2.9622,
"step": 1940000
},
{
"epoch": 19.94,
"learning_rate": 0.02523389495569119,
"loss": 2.9637,
"step": 1950000
},
{
"epoch": 20.04,
"learning_rate": 0.02510523986455192,
"loss": 2.9559,
"step": 1960000
},
{
"epoch": 20.15,
"learning_rate": 0.024976584773412653,
"loss": 2.9496,
"step": 1970000
},
{
"epoch": 20.25,
"learning_rate": 0.02484792968227339,
"loss": 2.9516,
"step": 1980000
},
{
"epoch": 20.35,
"learning_rate": 0.024719274591134124,
"loss": 2.9522,
"step": 1990000
},
{
"epoch": 20.45,
"learning_rate": 0.024590619499994856,
"loss": 2.9538,
"step": 2000000
},
{
"epoch": 20.56,
"learning_rate": 0.02446196440885559,
"loss": 2.9552,
"step": 2010000
},
{
"epoch": 20.66,
"learning_rate": 0.024333309317716323,
"loss": 2.957,
"step": 2020000
},
{
"epoch": 20.76,
"learning_rate": 0.024204654226577055,
"loss": 2.9572,
"step": 2030000
},
{
"epoch": 20.86,
"learning_rate": 0.024075999135437787,
"loss": 2.9586,
"step": 2040000
},
{
"epoch": 20.96,
"learning_rate": 0.023947344044298522,
"loss": 2.9609,
"step": 2050000
},
{
"epoch": 21.07,
"learning_rate": 0.023818688953159254,
"loss": 2.9429,
"step": 2060000
},
{
"epoch": 21.17,
"learning_rate": 0.02369003386201999,
"loss": 2.947,
"step": 2070000
},
{
"epoch": 21.27,
"learning_rate": 0.023561378770880725,
"loss": 2.9488,
"step": 2080000
},
{
"epoch": 21.37,
"learning_rate": 0.023432723679741457,
"loss": 2.9491,
"step": 2090000
},
{
"epoch": 21.48,
"learning_rate": 0.023304068588602192,
"loss": 2.9514,
"step": 2100000
},
{
"epoch": 21.58,
"learning_rate": 0.023175413497462924,
"loss": 2.9513,
"step": 2110000
},
{
"epoch": 21.68,
"learning_rate": 0.023046758406323656,
"loss": 2.9537,
"step": 2120000
},
{
"epoch": 21.78,
"learning_rate": 0.022918103315184388,
"loss": 2.9548,
"step": 2130000
},
{
"epoch": 21.88,
"learning_rate": 0.022789448224045123,
"loss": 2.9573,
"step": 2140000
},
{
"epoch": 21.99,
"learning_rate": 0.022660793132905855,
"loss": 2.9565,
"step": 2150000
},
{
"epoch": 22.09,
"learning_rate": 0.02253213804176659,
"loss": 2.9432,
"step": 2160000
},
{
"epoch": 22.19,
"learning_rate": 0.022403482950627322,
"loss": 2.9445,
"step": 2170000
},
{
"epoch": 22.29,
"learning_rate": 0.022274827859488058,
"loss": 2.9465,
"step": 2180000
},
{
"epoch": 22.4,
"learning_rate": 0.022146172768348793,
"loss": 2.9478,
"step": 2190000
},
{
"epoch": 22.5,
"learning_rate": 0.022017517677209525,
"loss": 2.9474,
"step": 2200000
},
{
"epoch": 22.6,
"learning_rate": 0.021888862586070257,
"loss": 2.9497,
"step": 2210000
},
{
"epoch": 22.7,
"learning_rate": 0.02176020749493099,
"loss": 2.9507,
"step": 2220000
},
{
"epoch": 22.81,
"learning_rate": 0.021631552403791724,
"loss": 2.9518,
"step": 2230000
},
{
"epoch": 22.91,
"learning_rate": 0.021502897312652456,
"loss": 2.9523,
"step": 2240000
},
{
"epoch": 23.01,
"learning_rate": 0.02137424222151319,
"loss": 2.951,
"step": 2250000
},
{
"epoch": 23.11,
"learning_rate": 0.021245587130373923,
"loss": 2.9394,
"step": 2260000
},
{
"epoch": 23.21,
"learning_rate": 0.02111693203923466,
"loss": 2.9426,
"step": 2270000
},
{
"epoch": 23.32,
"learning_rate": 0.020988276948095394,
"loss": 2.9425,
"step": 2280000
},
{
"epoch": 23.42,
"learning_rate": 0.020859621856956126,
"loss": 2.9449,
"step": 2290000
},
{
"epoch": 23.52,
"learning_rate": 0.020730966765816858,
"loss": 2.939,
"step": 2300000
},
{
"epoch": 23.62,
"learning_rate": 0.02060231167467759,
"loss": 2.939,
"step": 2310000
},
{
"epoch": 23.73,
"learning_rate": 0.020473656583538325,
"loss": 2.9414,
"step": 2320000
},
{
"epoch": 23.83,
"learning_rate": 0.020345001492399057,
"loss": 2.9433,
"step": 2330000
},
{
"epoch": 23.93,
"learning_rate": 0.020216346401259792,
"loss": 2.9436,
"step": 2340000
},
{
"epoch": 24.03,
"learning_rate": 0.020087691310120524,
"loss": 2.9421,
"step": 2350000
},
{
"epoch": 24.13,
"learning_rate": 0.01995903621898126,
"loss": 2.9385,
"step": 2360000
},
{
"epoch": 24.24,
"learning_rate": 0.019830381127841995,
"loss": 2.9413,
"step": 2370000
},
{
"epoch": 24.34,
"learning_rate": 0.019701726036702727,
"loss": 2.9426,
"step": 2380000
},
{
"epoch": 24.44,
"learning_rate": 0.01957307094556346,
"loss": 2.9423,
"step": 2390000
},
{
"epoch": 24.54,
"learning_rate": 0.01944441585442419,
"loss": 2.9442,
"step": 2400000
},
{
"epoch": 24.65,
"learning_rate": 0.019315760763284926,
"loss": 2.9457,
"step": 2410000
},
{
"epoch": 24.75,
"learning_rate": 0.019187105672145658,
"loss": 2.9455,
"step": 2420000
},
{
"epoch": 24.85,
"learning_rate": 0.019058450581006393,
"loss": 2.947,
"step": 2430000
},
{
"epoch": 24.95,
"learning_rate": 0.018929795489867125,
"loss": 2.9464,
"step": 2440000
},
{
"epoch": 25.06,
"learning_rate": 0.01880114039872786,
"loss": 2.9411,
"step": 2450000
},
{
"epoch": 25.16,
"learning_rate": 0.018672485307588593,
"loss": 2.9362,
"step": 2460000
},
{
"epoch": 25.26,
"learning_rate": 0.018543830216449324,
"loss": 2.9373,
"step": 2470000
},
{
"epoch": 25.36,
"learning_rate": 0.01841517512531006,
"loss": 2.9395,
"step": 2480000
},
{
"epoch": 25.46,
"learning_rate": 0.018286520034170792,
"loss": 2.9404,
"step": 2490000
},
{
"epoch": 25.57,
"learning_rate": 0.018157864943031527,
"loss": 2.9412,
"step": 2500000
},
{
"epoch": 25.67,
"learning_rate": 0.01802920985189226,
"loss": 2.9425,
"step": 2510000
},
{
"epoch": 25.77,
"learning_rate": 0.017900554760752994,
"loss": 2.943,
"step": 2520000
},
{
"epoch": 25.87,
"learning_rate": 0.017771899669613726,
"loss": 2.9446,
"step": 2530000
},
{
"epoch": 25.98,
"learning_rate": 0.01764324457847446,
"loss": 2.945,
"step": 2540000
},
{
"epoch": 26.08,
"learning_rate": 0.017514589487335194,
"loss": 2.9327,
"step": 2550000
},
{
"epoch": 26.18,
"learning_rate": 0.017385934396195925,
"loss": 2.9341,
"step": 2560000
},
{
"epoch": 26.28,
"learning_rate": 0.01725727930505666,
"loss": 2.9366,
"step": 2570000
},
{
"epoch": 26.38,
"learning_rate": 0.017128624213917393,
"loss": 2.9376,
"step": 2580000
},
{
"epoch": 26.49,
"learning_rate": 0.016999969122778128,
"loss": 2.9373,
"step": 2590000
},
{
"epoch": 26.59,
"learning_rate": 0.01687131403163886,
"loss": 2.9388,
"step": 2600000
},
{
"epoch": 26.69,
"learning_rate": 0.016742658940499595,
"loss": 2.9404,
"step": 2610000
},
{
"epoch": 26.79,
"learning_rate": 0.016614003849360327,
"loss": 2.9411,
"step": 2620000
},
{
"epoch": 26.9,
"learning_rate": 0.016485348758221063,
"loss": 2.943,
"step": 2630000
},
{
"epoch": 27.0,
"learning_rate": 0.016356693667081795,
"loss": 2.9421,
"step": 2640000
},
{
"epoch": 27.1,
"learning_rate": 0.016228038575942526,
"loss": 2.9313,
"step": 2650000
},
{
"epoch": 27.2,
"learning_rate": 0.016099383484803262,
"loss": 2.9337,
"step": 2660000
},
{
"epoch": 27.31,
"learning_rate": 0.015970728393663994,
"loss": 2.9341,
"step": 2670000
},
{
"epoch": 27.41,
"learning_rate": 0.01584207330252473,
"loss": 2.9353,
"step": 2680000
},
{
"epoch": 27.51,
"learning_rate": 0.01571341821138546,
"loss": 2.9359,
"step": 2690000
},
{
"epoch": 27.61,
"learning_rate": 0.015584763120246196,
"loss": 2.9363,
"step": 2700000
},
{
"epoch": 27.71,
"learning_rate": 0.015456108029106928,
"loss": 2.9387,
"step": 2710000
},
{
"epoch": 27.82,
"learning_rate": 0.015327452937967662,
"loss": 2.9388,
"step": 2720000
},
{
"epoch": 27.92,
"learning_rate": 0.015198797846828394,
"loss": 2.9399,
"step": 2730000
},
{
"epoch": 28.02,
"learning_rate": 0.01507014275568913,
"loss": 2.9384,
"step": 2740000
},
{
"epoch": 28.12,
"learning_rate": 0.014941487664549863,
"loss": 2.9305,
"step": 2750000
},
{
"epoch": 28.23,
"learning_rate": 0.014812832573410595,
"loss": 2.9325,
"step": 2760000
},
{
"epoch": 28.33,
"learning_rate": 0.01468417748227133,
"loss": 2.9332,
"step": 2770000
},
{
"epoch": 28.43,
"learning_rate": 0.014555522391132062,
"loss": 2.9339,
"step": 2780000
}
],
"max_steps": 3911360,
"num_train_epochs": 40,
"total_flos": 1.8137583487197538e+20,
"trial_name": null,
"trial_params": null
}