llama3_1_truth_model_check / trainer_state.json
Ogamon's picture
Initial commit
860958e verified
raw
history blame contribute delete
No virus
165 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 4.98793242156074,
"eval_steps": 500,
"global_step": 775,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.006436041834271922,
"grad_norm": 857.0773315429688,
"learning_rate": 5.000000000000001e-07,
"loss": 12.2736,
"num_input_tokens_seen": 6576,
"step": 1
},
{
"epoch": 0.012872083668543845,
"grad_norm": 899.0701293945312,
"learning_rate": 1.0000000000000002e-06,
"loss": 12.5518,
"num_input_tokens_seen": 13312,
"step": 2
},
{
"epoch": 0.019308125502815767,
"grad_norm": 833.7578125,
"learning_rate": 1.5e-06,
"loss": 11.8551,
"num_input_tokens_seen": 19952,
"step": 3
},
{
"epoch": 0.02574416733708769,
"grad_norm": 712.9902954101562,
"learning_rate": 2.0000000000000003e-06,
"loss": 10.9045,
"num_input_tokens_seen": 26640,
"step": 4
},
{
"epoch": 0.032180209171359615,
"grad_norm": 620.4878540039062,
"learning_rate": 2.5e-06,
"loss": 8.9845,
"num_input_tokens_seen": 33360,
"step": 5
},
{
"epoch": 0.038616251005631534,
"grad_norm": 376.7406921386719,
"learning_rate": 3e-06,
"loss": 6.35,
"num_input_tokens_seen": 39984,
"step": 6
},
{
"epoch": 0.04505229283990346,
"grad_norm": 356.7503967285156,
"learning_rate": 3.5e-06,
"loss": 5.4864,
"num_input_tokens_seen": 46496,
"step": 7
},
{
"epoch": 0.05148833467417538,
"grad_norm": 383.4678955078125,
"learning_rate": 4.000000000000001e-06,
"loss": 2.8013,
"num_input_tokens_seen": 53008,
"step": 8
},
{
"epoch": 0.057924376508447305,
"grad_norm": 285.64483642578125,
"learning_rate": 4.5e-06,
"loss": 0.9851,
"num_input_tokens_seen": 59856,
"step": 9
},
{
"epoch": 0.06436041834271923,
"grad_norm": 222.22850036621094,
"learning_rate": 5e-06,
"loss": 0.5088,
"num_input_tokens_seen": 66352,
"step": 10
},
{
"epoch": 0.07079646017699115,
"grad_norm": 338.3565368652344,
"learning_rate": 4.99997891923933e-06,
"loss": 2.022,
"num_input_tokens_seen": 73184,
"step": 11
},
{
"epoch": 0.07723250201126307,
"grad_norm": 97.87013244628906,
"learning_rate": 4.999915677312839e-06,
"loss": 0.5842,
"num_input_tokens_seen": 80064,
"step": 12
},
{
"epoch": 0.083668543845535,
"grad_norm": 137.05706787109375,
"learning_rate": 4.999810275287077e-06,
"loss": 0.8538,
"num_input_tokens_seen": 86688,
"step": 13
},
{
"epoch": 0.09010458567980692,
"grad_norm": 88.00048828125,
"learning_rate": 4.9996627149396075e-06,
"loss": 0.5843,
"num_input_tokens_seen": 93168,
"step": 14
},
{
"epoch": 0.09654062751407884,
"grad_norm": 61.654090881347656,
"learning_rate": 4.999472998758979e-06,
"loss": 0.4456,
"num_input_tokens_seen": 99696,
"step": 15
},
{
"epoch": 0.10297666934835076,
"grad_norm": 37.3619499206543,
"learning_rate": 4.99924112994468e-06,
"loss": 0.3542,
"num_input_tokens_seen": 106016,
"step": 16
},
{
"epoch": 0.10941271118262269,
"grad_norm": 41.219093322753906,
"learning_rate": 4.998967112407087e-06,
"loss": 0.3416,
"num_input_tokens_seen": 112672,
"step": 17
},
{
"epoch": 0.11584875301689461,
"grad_norm": 22.267297744750977,
"learning_rate": 4.9986509507673986e-06,
"loss": 0.2803,
"num_input_tokens_seen": 119312,
"step": 18
},
{
"epoch": 0.12228479485116653,
"grad_norm": 22.40268898010254,
"learning_rate": 4.998292650357558e-06,
"loss": 0.2603,
"num_input_tokens_seen": 126016,
"step": 19
},
{
"epoch": 0.12872083668543846,
"grad_norm": 23.54829978942871,
"learning_rate": 4.99789221722016e-06,
"loss": 0.2456,
"num_input_tokens_seen": 132704,
"step": 20
},
{
"epoch": 0.13515687851971037,
"grad_norm": 24.01500701904297,
"learning_rate": 4.997449658108354e-06,
"loss": 0.2269,
"num_input_tokens_seen": 139472,
"step": 21
},
{
"epoch": 0.1415929203539823,
"grad_norm": 26.877809524536133,
"learning_rate": 4.996964980485725e-06,
"loss": 0.2896,
"num_input_tokens_seen": 145968,
"step": 22
},
{
"epoch": 0.14802896218825423,
"grad_norm": 11.89682388305664,
"learning_rate": 4.996438192526173e-06,
"loss": 0.1415,
"num_input_tokens_seen": 152144,
"step": 23
},
{
"epoch": 0.15446500402252614,
"grad_norm": 12.671065330505371,
"learning_rate": 4.995869303113768e-06,
"loss": 0.2228,
"num_input_tokens_seen": 158432,
"step": 24
},
{
"epoch": 0.16090104585679807,
"grad_norm": 12.23610782623291,
"learning_rate": 4.995258321842611e-06,
"loss": 0.1537,
"num_input_tokens_seen": 164672,
"step": 25
},
{
"epoch": 0.16733708769107,
"grad_norm": 21.917552947998047,
"learning_rate": 4.994605259016658e-06,
"loss": 0.2146,
"num_input_tokens_seen": 170896,
"step": 26
},
{
"epoch": 0.1737731295253419,
"grad_norm": 6.669096946716309,
"learning_rate": 4.993910125649561e-06,
"loss": 0.1688,
"num_input_tokens_seen": 177904,
"step": 27
},
{
"epoch": 0.18020917135961384,
"grad_norm": 10.865259170532227,
"learning_rate": 4.99317293346447e-06,
"loss": 0.1592,
"num_input_tokens_seen": 184640,
"step": 28
},
{
"epoch": 0.18664521319388577,
"grad_norm": 22.876623153686523,
"learning_rate": 4.992393694893844e-06,
"loss": 0.2333,
"num_input_tokens_seen": 191008,
"step": 29
},
{
"epoch": 0.19308125502815768,
"grad_norm": 12.427145004272461,
"learning_rate": 4.991572423079236e-06,
"loss": 0.1812,
"num_input_tokens_seen": 197568,
"step": 30
},
{
"epoch": 0.1995172968624296,
"grad_norm": 15.451845169067383,
"learning_rate": 4.990709131871074e-06,
"loss": 0.1925,
"num_input_tokens_seen": 204272,
"step": 31
},
{
"epoch": 0.20595333869670152,
"grad_norm": 18.907636642456055,
"learning_rate": 4.989803835828426e-06,
"loss": 0.1864,
"num_input_tokens_seen": 210944,
"step": 32
},
{
"epoch": 0.21238938053097345,
"grad_norm": 15.60383129119873,
"learning_rate": 4.988856550218755e-06,
"loss": 0.1848,
"num_input_tokens_seen": 217584,
"step": 33
},
{
"epoch": 0.21882542236524538,
"grad_norm": 7.129302024841309,
"learning_rate": 4.987867291017662e-06,
"loss": 0.1466,
"num_input_tokens_seen": 224064,
"step": 34
},
{
"epoch": 0.2252614641995173,
"grad_norm": 15.756115913391113,
"learning_rate": 4.986836074908616e-06,
"loss": 0.1481,
"num_input_tokens_seen": 230880,
"step": 35
},
{
"epoch": 0.23169750603378922,
"grad_norm": 8.845354080200195,
"learning_rate": 4.985762919282674e-06,
"loss": 0.149,
"num_input_tokens_seen": 237312,
"step": 36
},
{
"epoch": 0.23813354786806115,
"grad_norm": 15.7093505859375,
"learning_rate": 4.984647842238185e-06,
"loss": 0.1506,
"num_input_tokens_seen": 243648,
"step": 37
},
{
"epoch": 0.24456958970233306,
"grad_norm": 11.331380844116211,
"learning_rate": 4.983490862580486e-06,
"loss": 0.1709,
"num_input_tokens_seen": 250096,
"step": 38
},
{
"epoch": 0.251005631536605,
"grad_norm": 6.254825115203857,
"learning_rate": 4.982291999821587e-06,
"loss": 0.0898,
"num_input_tokens_seen": 256432,
"step": 39
},
{
"epoch": 0.2574416733708769,
"grad_norm": 7.792216777801514,
"learning_rate": 4.98105127417984e-06,
"loss": 0.1856,
"num_input_tokens_seen": 263088,
"step": 40
},
{
"epoch": 0.26387771520514886,
"grad_norm": 4.186593055725098,
"learning_rate": 4.979768706579595e-06,
"loss": 0.0947,
"num_input_tokens_seen": 269904,
"step": 41
},
{
"epoch": 0.27031375703942073,
"grad_norm": 7.599153518676758,
"learning_rate": 4.978444318650855e-06,
"loss": 0.1367,
"num_input_tokens_seen": 276592,
"step": 42
},
{
"epoch": 0.27674979887369267,
"grad_norm": 10.765763282775879,
"learning_rate": 4.977078132728901e-06,
"loss": 0.1645,
"num_input_tokens_seen": 283424,
"step": 43
},
{
"epoch": 0.2831858407079646,
"grad_norm": 6.705766677856445,
"learning_rate": 4.975670171853926e-06,
"loss": 0.1179,
"num_input_tokens_seen": 290176,
"step": 44
},
{
"epoch": 0.28962188254223653,
"grad_norm": 6.055794715881348,
"learning_rate": 4.9742204597706386e-06,
"loss": 0.133,
"num_input_tokens_seen": 296752,
"step": 45
},
{
"epoch": 0.29605792437650846,
"grad_norm": 7.4584760665893555,
"learning_rate": 4.972729020927866e-06,
"loss": 0.083,
"num_input_tokens_seen": 303392,
"step": 46
},
{
"epoch": 0.3024939662107804,
"grad_norm": 10.979104995727539,
"learning_rate": 4.9711958804781385e-06,
"loss": 0.1748,
"num_input_tokens_seen": 310304,
"step": 47
},
{
"epoch": 0.3089300080450523,
"grad_norm": 13.912871360778809,
"learning_rate": 4.969621064277271e-06,
"loss": 0.1854,
"num_input_tokens_seen": 317440,
"step": 48
},
{
"epoch": 0.3153660498793242,
"grad_norm": 6.554210186004639,
"learning_rate": 4.968004598883923e-06,
"loss": 0.1232,
"num_input_tokens_seen": 324304,
"step": 49
},
{
"epoch": 0.32180209171359614,
"grad_norm": 6.358190536499023,
"learning_rate": 4.966346511559149e-06,
"loss": 0.1172,
"num_input_tokens_seen": 330832,
"step": 50
},
{
"epoch": 0.32823813354786807,
"grad_norm": 6.195626258850098,
"learning_rate": 4.964646830265944e-06,
"loss": 0.1404,
"num_input_tokens_seen": 337952,
"step": 51
},
{
"epoch": 0.33467417538214,
"grad_norm": 12.585171699523926,
"learning_rate": 4.962905583668766e-06,
"loss": 0.137,
"num_input_tokens_seen": 344384,
"step": 52
},
{
"epoch": 0.3411102172164119,
"grad_norm": 3.7672178745269775,
"learning_rate": 4.961122801133059e-06,
"loss": 0.1191,
"num_input_tokens_seen": 351184,
"step": 53
},
{
"epoch": 0.3475462590506838,
"grad_norm": 17.48076629638672,
"learning_rate": 4.9592985127247525e-06,
"loss": 0.1624,
"num_input_tokens_seen": 357696,
"step": 54
},
{
"epoch": 0.35398230088495575,
"grad_norm": 7.758498668670654,
"learning_rate": 4.957432749209755e-06,
"loss": 0.1256,
"num_input_tokens_seen": 364368,
"step": 55
},
{
"epoch": 0.3604183427192277,
"grad_norm": 10.048332214355469,
"learning_rate": 4.955525542053438e-06,
"loss": 0.1274,
"num_input_tokens_seen": 370896,
"step": 56
},
{
"epoch": 0.3668543845534996,
"grad_norm": 17.495296478271484,
"learning_rate": 4.953576923420105e-06,
"loss": 0.174,
"num_input_tokens_seen": 377168,
"step": 57
},
{
"epoch": 0.37329042638777155,
"grad_norm": 7.546329021453857,
"learning_rate": 4.9515869261724444e-06,
"loss": 0.0805,
"num_input_tokens_seen": 383728,
"step": 58
},
{
"epoch": 0.3797264682220434,
"grad_norm": 5.6687188148498535,
"learning_rate": 4.949555583870983e-06,
"loss": 0.1181,
"num_input_tokens_seen": 390448,
"step": 59
},
{
"epoch": 0.38616251005631536,
"grad_norm": 9.777739524841309,
"learning_rate": 4.9474829307735115e-06,
"loss": 0.1613,
"num_input_tokens_seen": 396960,
"step": 60
},
{
"epoch": 0.3925985518905873,
"grad_norm": 10.064454078674316,
"learning_rate": 4.9453690018345144e-06,
"loss": 0.0885,
"num_input_tokens_seen": 403680,
"step": 61
},
{
"epoch": 0.3990345937248592,
"grad_norm": 6.493910312652588,
"learning_rate": 4.943213832704575e-06,
"loss": 0.1526,
"num_input_tokens_seen": 410096,
"step": 62
},
{
"epoch": 0.40547063555913115,
"grad_norm": 7.949091911315918,
"learning_rate": 4.941017459729778e-06,
"loss": 0.1114,
"num_input_tokens_seen": 416672,
"step": 63
},
{
"epoch": 0.41190667739340303,
"grad_norm": 8.829463958740234,
"learning_rate": 4.938779919951092e-06,
"loss": 0.1139,
"num_input_tokens_seen": 423136,
"step": 64
},
{
"epoch": 0.41834271922767496,
"grad_norm": 3.495246410369873,
"learning_rate": 4.936501251103751e-06,
"loss": 0.0878,
"num_input_tokens_seen": 429888,
"step": 65
},
{
"epoch": 0.4247787610619469,
"grad_norm": 8.937992095947266,
"learning_rate": 4.934181491616613e-06,
"loss": 0.1047,
"num_input_tokens_seen": 436720,
"step": 66
},
{
"epoch": 0.43121480289621883,
"grad_norm": 12.225470542907715,
"learning_rate": 4.9318206806115125e-06,
"loss": 0.1323,
"num_input_tokens_seen": 443648,
"step": 67
},
{
"epoch": 0.43765084473049076,
"grad_norm": 8.192527770996094,
"learning_rate": 4.929418857902603e-06,
"loss": 0.095,
"num_input_tokens_seen": 450464,
"step": 68
},
{
"epoch": 0.4440868865647627,
"grad_norm": 9.30573844909668,
"learning_rate": 4.926976063995687e-06,
"loss": 0.2024,
"num_input_tokens_seen": 457296,
"step": 69
},
{
"epoch": 0.4505229283990346,
"grad_norm": 16.341676712036133,
"learning_rate": 4.9244923400875245e-06,
"loss": 0.1614,
"num_input_tokens_seen": 463984,
"step": 70
},
{
"epoch": 0.4569589702333065,
"grad_norm": 19.184734344482422,
"learning_rate": 4.921967728065147e-06,
"loss": 0.2073,
"num_input_tokens_seen": 470432,
"step": 71
},
{
"epoch": 0.46339501206757844,
"grad_norm": 9.802066802978516,
"learning_rate": 4.91940227050515e-06,
"loss": 0.108,
"num_input_tokens_seen": 476736,
"step": 72
},
{
"epoch": 0.46983105390185037,
"grad_norm": 8.744816780090332,
"learning_rate": 4.916796010672969e-06,
"loss": 0.1455,
"num_input_tokens_seen": 482896,
"step": 73
},
{
"epoch": 0.4762670957361223,
"grad_norm": 11.598526954650879,
"learning_rate": 4.914148992522157e-06,
"loss": 0.131,
"num_input_tokens_seen": 489504,
"step": 74
},
{
"epoch": 0.4827031375703942,
"grad_norm": 11.42316722869873,
"learning_rate": 4.911461260693639e-06,
"loss": 0.1416,
"num_input_tokens_seen": 496160,
"step": 75
},
{
"epoch": 0.4891391794046661,
"grad_norm": 3.1568145751953125,
"learning_rate": 4.908732860514958e-06,
"loss": 0.1045,
"num_input_tokens_seen": 502528,
"step": 76
},
{
"epoch": 0.49557522123893805,
"grad_norm": 9.457361221313477,
"learning_rate": 4.905963837999518e-06,
"loss": 0.1466,
"num_input_tokens_seen": 509104,
"step": 77
},
{
"epoch": 0.50201126307321,
"grad_norm": 9.220935821533203,
"learning_rate": 4.903154239845798e-06,
"loss": 0.1502,
"num_input_tokens_seen": 515760,
"step": 78
},
{
"epoch": 0.5084473049074819,
"grad_norm": 5.706662654876709,
"learning_rate": 4.900304113436571e-06,
"loss": 0.1235,
"num_input_tokens_seen": 522336,
"step": 79
},
{
"epoch": 0.5148833467417538,
"grad_norm": 8.74252700805664,
"learning_rate": 4.897413506838103e-06,
"loss": 0.1022,
"num_input_tokens_seen": 528960,
"step": 80
},
{
"epoch": 0.5213193885760258,
"grad_norm": 4.498232841491699,
"learning_rate": 4.894482468799344e-06,
"loss": 0.0922,
"num_input_tokens_seen": 535920,
"step": 81
},
{
"epoch": 0.5277554304102977,
"grad_norm": 3.6902291774749756,
"learning_rate": 4.891511048751102e-06,
"loss": 0.0699,
"num_input_tokens_seen": 542496,
"step": 82
},
{
"epoch": 0.5341914722445696,
"grad_norm": 5.754522323608398,
"learning_rate": 4.888499296805214e-06,
"loss": 0.1057,
"num_input_tokens_seen": 548752,
"step": 83
},
{
"epoch": 0.5406275140788415,
"grad_norm": 4.513391017913818,
"learning_rate": 4.8854472637536966e-06,
"loss": 0.0793,
"num_input_tokens_seen": 555696,
"step": 84
},
{
"epoch": 0.5470635559131134,
"grad_norm": 4.931502342224121,
"learning_rate": 4.882355001067892e-06,
"loss": 0.14,
"num_input_tokens_seen": 562192,
"step": 85
},
{
"epoch": 0.5534995977473853,
"grad_norm": 6.896547794342041,
"learning_rate": 4.8792225608976e-06,
"loss": 0.1538,
"num_input_tokens_seen": 568672,
"step": 86
},
{
"epoch": 0.5599356395816573,
"grad_norm": 3.4364850521087646,
"learning_rate": 4.8760499960702005e-06,
"loss": 0.1135,
"num_input_tokens_seen": 575440,
"step": 87
},
{
"epoch": 0.5663716814159292,
"grad_norm": 6.179934501647949,
"learning_rate": 4.8728373600897535e-06,
"loss": 0.1253,
"num_input_tokens_seen": 581808,
"step": 88
},
{
"epoch": 0.5728077232502011,
"grad_norm": 14.744488716125488,
"learning_rate": 4.869584707136109e-06,
"loss": 0.1408,
"num_input_tokens_seen": 588576,
"step": 89
},
{
"epoch": 0.5792437650844731,
"grad_norm": 8.414978981018066,
"learning_rate": 4.8662920920639866e-06,
"loss": 0.0916,
"num_input_tokens_seen": 595328,
"step": 90
},
{
"epoch": 0.585679806918745,
"grad_norm": 3.016206979751587,
"learning_rate": 4.86295957040205e-06,
"loss": 0.1016,
"num_input_tokens_seen": 601808,
"step": 91
},
{
"epoch": 0.5921158487530169,
"grad_norm": 3.084475040435791,
"learning_rate": 4.8595871983519705e-06,
"loss": 0.0936,
"num_input_tokens_seen": 608400,
"step": 92
},
{
"epoch": 0.5985518905872889,
"grad_norm": 5.78838586807251,
"learning_rate": 4.856175032787485e-06,
"loss": 0.1557,
"num_input_tokens_seen": 615296,
"step": 93
},
{
"epoch": 0.6049879324215608,
"grad_norm": 4.525265216827393,
"learning_rate": 4.852723131253429e-06,
"loss": 0.0879,
"num_input_tokens_seen": 621888,
"step": 94
},
{
"epoch": 0.6114239742558326,
"grad_norm": 8.129231452941895,
"learning_rate": 4.849231551964771e-06,
"loss": 0.1399,
"num_input_tokens_seen": 628768,
"step": 95
},
{
"epoch": 0.6178600160901045,
"grad_norm": 7.902085781097412,
"learning_rate": 4.845700353805629e-06,
"loss": 0.1724,
"num_input_tokens_seen": 635056,
"step": 96
},
{
"epoch": 0.6242960579243765,
"grad_norm": 7.928036689758301,
"learning_rate": 4.842129596328277e-06,
"loss": 0.1018,
"num_input_tokens_seen": 641872,
"step": 97
},
{
"epoch": 0.6307320997586484,
"grad_norm": 5.5206756591796875,
"learning_rate": 4.838519339752143e-06,
"loss": 0.0398,
"num_input_tokens_seen": 648752,
"step": 98
},
{
"epoch": 0.6371681415929203,
"grad_norm": 5.910008907318115,
"learning_rate": 4.834869644962789e-06,
"loss": 0.1094,
"num_input_tokens_seen": 655424,
"step": 99
},
{
"epoch": 0.6436041834271923,
"grad_norm": 17.12012481689453,
"learning_rate": 4.83118057351089e-06,
"loss": 0.1915,
"num_input_tokens_seen": 662224,
"step": 100
},
{
"epoch": 0.6500402252614642,
"grad_norm": 13.876479148864746,
"learning_rate": 4.827452187611192e-06,
"loss": 0.1518,
"num_input_tokens_seen": 668576,
"step": 101
},
{
"epoch": 0.6564762670957361,
"grad_norm": 9.082406044006348,
"learning_rate": 4.823684550141464e-06,
"loss": 0.141,
"num_input_tokens_seen": 675232,
"step": 102
},
{
"epoch": 0.6629123089300081,
"grad_norm": 7.364652633666992,
"learning_rate": 4.819877724641437e-06,
"loss": 0.1564,
"num_input_tokens_seen": 681856,
"step": 103
},
{
"epoch": 0.66934835076428,
"grad_norm": 16.54301643371582,
"learning_rate": 4.8160317753117326e-06,
"loss": 0.1267,
"num_input_tokens_seen": 688416,
"step": 104
},
{
"epoch": 0.6757843925985519,
"grad_norm": 25.702648162841797,
"learning_rate": 4.81214676701278e-06,
"loss": 0.228,
"num_input_tokens_seen": 695248,
"step": 105
},
{
"epoch": 0.6822204344328238,
"grad_norm": 17.066158294677734,
"learning_rate": 4.808222765263724e-06,
"loss": 0.1532,
"num_input_tokens_seen": 701952,
"step": 106
},
{
"epoch": 0.6886564762670957,
"grad_norm": 11.833669662475586,
"learning_rate": 4.8042598362413175e-06,
"loss": 0.1482,
"num_input_tokens_seen": 708368,
"step": 107
},
{
"epoch": 0.6950925181013676,
"grad_norm": 2.9714369773864746,
"learning_rate": 4.800258046778809e-06,
"loss": 0.1074,
"num_input_tokens_seen": 714768,
"step": 108
},
{
"epoch": 0.7015285599356396,
"grad_norm": 9.384042739868164,
"learning_rate": 4.796217464364808e-06,
"loss": 0.1163,
"num_input_tokens_seen": 721600,
"step": 109
},
{
"epoch": 0.7079646017699115,
"grad_norm": 12.922999382019043,
"learning_rate": 4.792138157142158e-06,
"loss": 0.1871,
"num_input_tokens_seen": 728448,
"step": 110
},
{
"epoch": 0.7144006436041834,
"grad_norm": 5.947402000427246,
"learning_rate": 4.788020193906776e-06,
"loss": 0.092,
"num_input_tokens_seen": 734720,
"step": 111
},
{
"epoch": 0.7208366854384554,
"grad_norm": 6.692570686340332,
"learning_rate": 4.783863644106502e-06,
"loss": 0.0959,
"num_input_tokens_seen": 741216,
"step": 112
},
{
"epoch": 0.7272727272727273,
"grad_norm": 4.1358771324157715,
"learning_rate": 4.779668577839921e-06,
"loss": 0.0853,
"num_input_tokens_seen": 747664,
"step": 113
},
{
"epoch": 0.7337087691069992,
"grad_norm": 4.695752143859863,
"learning_rate": 4.775435065855183e-06,
"loss": 0.1581,
"num_input_tokens_seen": 754480,
"step": 114
},
{
"epoch": 0.7401448109412712,
"grad_norm": 3.8357784748077393,
"learning_rate": 4.771163179548809e-06,
"loss": 0.0907,
"num_input_tokens_seen": 761344,
"step": 115
},
{
"epoch": 0.7465808527755431,
"grad_norm": 4.455271244049072,
"learning_rate": 4.766852990964492e-06,
"loss": 0.0826,
"num_input_tokens_seen": 768160,
"step": 116
},
{
"epoch": 0.7530168946098149,
"grad_norm": 3.682065963745117,
"learning_rate": 4.762504572791873e-06,
"loss": 0.0979,
"num_input_tokens_seen": 774496,
"step": 117
},
{
"epoch": 0.7594529364440868,
"grad_norm": 6.100201606750488,
"learning_rate": 4.7581179983653224e-06,
"loss": 0.1617,
"num_input_tokens_seen": 781232,
"step": 118
},
{
"epoch": 0.7658889782783588,
"grad_norm": 7.6822991371154785,
"learning_rate": 4.753693341662702e-06,
"loss": 0.1306,
"num_input_tokens_seen": 788064,
"step": 119
},
{
"epoch": 0.7723250201126307,
"grad_norm": 4.73075008392334,
"learning_rate": 4.749230677304114e-06,
"loss": 0.0955,
"num_input_tokens_seen": 794656,
"step": 120
},
{
"epoch": 0.7787610619469026,
"grad_norm": 4.435886859893799,
"learning_rate": 4.7447300805506455e-06,
"loss": 0.0683,
"num_input_tokens_seen": 801184,
"step": 121
},
{
"epoch": 0.7851971037811746,
"grad_norm": 3.517606735229492,
"learning_rate": 4.7401916273031e-06,
"loss": 0.1117,
"num_input_tokens_seen": 808000,
"step": 122
},
{
"epoch": 0.7916331456154465,
"grad_norm": 3.445953130722046,
"learning_rate": 4.7356153941007145e-06,
"loss": 0.1115,
"num_input_tokens_seen": 814608,
"step": 123
},
{
"epoch": 0.7980691874497184,
"grad_norm": 2.4660255908966064,
"learning_rate": 4.73100145811987e-06,
"loss": 0.0945,
"num_input_tokens_seen": 821072,
"step": 124
},
{
"epoch": 0.8045052292839904,
"grad_norm": 6.673710346221924,
"learning_rate": 4.726349897172791e-06,
"loss": 0.1125,
"num_input_tokens_seen": 827840,
"step": 125
},
{
"epoch": 0.8109412711182623,
"grad_norm": 7.910948753356934,
"learning_rate": 4.721660789706232e-06,
"loss": 0.15,
"num_input_tokens_seen": 834880,
"step": 126
},
{
"epoch": 0.8173773129525342,
"grad_norm": 4.5563154220581055,
"learning_rate": 4.716934214800155e-06,
"loss": 0.1015,
"num_input_tokens_seen": 841360,
"step": 127
},
{
"epoch": 0.8238133547868061,
"grad_norm": 5.7200422286987305,
"learning_rate": 4.712170252166395e-06,
"loss": 0.1271,
"num_input_tokens_seen": 847888,
"step": 128
},
{
"epoch": 0.830249396621078,
"grad_norm": 4.4525465965271,
"learning_rate": 4.707368982147318e-06,
"loss": 0.0762,
"num_input_tokens_seen": 854896,
"step": 129
},
{
"epoch": 0.8366854384553499,
"grad_norm": 4.427840232849121,
"learning_rate": 4.702530485714462e-06,
"loss": 0.1196,
"num_input_tokens_seen": 861600,
"step": 130
},
{
"epoch": 0.8431214802896219,
"grad_norm": 3.674197196960449,
"learning_rate": 4.697654844467175e-06,
"loss": 0.0866,
"num_input_tokens_seen": 868272,
"step": 131
},
{
"epoch": 0.8495575221238938,
"grad_norm": 7.5055413246154785,
"learning_rate": 4.69274214063124e-06,
"loss": 0.0718,
"num_input_tokens_seen": 875232,
"step": 132
},
{
"epoch": 0.8559935639581657,
"grad_norm": 7.523169040679932,
"learning_rate": 4.687792457057482e-06,
"loss": 0.0808,
"num_input_tokens_seen": 882112,
"step": 133
},
{
"epoch": 0.8624296057924377,
"grad_norm": 10.57685375213623,
"learning_rate": 4.682805877220378e-06,
"loss": 0.1069,
"num_input_tokens_seen": 888848,
"step": 134
},
{
"epoch": 0.8688656476267096,
"grad_norm": 6.235794544219971,
"learning_rate": 4.677782485216644e-06,
"loss": 0.0804,
"num_input_tokens_seen": 895136,
"step": 135
},
{
"epoch": 0.8753016894609815,
"grad_norm": 5.526005268096924,
"learning_rate": 4.672722365763821e-06,
"loss": 0.068,
"num_input_tokens_seen": 901552,
"step": 136
},
{
"epoch": 0.8817377312952535,
"grad_norm": 6.142871856689453,
"learning_rate": 4.667625604198842e-06,
"loss": 0.1193,
"num_input_tokens_seen": 908272,
"step": 137
},
{
"epoch": 0.8881737731295254,
"grad_norm": 17.300273895263672,
"learning_rate": 4.662492286476595e-06,
"loss": 0.1535,
"num_input_tokens_seen": 914928,
"step": 138
},
{
"epoch": 0.8946098149637972,
"grad_norm": 13.767914772033691,
"learning_rate": 4.657322499168475e-06,
"loss": 0.1303,
"num_input_tokens_seen": 921296,
"step": 139
},
{
"epoch": 0.9010458567980691,
"grad_norm": 5.356888294219971,
"learning_rate": 4.65211632946092e-06,
"loss": 0.0879,
"num_input_tokens_seen": 927728,
"step": 140
},
{
"epoch": 0.9074818986323411,
"grad_norm": 10.261467933654785,
"learning_rate": 4.646873865153945e-06,
"loss": 0.0986,
"num_input_tokens_seen": 934240,
"step": 141
},
{
"epoch": 0.913917940466613,
"grad_norm": 14.075957298278809,
"learning_rate": 4.641595194659657e-06,
"loss": 0.1219,
"num_input_tokens_seen": 940832,
"step": 142
},
{
"epoch": 0.9203539823008849,
"grad_norm": 11.964951515197754,
"learning_rate": 4.63628040700077e-06,
"loss": 0.1303,
"num_input_tokens_seen": 947856,
"step": 143
},
{
"epoch": 0.9267900241351569,
"grad_norm": 6.297915935516357,
"learning_rate": 4.630929591809095e-06,
"loss": 0.081,
"num_input_tokens_seen": 954160,
"step": 144
},
{
"epoch": 0.9332260659694288,
"grad_norm": 4.006863594055176,
"learning_rate": 4.625542839324036e-06,
"loss": 0.0979,
"num_input_tokens_seen": 960848,
"step": 145
},
{
"epoch": 0.9396621078037007,
"grad_norm": 9.041242599487305,
"learning_rate": 4.620120240391065e-06,
"loss": 0.1446,
"num_input_tokens_seen": 967440,
"step": 146
},
{
"epoch": 0.9460981496379727,
"grad_norm": 14.858406066894531,
"learning_rate": 4.614661886460191e-06,
"loss": 0.1267,
"num_input_tokens_seen": 973808,
"step": 147
},
{
"epoch": 0.9525341914722446,
"grad_norm": 12.371238708496094,
"learning_rate": 4.609167869584416e-06,
"loss": 0.1095,
"num_input_tokens_seen": 980352,
"step": 148
},
{
"epoch": 0.9589702333065165,
"grad_norm": 6.89439582824707,
"learning_rate": 4.6036382824181836e-06,
"loss": 0.1252,
"num_input_tokens_seen": 987088,
"step": 149
},
{
"epoch": 0.9654062751407884,
"grad_norm": 3.6482529640197754,
"learning_rate": 4.598073218215817e-06,
"loss": 0.0645,
"num_input_tokens_seen": 993648,
"step": 150
},
{
"epoch": 0.9718423169750603,
"grad_norm": 6.078918933868408,
"learning_rate": 4.592472770829945e-06,
"loss": 0.0974,
"num_input_tokens_seen": 1000272,
"step": 151
},
{
"epoch": 0.9782783588093322,
"grad_norm": 10.974119186401367,
"learning_rate": 4.586837034709921e-06,
"loss": 0.0833,
"num_input_tokens_seen": 1006912,
"step": 152
},
{
"epoch": 0.9847144006436042,
"grad_norm": 8.552461624145508,
"learning_rate": 4.581166104900228e-06,
"loss": 0.0787,
"num_input_tokens_seen": 1013328,
"step": 153
},
{
"epoch": 0.9911504424778761,
"grad_norm": 8.927652359008789,
"learning_rate": 4.575460077038877e-06,
"loss": 0.0814,
"num_input_tokens_seen": 1020128,
"step": 154
},
{
"epoch": 0.997586484312148,
"grad_norm": 2.613471269607544,
"learning_rate": 4.569719047355795e-06,
"loss": 0.0278,
"num_input_tokens_seen": 1026848,
"step": 155
},
{
"epoch": 1.00402252614642,
"grad_norm": 4.19236946105957,
"learning_rate": 4.5639431126712e-06,
"loss": 0.093,
"num_input_tokens_seen": 1033728,
"step": 156
},
{
"epoch": 1.010458567980692,
"grad_norm": 6.943019866943359,
"learning_rate": 4.5581323703939685e-06,
"loss": 0.073,
"num_input_tokens_seen": 1040720,
"step": 157
},
{
"epoch": 1.0168946098149638,
"grad_norm": 6.545025825500488,
"learning_rate": 4.552286918519996e-06,
"loss": 0.0625,
"num_input_tokens_seen": 1047168,
"step": 158
},
{
"epoch": 1.0233306516492358,
"grad_norm": 7.890603065490723,
"learning_rate": 4.5464068556305375e-06,
"loss": 0.0461,
"num_input_tokens_seen": 1053760,
"step": 159
},
{
"epoch": 1.0297666934835077,
"grad_norm": 5.44887638092041,
"learning_rate": 4.540492280890555e-06,
"loss": 0.0318,
"num_input_tokens_seen": 1060176,
"step": 160
},
{
"epoch": 1.0362027353177796,
"grad_norm": 1.036007285118103,
"learning_rate": 4.534543294047033e-06,
"loss": 0.0068,
"num_input_tokens_seen": 1066768,
"step": 161
},
{
"epoch": 1.0426387771520516,
"grad_norm": 5.863292694091797,
"learning_rate": 4.528559995427309e-06,
"loss": 0.0462,
"num_input_tokens_seen": 1073376,
"step": 162
},
{
"epoch": 1.0490748189863235,
"grad_norm": 8.744257926940918,
"learning_rate": 4.522542485937369e-06,
"loss": 0.0487,
"num_input_tokens_seen": 1079952,
"step": 163
},
{
"epoch": 1.0555108608205954,
"grad_norm": 6.485115051269531,
"learning_rate": 4.516490867060156e-06,
"loss": 0.0664,
"num_input_tokens_seen": 1086848,
"step": 164
},
{
"epoch": 1.0619469026548674,
"grad_norm": 3.8945565223693848,
"learning_rate": 4.5104052408538545e-06,
"loss": 0.0347,
"num_input_tokens_seen": 1093328,
"step": 165
},
{
"epoch": 1.068382944489139,
"grad_norm": 3.5805532932281494,
"learning_rate": 4.504285709950167e-06,
"loss": 0.0202,
"num_input_tokens_seen": 1099840,
"step": 166
},
{
"epoch": 1.074818986323411,
"grad_norm": 6.033172607421875,
"learning_rate": 4.498132377552587e-06,
"loss": 0.0573,
"num_input_tokens_seen": 1106528,
"step": 167
},
{
"epoch": 1.081255028157683,
"grad_norm": 8.104386329650879,
"learning_rate": 4.491945347434656e-06,
"loss": 0.0848,
"num_input_tokens_seen": 1113424,
"step": 168
},
{
"epoch": 1.0876910699919549,
"grad_norm": 6.130471229553223,
"learning_rate": 4.485724723938215e-06,
"loss": 0.0464,
"num_input_tokens_seen": 1120064,
"step": 169
},
{
"epoch": 1.0941271118262268,
"grad_norm": 8.871036529541016,
"learning_rate": 4.479470611971646e-06,
"loss": 0.1004,
"num_input_tokens_seen": 1126960,
"step": 170
},
{
"epoch": 1.1005631536604987,
"grad_norm": 7.802618026733398,
"learning_rate": 4.473183117008096e-06,
"loss": 0.0842,
"num_input_tokens_seen": 1133664,
"step": 171
},
{
"epoch": 1.1069991954947707,
"grad_norm": 2.848886489868164,
"learning_rate": 4.4668623450837085e-06,
"loss": 0.0452,
"num_input_tokens_seen": 1140048,
"step": 172
},
{
"epoch": 1.1134352373290426,
"grad_norm": 2.9371185302734375,
"learning_rate": 4.460508402795827e-06,
"loss": 0.0225,
"num_input_tokens_seen": 1146448,
"step": 173
},
{
"epoch": 1.1198712791633145,
"grad_norm": 2.1428751945495605,
"learning_rate": 4.4541213973012005e-06,
"loss": 0.0058,
"num_input_tokens_seen": 1152960,
"step": 174
},
{
"epoch": 1.1263073209975865,
"grad_norm": 6.480560302734375,
"learning_rate": 4.447701436314176e-06,
"loss": 0.0565,
"num_input_tokens_seen": 1159632,
"step": 175
},
{
"epoch": 1.1327433628318584,
"grad_norm": 8.678375244140625,
"learning_rate": 4.441248628104884e-06,
"loss": 0.0591,
"num_input_tokens_seen": 1166640,
"step": 176
},
{
"epoch": 1.1391794046661303,
"grad_norm": 8.184906005859375,
"learning_rate": 4.434763081497407e-06,
"loss": 0.0488,
"num_input_tokens_seen": 1173408,
"step": 177
},
{
"epoch": 1.1456154465004023,
"grad_norm": 3.727961540222168,
"learning_rate": 4.428244905867952e-06,
"loss": 0.0318,
"num_input_tokens_seen": 1179776,
"step": 178
},
{
"epoch": 1.1520514883346742,
"grad_norm": 7.119325160980225,
"learning_rate": 4.421694211142998e-06,
"loss": 0.064,
"num_input_tokens_seen": 1186720,
"step": 179
},
{
"epoch": 1.1584875301689461,
"grad_norm": 3.094886541366577,
"learning_rate": 4.415111107797445e-06,
"loss": 0.0465,
"num_input_tokens_seen": 1193408,
"step": 180
},
{
"epoch": 1.164923572003218,
"grad_norm": 5.577038288116455,
"learning_rate": 4.408495706852758e-06,
"loss": 0.0344,
"num_input_tokens_seen": 1200096,
"step": 181
},
{
"epoch": 1.17135961383749,
"grad_norm": 7.607036590576172,
"learning_rate": 4.401848119875081e-06,
"loss": 0.0747,
"num_input_tokens_seen": 1206848,
"step": 182
},
{
"epoch": 1.177795655671762,
"grad_norm": 5.953075885772705,
"learning_rate": 4.395168458973368e-06,
"loss": 0.073,
"num_input_tokens_seen": 1213632,
"step": 183
},
{
"epoch": 1.1842316975060339,
"grad_norm": 7.784894943237305,
"learning_rate": 4.388456836797484e-06,
"loss": 0.0652,
"num_input_tokens_seen": 1220336,
"step": 184
},
{
"epoch": 1.1906677393403058,
"grad_norm": 6.535793781280518,
"learning_rate": 4.381713366536312e-06,
"loss": 0.0881,
"num_input_tokens_seen": 1226736,
"step": 185
},
{
"epoch": 1.1971037811745777,
"grad_norm": 4.9065093994140625,
"learning_rate": 4.374938161915835e-06,
"loss": 0.0676,
"num_input_tokens_seen": 1233536,
"step": 186
},
{
"epoch": 1.2035398230088497,
"grad_norm": 5.7732648849487305,
"learning_rate": 4.368131337197228e-06,
"loss": 0.0481,
"num_input_tokens_seen": 1240032,
"step": 187
},
{
"epoch": 1.2099758648431216,
"grad_norm": 5.656060218811035,
"learning_rate": 4.361293007174926e-06,
"loss": 0.0477,
"num_input_tokens_seen": 1247008,
"step": 188
},
{
"epoch": 1.2164119066773935,
"grad_norm": 3.2787587642669678,
"learning_rate": 4.354423287174686e-06,
"loss": 0.0456,
"num_input_tokens_seen": 1254032,
"step": 189
},
{
"epoch": 1.2228479485116655,
"grad_norm": 9.526626586914062,
"learning_rate": 4.3475222930516484e-06,
"loss": 0.133,
"num_input_tokens_seen": 1261104,
"step": 190
},
{
"epoch": 1.2292839903459372,
"grad_norm": 4.272536277770996,
"learning_rate": 4.340590141188377e-06,
"loss": 0.0672,
"num_input_tokens_seen": 1267680,
"step": 191
},
{
"epoch": 1.235720032180209,
"grad_norm": 4.5576701164245605,
"learning_rate": 4.333626948492898e-06,
"loss": 0.0352,
"num_input_tokens_seen": 1274112,
"step": 192
},
{
"epoch": 1.242156074014481,
"grad_norm": 2.7765443325042725,
"learning_rate": 4.326632832396733e-06,
"loss": 0.0361,
"num_input_tokens_seen": 1280528,
"step": 193
},
{
"epoch": 1.248592115848753,
"grad_norm": 2.681631565093994,
"learning_rate": 4.319607910852911e-06,
"loss": 0.0432,
"num_input_tokens_seen": 1287232,
"step": 194
},
{
"epoch": 1.255028157683025,
"grad_norm": 7.467050075531006,
"learning_rate": 4.3125523023339825e-06,
"loss": 0.0398,
"num_input_tokens_seen": 1293792,
"step": 195
},
{
"epoch": 1.2614641995172968,
"grad_norm": 2.6330530643463135,
"learning_rate": 4.305466125830023e-06,
"loss": 0.0473,
"num_input_tokens_seen": 1300624,
"step": 196
},
{
"epoch": 1.2679002413515688,
"grad_norm": 5.228641510009766,
"learning_rate": 4.2983495008466285e-06,
"loss": 0.0394,
"num_input_tokens_seen": 1307520,
"step": 197
},
{
"epoch": 1.2743362831858407,
"grad_norm": 5.0004191398620605,
"learning_rate": 4.29120254740289e-06,
"loss": 0.0645,
"num_input_tokens_seen": 1313824,
"step": 198
},
{
"epoch": 1.2807723250201126,
"grad_norm": 5.468844890594482,
"learning_rate": 4.284025386029381e-06,
"loss": 0.0663,
"num_input_tokens_seen": 1319952,
"step": 199
},
{
"epoch": 1.2872083668543846,
"grad_norm": 6.145412921905518,
"learning_rate": 4.276818137766118e-06,
"loss": 0.1067,
"num_input_tokens_seen": 1326352,
"step": 200
},
{
"epoch": 1.2936444086886565,
"grad_norm": 5.632473945617676,
"learning_rate": 4.269580924160523e-06,
"loss": 0.0603,
"num_input_tokens_seen": 1332912,
"step": 201
},
{
"epoch": 1.3000804505229284,
"grad_norm": 1.326751947402954,
"learning_rate": 4.262313867265369e-06,
"loss": 0.0055,
"num_input_tokens_seen": 1339872,
"step": 202
},
{
"epoch": 1.3065164923572004,
"grad_norm": 6.162146091461182,
"learning_rate": 4.255017089636725e-06,
"loss": 0.0547,
"num_input_tokens_seen": 1346240,
"step": 203
},
{
"epoch": 1.3129525341914723,
"grad_norm": 2.500483989715576,
"learning_rate": 4.24769071433189e-06,
"loss": 0.0684,
"num_input_tokens_seen": 1353104,
"step": 204
},
{
"epoch": 1.3193885760257442,
"grad_norm": 5.962297439575195,
"learning_rate": 4.240334864907317e-06,
"loss": 0.0484,
"num_input_tokens_seen": 1359664,
"step": 205
},
{
"epoch": 1.3258246178600162,
"grad_norm": 4.183216571807861,
"learning_rate": 4.232949665416526e-06,
"loss": 0.0225,
"num_input_tokens_seen": 1366112,
"step": 206
},
{
"epoch": 1.332260659694288,
"grad_norm": 2.874197244644165,
"learning_rate": 4.225535240408014e-06,
"loss": 0.0278,
"num_input_tokens_seen": 1372912,
"step": 207
},
{
"epoch": 1.33869670152856,
"grad_norm": 5.831293106079102,
"learning_rate": 4.218091714923157e-06,
"loss": 0.034,
"num_input_tokens_seen": 1379200,
"step": 208
},
{
"epoch": 1.3451327433628317,
"grad_norm": 4.147435665130615,
"learning_rate": 4.210619214494099e-06,
"loss": 0.0453,
"num_input_tokens_seen": 1385520,
"step": 209
},
{
"epoch": 1.3515687851971037,
"grad_norm": 6.03895378112793,
"learning_rate": 4.203117865141635e-06,
"loss": 0.0564,
"num_input_tokens_seen": 1391968,
"step": 210
},
{
"epoch": 1.3580048270313756,
"grad_norm": 4.196593284606934,
"learning_rate": 4.195587793373085e-06,
"loss": 0.0318,
"num_input_tokens_seen": 1398576,
"step": 211
},
{
"epoch": 1.3644408688656475,
"grad_norm": 6.364063739776611,
"learning_rate": 4.188029126180161e-06,
"loss": 0.0575,
"num_input_tokens_seen": 1405280,
"step": 212
},
{
"epoch": 1.3708769106999195,
"grad_norm": 5.420915603637695,
"learning_rate": 4.180441991036827e-06,
"loss": 0.0448,
"num_input_tokens_seen": 1411968,
"step": 213
},
{
"epoch": 1.3773129525341914,
"grad_norm": 5.313647747039795,
"learning_rate": 4.172826515897146e-06,
"loss": 0.0493,
"num_input_tokens_seen": 1418576,
"step": 214
},
{
"epoch": 1.3837489943684633,
"grad_norm": 3.2573652267456055,
"learning_rate": 4.165182829193126e-06,
"loss": 0.0478,
"num_input_tokens_seen": 1425360,
"step": 215
},
{
"epoch": 1.3901850362027353,
"grad_norm": 4.227644443511963,
"learning_rate": 4.15751105983255e-06,
"loss": 0.0887,
"num_input_tokens_seen": 1432144,
"step": 216
},
{
"epoch": 1.3966210780370072,
"grad_norm": 6.514432907104492,
"learning_rate": 4.149811337196808e-06,
"loss": 0.0634,
"num_input_tokens_seen": 1439200,
"step": 217
},
{
"epoch": 1.4030571198712791,
"grad_norm": 3.3998050689697266,
"learning_rate": 4.142083791138703e-06,
"loss": 0.0349,
"num_input_tokens_seen": 1445728,
"step": 218
},
{
"epoch": 1.409493161705551,
"grad_norm": 5.725708961486816,
"learning_rate": 4.134328551980279e-06,
"loss": 0.0459,
"num_input_tokens_seen": 1452384,
"step": 219
},
{
"epoch": 1.415929203539823,
"grad_norm": 3.3524420261383057,
"learning_rate": 4.126545750510605e-06,
"loss": 0.0304,
"num_input_tokens_seen": 1459136,
"step": 220
},
{
"epoch": 1.422365245374095,
"grad_norm": 9.169583320617676,
"learning_rate": 4.118735517983584e-06,
"loss": 0.0658,
"num_input_tokens_seen": 1465632,
"step": 221
},
{
"epoch": 1.4288012872083669,
"grad_norm": 5.310299873352051,
"learning_rate": 4.110897986115729e-06,
"loss": 0.087,
"num_input_tokens_seen": 1472592,
"step": 222
},
{
"epoch": 1.4352373290426388,
"grad_norm": 4.850796222686768,
"learning_rate": 4.1030332870839466e-06,
"loss": 0.0952,
"num_input_tokens_seen": 1479168,
"step": 223
},
{
"epoch": 1.4416733708769107,
"grad_norm": 5.20851993560791,
"learning_rate": 4.0951415535233065e-06,
"loss": 0.0358,
"num_input_tokens_seen": 1485664,
"step": 224
},
{
"epoch": 1.4481094127111827,
"grad_norm": 4.104648113250732,
"learning_rate": 4.087222918524807e-06,
"loss": 0.0527,
"num_input_tokens_seen": 1492368,
"step": 225
},
{
"epoch": 1.4545454545454546,
"grad_norm": 2.5263378620147705,
"learning_rate": 4.079277515633127e-06,
"loss": 0.0452,
"num_input_tokens_seen": 1498752,
"step": 226
},
{
"epoch": 1.4609814963797265,
"grad_norm": 2.5317678451538086,
"learning_rate": 4.0713054788443776e-06,
"loss": 0.0313,
"num_input_tokens_seen": 1505296,
"step": 227
},
{
"epoch": 1.4674175382139985,
"grad_norm": 5.61666202545166,
"learning_rate": 4.063306942603835e-06,
"loss": 0.0544,
"num_input_tokens_seen": 1511584,
"step": 228
},
{
"epoch": 1.4738535800482704,
"grad_norm": 3.4415996074676514,
"learning_rate": 4.0552820418036855e-06,
"loss": 0.0428,
"num_input_tokens_seen": 1517776,
"step": 229
},
{
"epoch": 1.4802896218825423,
"grad_norm": 2.44814395904541,
"learning_rate": 4.0472309117807365e-06,
"loss": 0.0183,
"num_input_tokens_seen": 1524416,
"step": 230
},
{
"epoch": 1.4867256637168142,
"grad_norm": 3.827312469482422,
"learning_rate": 4.039153688314146e-06,
"loss": 0.0662,
"num_input_tokens_seen": 1530864,
"step": 231
},
{
"epoch": 1.4931617055510862,
"grad_norm": 2.9351532459259033,
"learning_rate": 4.031050507623125e-06,
"loss": 0.0258,
"num_input_tokens_seen": 1537216,
"step": 232
},
{
"epoch": 1.4995977473853581,
"grad_norm": 6.040038585662842,
"learning_rate": 4.022921506364644e-06,
"loss": 0.0584,
"num_input_tokens_seen": 1543824,
"step": 233
},
{
"epoch": 1.50603378921963,
"grad_norm": 2.7363831996917725,
"learning_rate": 4.014766821631128e-06,
"loss": 0.0916,
"num_input_tokens_seen": 1550432,
"step": 234
},
{
"epoch": 1.512469831053902,
"grad_norm": 4.466485977172852,
"learning_rate": 4.006586590948141e-06,
"loss": 0.0397,
"num_input_tokens_seen": 1556912,
"step": 235
},
{
"epoch": 1.518905872888174,
"grad_norm": 4.36499547958374,
"learning_rate": 3.998380952272073e-06,
"loss": 0.0405,
"num_input_tokens_seen": 1563456,
"step": 236
},
{
"epoch": 1.5253419147224458,
"grad_norm": 3.1068978309631348,
"learning_rate": 3.990150043987806e-06,
"loss": 0.0645,
"num_input_tokens_seen": 1570240,
"step": 237
},
{
"epoch": 1.5317779565567178,
"grad_norm": 4.554339408874512,
"learning_rate": 3.981894004906388e-06,
"loss": 0.0389,
"num_input_tokens_seen": 1576896,
"step": 238
},
{
"epoch": 1.5382139983909895,
"grad_norm": 2.1207427978515625,
"learning_rate": 3.973612974262685e-06,
"loss": 0.0341,
"num_input_tokens_seen": 1583440,
"step": 239
},
{
"epoch": 1.5446500402252614,
"grad_norm": 4.71979284286499,
"learning_rate": 3.965307091713037e-06,
"loss": 0.0625,
"num_input_tokens_seen": 1589968,
"step": 240
},
{
"epoch": 1.5510860820595334,
"grad_norm": 3.9797351360321045,
"learning_rate": 3.956976497332903e-06,
"loss": 0.0651,
"num_input_tokens_seen": 1596416,
"step": 241
},
{
"epoch": 1.5575221238938053,
"grad_norm": 4.844697952270508,
"learning_rate": 3.948621331614495e-06,
"loss": 0.0391,
"num_input_tokens_seen": 1602944,
"step": 242
},
{
"epoch": 1.5639581657280772,
"grad_norm": 4.572307109832764,
"learning_rate": 3.9402417354644115e-06,
"loss": 0.0486,
"num_input_tokens_seen": 1609632,
"step": 243
},
{
"epoch": 1.5703942075623492,
"grad_norm": 7.0537309646606445,
"learning_rate": 3.9318378502012636e-06,
"loss": 0.1192,
"num_input_tokens_seen": 1616096,
"step": 244
},
{
"epoch": 1.576830249396621,
"grad_norm": 4.42478609085083,
"learning_rate": 3.923409817553284e-06,
"loss": 0.0679,
"num_input_tokens_seen": 1622848,
"step": 245
},
{
"epoch": 1.583266291230893,
"grad_norm": 5.157562255859375,
"learning_rate": 3.914957779655946e-06,
"loss": 0.0493,
"num_input_tokens_seen": 1629600,
"step": 246
},
{
"epoch": 1.589702333065165,
"grad_norm": 2.8394153118133545,
"learning_rate": 3.906481879049559e-06,
"loss": 0.0456,
"num_input_tokens_seen": 1636192,
"step": 247
},
{
"epoch": 1.5961383748994369,
"grad_norm": 2.4742684364318848,
"learning_rate": 3.897982258676867e-06,
"loss": 0.0391,
"num_input_tokens_seen": 1642832,
"step": 248
},
{
"epoch": 1.6025744167337088,
"grad_norm": 4.165124893188477,
"learning_rate": 3.8894590618806435e-06,
"loss": 0.0501,
"num_input_tokens_seen": 1649904,
"step": 249
},
{
"epoch": 1.6090104585679805,
"grad_norm": 2.7913286685943604,
"learning_rate": 3.880912432401265e-06,
"loss": 0.0397,
"num_input_tokens_seen": 1656704,
"step": 250
},
{
"epoch": 1.6154465004022525,
"grad_norm": 4.8400397300720215,
"learning_rate": 3.872342514374291e-06,
"loss": 0.0846,
"num_input_tokens_seen": 1663680,
"step": 251
},
{
"epoch": 1.6218825422365244,
"grad_norm": 3.111396074295044,
"learning_rate": 3.863749452328035e-06,
"loss": 0.0443,
"num_input_tokens_seen": 1670160,
"step": 252
},
{
"epoch": 1.6283185840707963,
"grad_norm": 3.1794304847717285,
"learning_rate": 3.855133391181124e-06,
"loss": 0.045,
"num_input_tokens_seen": 1676832,
"step": 253
},
{
"epoch": 1.6347546259050683,
"grad_norm": 1.6655223369598389,
"learning_rate": 3.846494476240057e-06,
"loss": 0.0172,
"num_input_tokens_seen": 1683664,
"step": 254
},
{
"epoch": 1.6411906677393402,
"grad_norm": 4.251989841461182,
"learning_rate": 3.837832853196751e-06,
"loss": 0.0949,
"num_input_tokens_seen": 1690208,
"step": 255
},
{
"epoch": 1.6476267095736121,
"grad_norm": 7.070593357086182,
"learning_rate": 3.8291486681260904e-06,
"loss": 0.0277,
"num_input_tokens_seen": 1697296,
"step": 256
},
{
"epoch": 1.654062751407884,
"grad_norm": 2.8217155933380127,
"learning_rate": 3.820442067483455e-06,
"loss": 0.0247,
"num_input_tokens_seen": 1703504,
"step": 257
},
{
"epoch": 1.660498793242156,
"grad_norm": 5.125271320343018,
"learning_rate": 3.811713198102258e-06,
"loss": 0.0549,
"num_input_tokens_seen": 1710016,
"step": 258
},
{
"epoch": 1.666934835076428,
"grad_norm": 5.227617263793945,
"learning_rate": 3.802962207191463e-06,
"loss": 0.0342,
"num_input_tokens_seen": 1716960,
"step": 259
},
{
"epoch": 1.6733708769106999,
"grad_norm": 3.3697738647460938,
"learning_rate": 3.794189242333107e-06,
"loss": 0.0617,
"num_input_tokens_seen": 1723504,
"step": 260
},
{
"epoch": 1.6798069187449718,
"grad_norm": 2.9104015827178955,
"learning_rate": 3.785394451479806e-06,
"loss": 0.0675,
"num_input_tokens_seen": 1730160,
"step": 261
},
{
"epoch": 1.6862429605792437,
"grad_norm": 4.513949394226074,
"learning_rate": 3.7765779829522674e-06,
"loss": 0.1055,
"num_input_tokens_seen": 1736752,
"step": 262
},
{
"epoch": 1.6926790024135157,
"grad_norm": 3.0852975845336914,
"learning_rate": 3.7677399854367815e-06,
"loss": 0.0355,
"num_input_tokens_seen": 1743328,
"step": 263
},
{
"epoch": 1.6991150442477876,
"grad_norm": 3.222297191619873,
"learning_rate": 3.7588806079827147e-06,
"loss": 0.0622,
"num_input_tokens_seen": 1749776,
"step": 264
},
{
"epoch": 1.7055510860820595,
"grad_norm": 2.017244338989258,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.0297,
"num_input_tokens_seen": 1756512,
"step": 265
},
{
"epoch": 1.7119871279163315,
"grad_norm": 2.465116262435913,
"learning_rate": 3.7410983112566166e-06,
"loss": 0.0312,
"num_input_tokens_seen": 1762928,
"step": 266
},
{
"epoch": 1.7184231697506034,
"grad_norm": 2.8471832275390625,
"learning_rate": 3.7321756918760587e-06,
"loss": 0.0811,
"num_input_tokens_seen": 1769392,
"step": 267
},
{
"epoch": 1.7248592115848753,
"grad_norm": 3.4750540256500244,
"learning_rate": 3.7232322923348093e-06,
"loss": 0.067,
"num_input_tokens_seen": 1776032,
"step": 268
},
{
"epoch": 1.7312952534191473,
"grad_norm": 2.845557928085327,
"learning_rate": 3.7142682634598016e-06,
"loss": 0.0553,
"num_input_tokens_seen": 1782512,
"step": 269
},
{
"epoch": 1.7377312952534192,
"grad_norm": 2.0945403575897217,
"learning_rate": 3.7052837564258728e-06,
"loss": 0.021,
"num_input_tokens_seen": 1789280,
"step": 270
},
{
"epoch": 1.7441673370876911,
"grad_norm": 2.614729642868042,
"learning_rate": 3.6962789227532165e-06,
"loss": 0.0589,
"num_input_tokens_seen": 1795696,
"step": 271
},
{
"epoch": 1.750603378921963,
"grad_norm": 3.331339120864868,
"learning_rate": 3.6872539143048287e-06,
"loss": 0.0521,
"num_input_tokens_seen": 1802448,
"step": 272
},
{
"epoch": 1.757039420756235,
"grad_norm": 2.845620632171631,
"learning_rate": 3.6782088832839436e-06,
"loss": 0.0402,
"num_input_tokens_seen": 1809264,
"step": 273
},
{
"epoch": 1.763475462590507,
"grad_norm": 3.3971211910247803,
"learning_rate": 3.6691439822314672e-06,
"loss": 0.0363,
"num_input_tokens_seen": 1815808,
"step": 274
},
{
"epoch": 1.7699115044247788,
"grad_norm": 5.249027729034424,
"learning_rate": 3.660059364023409e-06,
"loss": 0.0523,
"num_input_tokens_seen": 1822352,
"step": 275
},
{
"epoch": 1.7763475462590508,
"grad_norm": 3.6546497344970703,
"learning_rate": 3.650955181868298e-06,
"loss": 0.0255,
"num_input_tokens_seen": 1829056,
"step": 276
},
{
"epoch": 1.7827835880933227,
"grad_norm": 7.767543792724609,
"learning_rate": 3.641831589304602e-06,
"loss": 0.1031,
"num_input_tokens_seen": 1835696,
"step": 277
},
{
"epoch": 1.7892196299275946,
"grad_norm": 1.5550068616867065,
"learning_rate": 3.6326887401981386e-06,
"loss": 0.0452,
"num_input_tokens_seen": 1842288,
"step": 278
},
{
"epoch": 1.7956556717618666,
"grad_norm": 4.8318986892700195,
"learning_rate": 3.6235267887394774e-06,
"loss": 0.0537,
"num_input_tokens_seen": 1848960,
"step": 279
},
{
"epoch": 1.8020917135961385,
"grad_norm": 4.691814422607422,
"learning_rate": 3.6143458894413463e-06,
"loss": 0.0572,
"num_input_tokens_seen": 1855648,
"step": 280
},
{
"epoch": 1.8085277554304104,
"grad_norm": 2.6937472820281982,
"learning_rate": 3.6051461971360146e-06,
"loss": 0.0298,
"num_input_tokens_seen": 1862160,
"step": 281
},
{
"epoch": 1.8149637972646824,
"grad_norm": 4.052839279174805,
"learning_rate": 3.595927866972694e-06,
"loss": 0.037,
"num_input_tokens_seen": 1868896,
"step": 282
},
{
"epoch": 1.8213998390989543,
"grad_norm": 5.030338287353516,
"learning_rate": 3.586691054414913e-06,
"loss": 0.0783,
"num_input_tokens_seen": 1875248,
"step": 283
},
{
"epoch": 1.827835880933226,
"grad_norm": 1.9826079607009888,
"learning_rate": 3.577435915237899e-06,
"loss": 0.0436,
"num_input_tokens_seen": 1881728,
"step": 284
},
{
"epoch": 1.834271922767498,
"grad_norm": 1.8905837535858154,
"learning_rate": 3.5681626055259526e-06,
"loss": 0.0258,
"num_input_tokens_seen": 1888384,
"step": 285
},
{
"epoch": 1.8407079646017699,
"grad_norm": 1.9678194522857666,
"learning_rate": 3.558871281669811e-06,
"loss": 0.0235,
"num_input_tokens_seen": 1894864,
"step": 286
},
{
"epoch": 1.8471440064360418,
"grad_norm": 4.199605464935303,
"learning_rate": 3.549562100364014e-06,
"loss": 0.0541,
"num_input_tokens_seen": 1901680,
"step": 287
},
{
"epoch": 1.8535800482703138,
"grad_norm": 4.100510120391846,
"learning_rate": 3.5402352186042602e-06,
"loss": 0.0767,
"num_input_tokens_seen": 1908304,
"step": 288
},
{
"epoch": 1.8600160901045857,
"grad_norm": 6.471580982208252,
"learning_rate": 3.530890793684759e-06,
"loss": 0.0558,
"num_input_tokens_seen": 1914736,
"step": 289
},
{
"epoch": 1.8664521319388576,
"grad_norm": 6.2181525230407715,
"learning_rate": 3.521528983195579e-06,
"loss": 0.0483,
"num_input_tokens_seen": 1921088,
"step": 290
},
{
"epoch": 1.8728881737731295,
"grad_norm": 3.5814297199249268,
"learning_rate": 3.512149945019989e-06,
"loss": 0.0389,
"num_input_tokens_seen": 1927408,
"step": 291
},
{
"epoch": 1.8793242156074015,
"grad_norm": 3.193094491958618,
"learning_rate": 3.502753837331797e-06,
"loss": 0.034,
"num_input_tokens_seen": 1934160,
"step": 292
},
{
"epoch": 1.8857602574416734,
"grad_norm": 3.2676048278808594,
"learning_rate": 3.4933408185926805e-06,
"loss": 0.0921,
"num_input_tokens_seen": 1940912,
"step": 293
},
{
"epoch": 1.8921962992759453,
"grad_norm": 4.060972690582275,
"learning_rate": 3.4839110475495153e-06,
"loss": 0.0661,
"num_input_tokens_seen": 1947488,
"step": 294
},
{
"epoch": 1.898632341110217,
"grad_norm": 4.40585470199585,
"learning_rate": 3.4744646832316985e-06,
"loss": 0.0301,
"num_input_tokens_seen": 1954000,
"step": 295
},
{
"epoch": 1.905068382944489,
"grad_norm": 4.472731113433838,
"learning_rate": 3.465001884948468e-06,
"loss": 0.0878,
"num_input_tokens_seen": 1960400,
"step": 296
},
{
"epoch": 1.911504424778761,
"grad_norm": 3.2221555709838867,
"learning_rate": 3.45552281228621e-06,
"loss": 0.1126,
"num_input_tokens_seen": 1967728,
"step": 297
},
{
"epoch": 1.9179404666130329,
"grad_norm": 3.6210269927978516,
"learning_rate": 3.446027625105776e-06,
"loss": 0.0679,
"num_input_tokens_seen": 1974096,
"step": 298
},
{
"epoch": 1.9243765084473048,
"grad_norm": 2.038454055786133,
"learning_rate": 3.436516483539781e-06,
"loss": 0.031,
"num_input_tokens_seen": 1980672,
"step": 299
},
{
"epoch": 1.9308125502815767,
"grad_norm": 2.2427828311920166,
"learning_rate": 3.4269895479899023e-06,
"loss": 0.0687,
"num_input_tokens_seen": 1987104,
"step": 300
},
{
"epoch": 1.9372485921158487,
"grad_norm": 6.37827730178833,
"learning_rate": 3.4174469791241805e-06,
"loss": 0.0497,
"num_input_tokens_seen": 1994064,
"step": 301
},
{
"epoch": 1.9436846339501206,
"grad_norm": 9.542262077331543,
"learning_rate": 3.4078889378743036e-06,
"loss": 0.0829,
"num_input_tokens_seen": 2001056,
"step": 302
},
{
"epoch": 1.9501206757843925,
"grad_norm": 6.237174034118652,
"learning_rate": 3.3983155854328942e-06,
"loss": 0.0578,
"num_input_tokens_seen": 2007712,
"step": 303
},
{
"epoch": 1.9565567176186645,
"grad_norm": 2.3653266429901123,
"learning_rate": 3.388727083250795e-06,
"loss": 0.0398,
"num_input_tokens_seen": 2014368,
"step": 304
},
{
"epoch": 1.9629927594529364,
"grad_norm": 3.9448723793029785,
"learning_rate": 3.379123593034342e-06,
"loss": 0.0754,
"num_input_tokens_seen": 2020592,
"step": 305
},
{
"epoch": 1.9694288012872083,
"grad_norm": 2.1158804893493652,
"learning_rate": 3.369505276742638e-06,
"loss": 0.0395,
"num_input_tokens_seen": 2026864,
"step": 306
},
{
"epoch": 1.9758648431214803,
"grad_norm": 5.131661891937256,
"learning_rate": 3.359872296584821e-06,
"loss": 0.0575,
"num_input_tokens_seen": 2033440,
"step": 307
},
{
"epoch": 1.9823008849557522,
"grad_norm": 5.0783867835998535,
"learning_rate": 3.350224815017331e-06,
"loss": 0.0472,
"num_input_tokens_seen": 2039712,
"step": 308
},
{
"epoch": 1.9887369267900241,
"grad_norm": 6.688424587249756,
"learning_rate": 3.3405629947411687e-06,
"loss": 0.0498,
"num_input_tokens_seen": 2046576,
"step": 309
},
{
"epoch": 1.995172968624296,
"grad_norm": 5.24268102645874,
"learning_rate": 3.3308869986991493e-06,
"loss": 0.0447,
"num_input_tokens_seen": 2053248,
"step": 310
},
{
"epoch": 2.001609010458568,
"grad_norm": 1.7300570011138916,
"learning_rate": 3.32119699007316e-06,
"loss": 0.0155,
"num_input_tokens_seen": 2059840,
"step": 311
},
{
"epoch": 2.00804505229284,
"grad_norm": 2.5391845703125,
"learning_rate": 3.311493132281402e-06,
"loss": 0.0183,
"num_input_tokens_seen": 2066384,
"step": 312
},
{
"epoch": 2.014481094127112,
"grad_norm": 1.9404152631759644,
"learning_rate": 3.3017755889756382e-06,
"loss": 0.0102,
"num_input_tokens_seen": 2073088,
"step": 313
},
{
"epoch": 2.020917135961384,
"grad_norm": 1.1497960090637207,
"learning_rate": 3.292044524038433e-06,
"loss": 0.0119,
"num_input_tokens_seen": 2079600,
"step": 314
},
{
"epoch": 2.0273531777956557,
"grad_norm": 0.6188907027244568,
"learning_rate": 3.2823001015803863e-06,
"loss": 0.0037,
"num_input_tokens_seen": 2086080,
"step": 315
},
{
"epoch": 2.0337892196299276,
"grad_norm": 2.5652434825897217,
"learning_rate": 3.272542485937369e-06,
"loss": 0.0048,
"num_input_tokens_seen": 2092768,
"step": 316
},
{
"epoch": 2.0402252614641996,
"grad_norm": 1.3636257648468018,
"learning_rate": 3.2627718416677484e-06,
"loss": 0.004,
"num_input_tokens_seen": 2099296,
"step": 317
},
{
"epoch": 2.0466613032984715,
"grad_norm": 3.7406702041625977,
"learning_rate": 3.2529883335496163e-06,
"loss": 0.0472,
"num_input_tokens_seen": 2106176,
"step": 318
},
{
"epoch": 2.0530973451327434,
"grad_norm": 0.2876489460468292,
"learning_rate": 3.243192126578007e-06,
"loss": 0.0008,
"num_input_tokens_seen": 2112560,
"step": 319
},
{
"epoch": 2.0595333869670154,
"grad_norm": 3.388899087905884,
"learning_rate": 3.2333833859621155e-06,
"loss": 0.0332,
"num_input_tokens_seen": 2119296,
"step": 320
},
{
"epoch": 2.0659694288012873,
"grad_norm": 2.6212401390075684,
"learning_rate": 3.223562277122513e-06,
"loss": 0.0434,
"num_input_tokens_seen": 2125632,
"step": 321
},
{
"epoch": 2.0724054706355592,
"grad_norm": 3.6854021549224854,
"learning_rate": 3.213728965688356e-06,
"loss": 0.0105,
"num_input_tokens_seen": 2132096,
"step": 322
},
{
"epoch": 2.078841512469831,
"grad_norm": 3.9269893169403076,
"learning_rate": 3.2038836174945907e-06,
"loss": 0.0188,
"num_input_tokens_seen": 2138336,
"step": 323
},
{
"epoch": 2.085277554304103,
"grad_norm": 2.3363194465637207,
"learning_rate": 3.194026398579162e-06,
"loss": 0.0382,
"num_input_tokens_seen": 2144672,
"step": 324
},
{
"epoch": 2.091713596138375,
"grad_norm": 0.16176919639110565,
"learning_rate": 3.184157475180208e-06,
"loss": 0.0002,
"num_input_tokens_seen": 2151216,
"step": 325
},
{
"epoch": 2.098149637972647,
"grad_norm": 7.4007368087768555,
"learning_rate": 3.1742770137332567e-06,
"loss": 0.0473,
"num_input_tokens_seen": 2158000,
"step": 326
},
{
"epoch": 2.104585679806919,
"grad_norm": 0.3990660607814789,
"learning_rate": 3.164385180868425e-06,
"loss": 0.0008,
"num_input_tokens_seen": 2164448,
"step": 327
},
{
"epoch": 2.111021721641191,
"grad_norm": 5.447741508483887,
"learning_rate": 3.1544821434076013e-06,
"loss": 0.0123,
"num_input_tokens_seen": 2171120,
"step": 328
},
{
"epoch": 2.1174577634754628,
"grad_norm": 4.229776382446289,
"learning_rate": 3.144568068361634e-06,
"loss": 0.03,
"num_input_tokens_seen": 2177648,
"step": 329
},
{
"epoch": 2.1238938053097347,
"grad_norm": 5.920961380004883,
"learning_rate": 3.1346431229275197e-06,
"loss": 0.0207,
"num_input_tokens_seen": 2183856,
"step": 330
},
{
"epoch": 2.1303298471440066,
"grad_norm": 11.779773712158203,
"learning_rate": 3.124707474485577e-06,
"loss": 0.0172,
"num_input_tokens_seen": 2190608,
"step": 331
},
{
"epoch": 2.136765888978278,
"grad_norm": 8.82557201385498,
"learning_rate": 3.1147612905966286e-06,
"loss": 0.0115,
"num_input_tokens_seen": 2197232,
"step": 332
},
{
"epoch": 2.14320193081255,
"grad_norm": 1.1176470518112183,
"learning_rate": 3.1048047389991693e-06,
"loss": 0.0217,
"num_input_tokens_seen": 2203456,
"step": 333
},
{
"epoch": 2.149637972646822,
"grad_norm": 0.8806192278862,
"learning_rate": 3.094837987606547e-06,
"loss": 0.0035,
"num_input_tokens_seen": 2209856,
"step": 334
},
{
"epoch": 2.156074014481094,
"grad_norm": 6.793837547302246,
"learning_rate": 3.084861204504122e-06,
"loss": 0.0426,
"num_input_tokens_seen": 2216400,
"step": 335
},
{
"epoch": 2.162510056315366,
"grad_norm": 5.5410075187683105,
"learning_rate": 3.0748745579464347e-06,
"loss": 0.0382,
"num_input_tokens_seen": 2222864,
"step": 336
},
{
"epoch": 2.168946098149638,
"grad_norm": 1.8118884563446045,
"learning_rate": 3.0648782163543696e-06,
"loss": 0.0082,
"num_input_tokens_seen": 2229760,
"step": 337
},
{
"epoch": 2.1753821399839097,
"grad_norm": 2.607206106185913,
"learning_rate": 3.0548723483123157e-06,
"loss": 0.0338,
"num_input_tokens_seen": 2236368,
"step": 338
},
{
"epoch": 2.1818181818181817,
"grad_norm": 1.080344557762146,
"learning_rate": 3.0448571225653195e-06,
"loss": 0.0141,
"num_input_tokens_seen": 2242816,
"step": 339
},
{
"epoch": 2.1882542236524536,
"grad_norm": 2.380739212036133,
"learning_rate": 3.0348327080162438e-06,
"loss": 0.0287,
"num_input_tokens_seen": 2249488,
"step": 340
},
{
"epoch": 2.1946902654867255,
"grad_norm": 1.0098868608474731,
"learning_rate": 3.0247992737229147e-06,
"loss": 0.0027,
"num_input_tokens_seen": 2255968,
"step": 341
},
{
"epoch": 2.2011263073209975,
"grad_norm": 0.7940512895584106,
"learning_rate": 3.014756988895275e-06,
"loss": 0.0026,
"num_input_tokens_seen": 2262544,
"step": 342
},
{
"epoch": 2.2075623491552694,
"grad_norm": 2.9759926795959473,
"learning_rate": 3.0047060228925256e-06,
"loss": 0.039,
"num_input_tokens_seen": 2269312,
"step": 343
},
{
"epoch": 2.2139983909895413,
"grad_norm": 4.84032678604126,
"learning_rate": 2.994646545220275e-06,
"loss": 0.0154,
"num_input_tokens_seen": 2275968,
"step": 344
},
{
"epoch": 2.2204344328238133,
"grad_norm": 2.9671568870544434,
"learning_rate": 2.9845787255276753e-06,
"loss": 0.0231,
"num_input_tokens_seen": 2282976,
"step": 345
},
{
"epoch": 2.226870474658085,
"grad_norm": 5.410647392272949,
"learning_rate": 2.9745027336045652e-06,
"loss": 0.04,
"num_input_tokens_seen": 2289696,
"step": 346
},
{
"epoch": 2.233306516492357,
"grad_norm": 5.828602313995361,
"learning_rate": 2.964418739378603e-06,
"loss": 0.0282,
"num_input_tokens_seen": 2296272,
"step": 347
},
{
"epoch": 2.239742558326629,
"grad_norm": 1.9481452703475952,
"learning_rate": 2.954326912912404e-06,
"loss": 0.0143,
"num_input_tokens_seen": 2303120,
"step": 348
},
{
"epoch": 2.246178600160901,
"grad_norm": 3.2762415409088135,
"learning_rate": 2.9442274244006725e-06,
"loss": 0.0194,
"num_input_tokens_seen": 2309728,
"step": 349
},
{
"epoch": 2.252614641995173,
"grad_norm": 2.3237709999084473,
"learning_rate": 2.9341204441673267e-06,
"loss": 0.0051,
"num_input_tokens_seen": 2316144,
"step": 350
},
{
"epoch": 2.259050683829445,
"grad_norm": 1.7801238298416138,
"learning_rate": 2.924006142662632e-06,
"loss": 0.0162,
"num_input_tokens_seen": 2322768,
"step": 351
},
{
"epoch": 2.265486725663717,
"grad_norm": 4.876129150390625,
"learning_rate": 2.913884690460325e-06,
"loss": 0.0313,
"num_input_tokens_seen": 2329312,
"step": 352
},
{
"epoch": 2.2719227674979887,
"grad_norm": 0.9637519717216492,
"learning_rate": 2.903756258254734e-06,
"loss": 0.0041,
"num_input_tokens_seen": 2335824,
"step": 353
},
{
"epoch": 2.2783588093322606,
"grad_norm": 2.7481493949890137,
"learning_rate": 2.8936210168579043e-06,
"loss": 0.0321,
"num_input_tokens_seen": 2342272,
"step": 354
},
{
"epoch": 2.2847948511665326,
"grad_norm": 1.682763934135437,
"learning_rate": 2.883479137196714e-06,
"loss": 0.0064,
"num_input_tokens_seen": 2349056,
"step": 355
},
{
"epoch": 2.2912308930008045,
"grad_norm": 5.632142066955566,
"learning_rate": 2.8733307903099926e-06,
"loss": 0.0237,
"num_input_tokens_seen": 2355552,
"step": 356
},
{
"epoch": 2.2976669348350764,
"grad_norm": 2.460470199584961,
"learning_rate": 2.8631761473456377e-06,
"loss": 0.0152,
"num_input_tokens_seen": 2361808,
"step": 357
},
{
"epoch": 2.3041029766693484,
"grad_norm": 0.9998040199279785,
"learning_rate": 2.853015379557729e-06,
"loss": 0.0038,
"num_input_tokens_seen": 2368288,
"step": 358
},
{
"epoch": 2.3105390185036203,
"grad_norm": 3.164407968521118,
"learning_rate": 2.842848658303637e-06,
"loss": 0.0168,
"num_input_tokens_seen": 2374960,
"step": 359
},
{
"epoch": 2.3169750603378922,
"grad_norm": 2.3879611492156982,
"learning_rate": 2.832676155041135e-06,
"loss": 0.0049,
"num_input_tokens_seen": 2381776,
"step": 360
},
{
"epoch": 2.323411102172164,
"grad_norm": 1.3164470195770264,
"learning_rate": 2.822498041325509e-06,
"loss": 0.0114,
"num_input_tokens_seen": 2388112,
"step": 361
},
{
"epoch": 2.329847144006436,
"grad_norm": 2.3726656436920166,
"learning_rate": 2.8123144888066623e-06,
"loss": 0.022,
"num_input_tokens_seen": 2394736,
"step": 362
},
{
"epoch": 2.336283185840708,
"grad_norm": 1.7789826393127441,
"learning_rate": 2.802125669226222e-06,
"loss": 0.0154,
"num_input_tokens_seen": 2401248,
"step": 363
},
{
"epoch": 2.34271922767498,
"grad_norm": 3.68959641456604,
"learning_rate": 2.7919317544146405e-06,
"loss": 0.0204,
"num_input_tokens_seen": 2407872,
"step": 364
},
{
"epoch": 2.349155269509252,
"grad_norm": 2.4927353858947754,
"learning_rate": 2.7817329162883033e-06,
"loss": 0.0334,
"num_input_tokens_seen": 2414432,
"step": 365
},
{
"epoch": 2.355591311343524,
"grad_norm": 4.594964504241943,
"learning_rate": 2.7715293268466204e-06,
"loss": 0.0132,
"num_input_tokens_seen": 2420848,
"step": 366
},
{
"epoch": 2.3620273531777958,
"grad_norm": 4.325422286987305,
"learning_rate": 2.761321158169134e-06,
"loss": 0.0291,
"num_input_tokens_seen": 2427728,
"step": 367
},
{
"epoch": 2.3684633950120677,
"grad_norm": 2.46122407913208,
"learning_rate": 2.7511085824126133e-06,
"loss": 0.0089,
"num_input_tokens_seen": 2434880,
"step": 368
},
{
"epoch": 2.3748994368463396,
"grad_norm": 2.729311227798462,
"learning_rate": 2.74089177180815e-06,
"loss": 0.0306,
"num_input_tokens_seen": 2441168,
"step": 369
},
{
"epoch": 2.3813354786806116,
"grad_norm": 5.095163345336914,
"learning_rate": 2.730670898658255e-06,
"loss": 0.0297,
"num_input_tokens_seen": 2447920,
"step": 370
},
{
"epoch": 2.3877715205148835,
"grad_norm": 1.902287483215332,
"learning_rate": 2.7204461353339546e-06,
"loss": 0.0247,
"num_input_tokens_seen": 2454704,
"step": 371
},
{
"epoch": 2.3942075623491554,
"grad_norm": 3.267244577407837,
"learning_rate": 2.7102176542718783e-06,
"loss": 0.0234,
"num_input_tokens_seen": 2461216,
"step": 372
},
{
"epoch": 2.4006436041834274,
"grad_norm": 4.101126670837402,
"learning_rate": 2.699985627971354e-06,
"loss": 0.0192,
"num_input_tokens_seen": 2468032,
"step": 373
},
{
"epoch": 2.4070796460176993,
"grad_norm": 4.104948997497559,
"learning_rate": 2.689750228991503e-06,
"loss": 0.0324,
"num_input_tokens_seen": 2474544,
"step": 374
},
{
"epoch": 2.4135156878519712,
"grad_norm": 2.1446776390075684,
"learning_rate": 2.679511629948319e-06,
"loss": 0.0332,
"num_input_tokens_seen": 2481312,
"step": 375
},
{
"epoch": 2.419951729686243,
"grad_norm": 0.7457873225212097,
"learning_rate": 2.669270003511769e-06,
"loss": 0.0043,
"num_input_tokens_seen": 2487888,
"step": 376
},
{
"epoch": 2.426387771520515,
"grad_norm": 2.1420276165008545,
"learning_rate": 2.6590255224028725e-06,
"loss": 0.0197,
"num_input_tokens_seen": 2494784,
"step": 377
},
{
"epoch": 2.432823813354787,
"grad_norm": 3.0415239334106445,
"learning_rate": 2.648778359390794e-06,
"loss": 0.0366,
"num_input_tokens_seen": 2501712,
"step": 378
},
{
"epoch": 2.439259855189059,
"grad_norm": 3.6502788066864014,
"learning_rate": 2.638528687289925e-06,
"loss": 0.0173,
"num_input_tokens_seen": 2508592,
"step": 379
},
{
"epoch": 2.445695897023331,
"grad_norm": 2.2913506031036377,
"learning_rate": 2.6282766789569742e-06,
"loss": 0.0102,
"num_input_tokens_seen": 2515216,
"step": 380
},
{
"epoch": 2.4521319388576024,
"grad_norm": 3.3507297039031982,
"learning_rate": 2.618022507288049e-06,
"loss": 0.0361,
"num_input_tokens_seen": 2522064,
"step": 381
},
{
"epoch": 2.4585679806918743,
"grad_norm": 2.98098087310791,
"learning_rate": 2.6077663452157398e-06,
"loss": 0.0292,
"num_input_tokens_seen": 2528608,
"step": 382
},
{
"epoch": 2.4650040225261463,
"grad_norm": 1.4962135553359985,
"learning_rate": 2.5975083657062043e-06,
"loss": 0.0095,
"num_input_tokens_seen": 2535328,
"step": 383
},
{
"epoch": 2.471440064360418,
"grad_norm": 2.0819742679595947,
"learning_rate": 2.587248741756253e-06,
"loss": 0.015,
"num_input_tokens_seen": 2542224,
"step": 384
},
{
"epoch": 2.47787610619469,
"grad_norm": 1.8906433582305908,
"learning_rate": 2.576987646390426e-06,
"loss": 0.0276,
"num_input_tokens_seen": 2548976,
"step": 385
},
{
"epoch": 2.484312148028962,
"grad_norm": 2.451510190963745,
"learning_rate": 2.566725252658081e-06,
"loss": 0.0284,
"num_input_tokens_seen": 2555568,
"step": 386
},
{
"epoch": 2.490748189863234,
"grad_norm": 3.7337939739227295,
"learning_rate": 2.5564617336304703e-06,
"loss": 0.0366,
"num_input_tokens_seen": 2562128,
"step": 387
},
{
"epoch": 2.497184231697506,
"grad_norm": 1.6401593685150146,
"learning_rate": 2.546197262397825e-06,
"loss": 0.0322,
"num_input_tokens_seen": 2568640,
"step": 388
},
{
"epoch": 2.503620273531778,
"grad_norm": 0.9136457443237305,
"learning_rate": 2.535932012066434e-06,
"loss": 0.0057,
"num_input_tokens_seen": 2575024,
"step": 389
},
{
"epoch": 2.51005631536605,
"grad_norm": 1.119612455368042,
"learning_rate": 2.525666155755725e-06,
"loss": 0.0054,
"num_input_tokens_seen": 2581520,
"step": 390
},
{
"epoch": 2.5164923572003217,
"grad_norm": 2.4770889282226562,
"learning_rate": 2.515399866595347e-06,
"loss": 0.0199,
"num_input_tokens_seen": 2588528,
"step": 391
},
{
"epoch": 2.5229283990345936,
"grad_norm": 0.35335639119148254,
"learning_rate": 2.5051333177222476e-06,
"loss": 0.0045,
"num_input_tokens_seen": 2594992,
"step": 392
},
{
"epoch": 2.5293644408688656,
"grad_norm": 2.8933093547821045,
"learning_rate": 2.4948666822777536e-06,
"loss": 0.0283,
"num_input_tokens_seen": 2601568,
"step": 393
},
{
"epoch": 2.5358004827031375,
"grad_norm": 1.7032990455627441,
"learning_rate": 2.4846001334046537e-06,
"loss": 0.0248,
"num_input_tokens_seen": 2608160,
"step": 394
},
{
"epoch": 2.5422365245374094,
"grad_norm": 1.9688091278076172,
"learning_rate": 2.474333844244276e-06,
"loss": 0.0132,
"num_input_tokens_seen": 2614656,
"step": 395
},
{
"epoch": 2.5486725663716814,
"grad_norm": 3.135990619659424,
"learning_rate": 2.464067987933567e-06,
"loss": 0.04,
"num_input_tokens_seen": 2621600,
"step": 396
},
{
"epoch": 2.5551086082059533,
"grad_norm": 0.7140212059020996,
"learning_rate": 2.453802737602176e-06,
"loss": 0.0029,
"num_input_tokens_seen": 2627984,
"step": 397
},
{
"epoch": 2.5615446500402252,
"grad_norm": 3.9643640518188477,
"learning_rate": 2.4435382663695305e-06,
"loss": 0.0254,
"num_input_tokens_seen": 2634720,
"step": 398
},
{
"epoch": 2.567980691874497,
"grad_norm": 2.284302234649658,
"learning_rate": 2.4332747473419193e-06,
"loss": 0.0108,
"num_input_tokens_seen": 2641456,
"step": 399
},
{
"epoch": 2.574416733708769,
"grad_norm": 2.6400082111358643,
"learning_rate": 2.4230123536095746e-06,
"loss": 0.0269,
"num_input_tokens_seen": 2647760,
"step": 400
},
{
"epoch": 2.580852775543041,
"grad_norm": 3.1969995498657227,
"learning_rate": 2.4127512582437486e-06,
"loss": 0.0111,
"num_input_tokens_seen": 2654608,
"step": 401
},
{
"epoch": 2.587288817377313,
"grad_norm": 3.651118516921997,
"learning_rate": 2.4024916342937966e-06,
"loss": 0.0222,
"num_input_tokens_seen": 2661072,
"step": 402
},
{
"epoch": 2.593724859211585,
"grad_norm": 2.1281003952026367,
"learning_rate": 2.392233654784262e-06,
"loss": 0.0101,
"num_input_tokens_seen": 2667712,
"step": 403
},
{
"epoch": 2.600160901045857,
"grad_norm": 2.6782784461975098,
"learning_rate": 2.3819774927119523e-06,
"loss": 0.0138,
"num_input_tokens_seen": 2674496,
"step": 404
},
{
"epoch": 2.6065969428801288,
"grad_norm": 2.2902138233184814,
"learning_rate": 2.3717233210430258e-06,
"loss": 0.0281,
"num_input_tokens_seen": 2680816,
"step": 405
},
{
"epoch": 2.6130329847144007,
"grad_norm": 1.9150536060333252,
"learning_rate": 2.3614713127100752e-06,
"loss": 0.0042,
"num_input_tokens_seen": 2687632,
"step": 406
},
{
"epoch": 2.6194690265486726,
"grad_norm": 0.3568836748600006,
"learning_rate": 2.3512216406092066e-06,
"loss": 0.0015,
"num_input_tokens_seen": 2694464,
"step": 407
},
{
"epoch": 2.6259050683829446,
"grad_norm": 2.3506011962890625,
"learning_rate": 2.340974477597128e-06,
"loss": 0.0279,
"num_input_tokens_seen": 2701344,
"step": 408
},
{
"epoch": 2.6323411102172165,
"grad_norm": 2.780200481414795,
"learning_rate": 2.3307299964882314e-06,
"loss": 0.0399,
"num_input_tokens_seen": 2707536,
"step": 409
},
{
"epoch": 2.6387771520514884,
"grad_norm": 1.1793303489685059,
"learning_rate": 2.3204883700516813e-06,
"loss": 0.0074,
"num_input_tokens_seen": 2714544,
"step": 410
},
{
"epoch": 2.6452131938857604,
"grad_norm": 1.7807022333145142,
"learning_rate": 2.310249771008498e-06,
"loss": 0.0078,
"num_input_tokens_seen": 2721056,
"step": 411
},
{
"epoch": 2.6516492357200323,
"grad_norm": 12.764676094055176,
"learning_rate": 2.3000143720286463e-06,
"loss": 0.0406,
"num_input_tokens_seen": 2727664,
"step": 412
},
{
"epoch": 2.6580852775543042,
"grad_norm": 0.44338610768318176,
"learning_rate": 2.2897823457281225e-06,
"loss": 0.0023,
"num_input_tokens_seen": 2733600,
"step": 413
},
{
"epoch": 2.664521319388576,
"grad_norm": 3.5756232738494873,
"learning_rate": 2.2795538646660462e-06,
"loss": 0.006,
"num_input_tokens_seen": 2740400,
"step": 414
},
{
"epoch": 2.670957361222848,
"grad_norm": 1.4301191568374634,
"learning_rate": 2.269329101341745e-06,
"loss": 0.0236,
"num_input_tokens_seen": 2747248,
"step": 415
},
{
"epoch": 2.67739340305712,
"grad_norm": 2.0859804153442383,
"learning_rate": 2.2591082281918507e-06,
"loss": 0.0136,
"num_input_tokens_seen": 2753776,
"step": 416
},
{
"epoch": 2.6838294448913915,
"grad_norm": 2.9704370498657227,
"learning_rate": 2.2488914175873876e-06,
"loss": 0.015,
"num_input_tokens_seen": 2760720,
"step": 417
},
{
"epoch": 2.6902654867256635,
"grad_norm": 3.1178269386291504,
"learning_rate": 2.238678841830867e-06,
"loss": 0.0483,
"num_input_tokens_seen": 2767136,
"step": 418
},
{
"epoch": 2.6967015285599354,
"grad_norm": 0.6049777269363403,
"learning_rate": 2.2284706731533805e-06,
"loss": 0.0014,
"num_input_tokens_seen": 2773680,
"step": 419
},
{
"epoch": 2.7031375703942073,
"grad_norm": 3.5615270137786865,
"learning_rate": 2.2182670837116975e-06,
"loss": 0.0279,
"num_input_tokens_seen": 2780160,
"step": 420
},
{
"epoch": 2.7095736122284793,
"grad_norm": 3.4241111278533936,
"learning_rate": 2.20806824558536e-06,
"loss": 0.0705,
"num_input_tokens_seen": 2786912,
"step": 421
},
{
"epoch": 2.716009654062751,
"grad_norm": 1.0644826889038086,
"learning_rate": 2.197874330773779e-06,
"loss": 0.005,
"num_input_tokens_seen": 2793888,
"step": 422
},
{
"epoch": 2.722445695897023,
"grad_norm": 5.071107387542725,
"learning_rate": 2.1876855111933385e-06,
"loss": 0.0453,
"num_input_tokens_seen": 2800320,
"step": 423
},
{
"epoch": 2.728881737731295,
"grad_norm": 1.9479647874832153,
"learning_rate": 2.1775019586744924e-06,
"loss": 0.0095,
"num_input_tokens_seen": 2807088,
"step": 424
},
{
"epoch": 2.735317779565567,
"grad_norm": 2.730952262878418,
"learning_rate": 2.167323844958867e-06,
"loss": 0.0095,
"num_input_tokens_seen": 2813312,
"step": 425
},
{
"epoch": 2.741753821399839,
"grad_norm": 2.1456387042999268,
"learning_rate": 2.1571513416963647e-06,
"loss": 0.0138,
"num_input_tokens_seen": 2819936,
"step": 426
},
{
"epoch": 2.748189863234111,
"grad_norm": 2.14911150932312,
"learning_rate": 2.1469846204422724e-06,
"loss": 0.0272,
"num_input_tokens_seen": 2826224,
"step": 427
},
{
"epoch": 2.754625905068383,
"grad_norm": 0.5957837700843811,
"learning_rate": 2.136823852654363e-06,
"loss": 0.0026,
"num_input_tokens_seen": 2832960,
"step": 428
},
{
"epoch": 2.7610619469026547,
"grad_norm": 0.3253982961177826,
"learning_rate": 2.126669209690008e-06,
"loss": 0.0016,
"num_input_tokens_seen": 2839888,
"step": 429
},
{
"epoch": 2.7674979887369267,
"grad_norm": 3.472017765045166,
"learning_rate": 2.1165208628032863e-06,
"loss": 0.0633,
"num_input_tokens_seen": 2846688,
"step": 430
},
{
"epoch": 2.7739340305711986,
"grad_norm": 2.029026985168457,
"learning_rate": 2.1063789831420957e-06,
"loss": 0.0191,
"num_input_tokens_seen": 2853184,
"step": 431
},
{
"epoch": 2.7803700724054705,
"grad_norm": 2.316349506378174,
"learning_rate": 2.096243741745266e-06,
"loss": 0.0075,
"num_input_tokens_seen": 2859632,
"step": 432
},
{
"epoch": 2.7868061142397424,
"grad_norm": 3.786245346069336,
"learning_rate": 2.086115309539675e-06,
"loss": 0.0371,
"num_input_tokens_seen": 2865920,
"step": 433
},
{
"epoch": 2.7932421560740144,
"grad_norm": 1.864402413368225,
"learning_rate": 2.0759938573373683e-06,
"loss": 0.0275,
"num_input_tokens_seen": 2872464,
"step": 434
},
{
"epoch": 2.7996781979082863,
"grad_norm": 8.142292022705078,
"learning_rate": 2.0658795558326745e-06,
"loss": 0.0441,
"num_input_tokens_seen": 2879168,
"step": 435
},
{
"epoch": 2.8061142397425582,
"grad_norm": 1.3945283889770508,
"learning_rate": 2.0557725755993283e-06,
"loss": 0.0074,
"num_input_tokens_seen": 2885520,
"step": 436
},
{
"epoch": 2.81255028157683,
"grad_norm": 1.951145887374878,
"learning_rate": 2.0456730870875964e-06,
"loss": 0.0412,
"num_input_tokens_seen": 2892368,
"step": 437
},
{
"epoch": 2.818986323411102,
"grad_norm": 1.666693925857544,
"learning_rate": 2.035581260621398e-06,
"loss": 0.007,
"num_input_tokens_seen": 2898640,
"step": 438
},
{
"epoch": 2.825422365245374,
"grad_norm": 0.8178473114967346,
"learning_rate": 2.0254972663954356e-06,
"loss": 0.0195,
"num_input_tokens_seen": 2905312,
"step": 439
},
{
"epoch": 2.831858407079646,
"grad_norm": 2.1499900817871094,
"learning_rate": 2.015421274472325e-06,
"loss": 0.0104,
"num_input_tokens_seen": 2911872,
"step": 440
},
{
"epoch": 2.838294448913918,
"grad_norm": 3.162245273590088,
"learning_rate": 2.005353454779726e-06,
"loss": 0.0196,
"num_input_tokens_seen": 2918496,
"step": 441
},
{
"epoch": 2.84473049074819,
"grad_norm": 1.1920592784881592,
"learning_rate": 1.995293977107475e-06,
"loss": 0.0131,
"num_input_tokens_seen": 2924944,
"step": 442
},
{
"epoch": 2.8511665325824618,
"grad_norm": 1.091436743736267,
"learning_rate": 1.9852430111047254e-06,
"loss": 0.0072,
"num_input_tokens_seen": 2931440,
"step": 443
},
{
"epoch": 2.8576025744167337,
"grad_norm": 2.0469212532043457,
"learning_rate": 1.9752007262770857e-06,
"loss": 0.0058,
"num_input_tokens_seen": 2938304,
"step": 444
},
{
"epoch": 2.8640386162510056,
"grad_norm": 1.6995850801467896,
"learning_rate": 1.965167291983757e-06,
"loss": 0.0242,
"num_input_tokens_seen": 2945168,
"step": 445
},
{
"epoch": 2.8704746580852776,
"grad_norm": 5.4955735206604,
"learning_rate": 1.955142877434681e-06,
"loss": 0.0323,
"num_input_tokens_seen": 2951952,
"step": 446
},
{
"epoch": 2.8769106999195495,
"grad_norm": 1.5203238725662231,
"learning_rate": 1.9451276516876856e-06,
"loss": 0.0186,
"num_input_tokens_seen": 2958432,
"step": 447
},
{
"epoch": 2.8833467417538214,
"grad_norm": 1.398633599281311,
"learning_rate": 1.9351217836456316e-06,
"loss": 0.0071,
"num_input_tokens_seen": 2965328,
"step": 448
},
{
"epoch": 2.8897827835880934,
"grad_norm": 1.4775344133377075,
"learning_rate": 1.9251254420535665e-06,
"loss": 0.0177,
"num_input_tokens_seen": 2971872,
"step": 449
},
{
"epoch": 2.8962188254223653,
"grad_norm": 3.7046666145324707,
"learning_rate": 1.9151387954958792e-06,
"loss": 0.044,
"num_input_tokens_seen": 2978784,
"step": 450
},
{
"epoch": 2.9026548672566372,
"grad_norm": 1.9969475269317627,
"learning_rate": 1.9051620123934538e-06,
"loss": 0.0119,
"num_input_tokens_seen": 2985760,
"step": 451
},
{
"epoch": 2.909090909090909,
"grad_norm": 1.3861935138702393,
"learning_rate": 1.895195261000831e-06,
"loss": 0.0121,
"num_input_tokens_seen": 2992352,
"step": 452
},
{
"epoch": 2.915526950925181,
"grad_norm": 2.0632236003875732,
"learning_rate": 1.885238709403372e-06,
"loss": 0.0319,
"num_input_tokens_seen": 2998800,
"step": 453
},
{
"epoch": 2.921962992759453,
"grad_norm": 0.31324344873428345,
"learning_rate": 1.8752925255144228e-06,
"loss": 0.0024,
"num_input_tokens_seen": 3005392,
"step": 454
},
{
"epoch": 2.928399034593725,
"grad_norm": 1.0096696615219116,
"learning_rate": 1.8653568770724805e-06,
"loss": 0.0102,
"num_input_tokens_seen": 3012016,
"step": 455
},
{
"epoch": 2.934835076427997,
"grad_norm": 4.725823879241943,
"learning_rate": 1.8554319316383657e-06,
"loss": 0.0419,
"num_input_tokens_seen": 3018768,
"step": 456
},
{
"epoch": 2.941271118262269,
"grad_norm": 1.6467297077178955,
"learning_rate": 1.8455178565923993e-06,
"loss": 0.0109,
"num_input_tokens_seen": 3025328,
"step": 457
},
{
"epoch": 2.9477071600965408,
"grad_norm": 1.3065979480743408,
"learning_rate": 1.8356148191315753e-06,
"loss": 0.0092,
"num_input_tokens_seen": 3032080,
"step": 458
},
{
"epoch": 2.9541432019308127,
"grad_norm": 2.6485443115234375,
"learning_rate": 1.8257229862667437e-06,
"loss": 0.0449,
"num_input_tokens_seen": 3038880,
"step": 459
},
{
"epoch": 2.9605792437650846,
"grad_norm": 0.9736925363540649,
"learning_rate": 1.8158425248197931e-06,
"loss": 0.014,
"num_input_tokens_seen": 3045552,
"step": 460
},
{
"epoch": 2.9670152855993566,
"grad_norm": 0.423833429813385,
"learning_rate": 1.8059736014208388e-06,
"loss": 0.0035,
"num_input_tokens_seen": 3052288,
"step": 461
},
{
"epoch": 2.9734513274336285,
"grad_norm": 3.7729272842407227,
"learning_rate": 1.7961163825054101e-06,
"loss": 0.016,
"num_input_tokens_seen": 3058768,
"step": 462
},
{
"epoch": 2.9798873692679004,
"grad_norm": 2.9312222003936768,
"learning_rate": 1.7862710343116451e-06,
"loss": 0.0151,
"num_input_tokens_seen": 3065584,
"step": 463
},
{
"epoch": 2.9863234111021724,
"grad_norm": 0.6318484544754028,
"learning_rate": 1.7764377228774877e-06,
"loss": 0.0039,
"num_input_tokens_seen": 3072368,
"step": 464
},
{
"epoch": 2.9927594529364443,
"grad_norm": 5.504857063293457,
"learning_rate": 1.7666166140378853e-06,
"loss": 0.0361,
"num_input_tokens_seen": 3078864,
"step": 465
},
{
"epoch": 2.9991954947707162,
"grad_norm": 2.98315167427063,
"learning_rate": 1.7568078734219934e-06,
"loss": 0.0609,
"num_input_tokens_seen": 3085664,
"step": 466
},
{
"epoch": 3.0056315366049877,
"grad_norm": 0.24189546704292297,
"learning_rate": 1.747011666450384e-06,
"loss": 0.0027,
"num_input_tokens_seen": 3091568,
"step": 467
},
{
"epoch": 3.0120675784392597,
"grad_norm": 3.122098922729492,
"learning_rate": 1.737228158332252e-06,
"loss": 0.0097,
"num_input_tokens_seen": 3098544,
"step": 468
},
{
"epoch": 3.0185036202735316,
"grad_norm": 2.117048740386963,
"learning_rate": 1.7274575140626318e-06,
"loss": 0.0091,
"num_input_tokens_seen": 3105120,
"step": 469
},
{
"epoch": 3.0249396621078035,
"grad_norm": 0.3818783760070801,
"learning_rate": 1.7176998984196148e-06,
"loss": 0.0026,
"num_input_tokens_seen": 3111552,
"step": 470
},
{
"epoch": 3.0313757039420755,
"grad_norm": 3.4925177097320557,
"learning_rate": 1.7079554759615685e-06,
"loss": 0.0311,
"num_input_tokens_seen": 3118192,
"step": 471
},
{
"epoch": 3.0378117457763474,
"grad_norm": 0.1732572317123413,
"learning_rate": 1.6982244110243626e-06,
"loss": 0.0014,
"num_input_tokens_seen": 3124640,
"step": 472
},
{
"epoch": 3.0442477876106193,
"grad_norm": 1.305844783782959,
"learning_rate": 1.6885068677185989e-06,
"loss": 0.0185,
"num_input_tokens_seen": 3130992,
"step": 473
},
{
"epoch": 3.0506838294448912,
"grad_norm": 0.9071294665336609,
"learning_rate": 1.678803009926841e-06,
"loss": 0.0075,
"num_input_tokens_seen": 3137696,
"step": 474
},
{
"epoch": 3.057119871279163,
"grad_norm": 0.9389513731002808,
"learning_rate": 1.6691130013008514e-06,
"loss": 0.0069,
"num_input_tokens_seen": 3144560,
"step": 475
},
{
"epoch": 3.063555913113435,
"grad_norm": 0.15343110263347626,
"learning_rate": 1.6594370052588328e-06,
"loss": 0.0009,
"num_input_tokens_seen": 3151072,
"step": 476
},
{
"epoch": 3.069991954947707,
"grad_norm": 0.5078912973403931,
"learning_rate": 1.6497751849826692e-06,
"loss": 0.0015,
"num_input_tokens_seen": 3158016,
"step": 477
},
{
"epoch": 3.076427996781979,
"grad_norm": 0.14821191132068634,
"learning_rate": 1.6401277034151798e-06,
"loss": 0.0007,
"num_input_tokens_seen": 3164560,
"step": 478
},
{
"epoch": 3.082864038616251,
"grad_norm": 0.3397853672504425,
"learning_rate": 1.630494723257363e-06,
"loss": 0.0012,
"num_input_tokens_seen": 3171088,
"step": 479
},
{
"epoch": 3.089300080450523,
"grad_norm": 0.25013279914855957,
"learning_rate": 1.620876406965658e-06,
"loss": 0.0018,
"num_input_tokens_seen": 3177952,
"step": 480
},
{
"epoch": 3.0957361222847948,
"grad_norm": 0.04799158126115799,
"learning_rate": 1.611272916749205e-06,
"loss": 0.0003,
"num_input_tokens_seen": 3184592,
"step": 481
},
{
"epoch": 3.1021721641190667,
"grad_norm": 2.0195066928863525,
"learning_rate": 1.6016844145671062e-06,
"loss": 0.0044,
"num_input_tokens_seen": 3190896,
"step": 482
},
{
"epoch": 3.1086082059533386,
"grad_norm": 0.6244819164276123,
"learning_rate": 1.5921110621256972e-06,
"loss": 0.0019,
"num_input_tokens_seen": 3197376,
"step": 483
},
{
"epoch": 3.1150442477876106,
"grad_norm": 2.540050506591797,
"learning_rate": 1.58255302087582e-06,
"loss": 0.0059,
"num_input_tokens_seen": 3203776,
"step": 484
},
{
"epoch": 3.1214802896218825,
"grad_norm": 0.7487736344337463,
"learning_rate": 1.5730104520100984e-06,
"loss": 0.0036,
"num_input_tokens_seen": 3210464,
"step": 485
},
{
"epoch": 3.1279163314561544,
"grad_norm": 0.052535440772771835,
"learning_rate": 1.56348351646022e-06,
"loss": 0.0002,
"num_input_tokens_seen": 3217056,
"step": 486
},
{
"epoch": 3.1343523732904264,
"grad_norm": 2.5393643379211426,
"learning_rate": 1.5539723748942246e-06,
"loss": 0.0019,
"num_input_tokens_seen": 3223840,
"step": 487
},
{
"epoch": 3.1407884151246983,
"grad_norm": 0.28790536522865295,
"learning_rate": 1.544477187713791e-06,
"loss": 0.0009,
"num_input_tokens_seen": 3230592,
"step": 488
},
{
"epoch": 3.1472244569589702,
"grad_norm": 2.5697410106658936,
"learning_rate": 1.534998115051533e-06,
"loss": 0.0318,
"num_input_tokens_seen": 3237216,
"step": 489
},
{
"epoch": 3.153660498793242,
"grad_norm": 1.5203006267547607,
"learning_rate": 1.5255353167683017e-06,
"loss": 0.0216,
"num_input_tokens_seen": 3243920,
"step": 490
},
{
"epoch": 3.160096540627514,
"grad_norm": 0.1484091877937317,
"learning_rate": 1.5160889524504857e-06,
"loss": 0.0004,
"num_input_tokens_seen": 3250656,
"step": 491
},
{
"epoch": 3.166532582461786,
"grad_norm": 3.3526744842529297,
"learning_rate": 1.50665918140732e-06,
"loss": 0.0286,
"num_input_tokens_seen": 3257312,
"step": 492
},
{
"epoch": 3.172968624296058,
"grad_norm": 1.3879235982894897,
"learning_rate": 1.4972461626682033e-06,
"loss": 0.0254,
"num_input_tokens_seen": 3264112,
"step": 493
},
{
"epoch": 3.17940466613033,
"grad_norm": 1.1939952373504639,
"learning_rate": 1.4878500549800115e-06,
"loss": 0.0039,
"num_input_tokens_seen": 3270528,
"step": 494
},
{
"epoch": 3.185840707964602,
"grad_norm": 0.20248474180698395,
"learning_rate": 1.4784710168044215e-06,
"loss": 0.0005,
"num_input_tokens_seen": 3277008,
"step": 495
},
{
"epoch": 3.1922767497988738,
"grad_norm": 1.903956413269043,
"learning_rate": 1.4691092063152417e-06,
"loss": 0.0196,
"num_input_tokens_seen": 3283376,
"step": 496
},
{
"epoch": 3.1987127916331457,
"grad_norm": 0.3746008276939392,
"learning_rate": 1.459764781395741e-06,
"loss": 0.0015,
"num_input_tokens_seen": 3289664,
"step": 497
},
{
"epoch": 3.2051488334674176,
"grad_norm": 4.635190486907959,
"learning_rate": 1.4504378996359867e-06,
"loss": 0.0088,
"num_input_tokens_seen": 3296576,
"step": 498
},
{
"epoch": 3.2115848753016896,
"grad_norm": 1.4451507329940796,
"learning_rate": 1.4411287183301902e-06,
"loss": 0.0023,
"num_input_tokens_seen": 3303120,
"step": 499
},
{
"epoch": 3.2180209171359615,
"grad_norm": 1.252470850944519,
"learning_rate": 1.4318373944740485e-06,
"loss": 0.0071,
"num_input_tokens_seen": 3310384,
"step": 500
},
{
"epoch": 3.2244569589702334,
"grad_norm": 0.6509237289428711,
"learning_rate": 1.4225640847621006e-06,
"loss": 0.0006,
"num_input_tokens_seen": 3316768,
"step": 501
},
{
"epoch": 3.2308930008045054,
"grad_norm": 0.2248382717370987,
"learning_rate": 1.4133089455850878e-06,
"loss": 0.0011,
"num_input_tokens_seen": 3323488,
"step": 502
},
{
"epoch": 3.2373290426387773,
"grad_norm": 1.0306220054626465,
"learning_rate": 1.4040721330273063e-06,
"loss": 0.0057,
"num_input_tokens_seen": 3330000,
"step": 503
},
{
"epoch": 3.2437650844730492,
"grad_norm": 0.1734343320131302,
"learning_rate": 1.3948538028639851e-06,
"loss": 0.0006,
"num_input_tokens_seen": 3336592,
"step": 504
},
{
"epoch": 3.250201126307321,
"grad_norm": 0.9872696399688721,
"learning_rate": 1.3856541105586545e-06,
"loss": 0.0066,
"num_input_tokens_seen": 3343136,
"step": 505
},
{
"epoch": 3.256637168141593,
"grad_norm": 0.8048367500305176,
"learning_rate": 1.3764732112605223e-06,
"loss": 0.0079,
"num_input_tokens_seen": 3349680,
"step": 506
},
{
"epoch": 3.263073209975865,
"grad_norm": 1.8275296688079834,
"learning_rate": 1.367311259801863e-06,
"loss": 0.0215,
"num_input_tokens_seen": 3356304,
"step": 507
},
{
"epoch": 3.2695092518101365,
"grad_norm": 1.90727698802948,
"learning_rate": 1.3581684106953987e-06,
"loss": 0.0031,
"num_input_tokens_seen": 3363008,
"step": 508
},
{
"epoch": 3.2759452936444085,
"grad_norm": 2.614037275314331,
"learning_rate": 1.3490448181317025e-06,
"loss": 0.0024,
"num_input_tokens_seen": 3369728,
"step": 509
},
{
"epoch": 3.2823813354786804,
"grad_norm": 1.9239071607589722,
"learning_rate": 1.3399406359765921e-06,
"loss": 0.0094,
"num_input_tokens_seen": 3375968,
"step": 510
},
{
"epoch": 3.2888173773129523,
"grad_norm": 1.1601731777191162,
"learning_rate": 1.3308560177685334e-06,
"loss": 0.0054,
"num_input_tokens_seen": 3383024,
"step": 511
},
{
"epoch": 3.2952534191472242,
"grad_norm": 0.31424281001091003,
"learning_rate": 1.3217911167160575e-06,
"loss": 0.0008,
"num_input_tokens_seen": 3389488,
"step": 512
},
{
"epoch": 3.301689460981496,
"grad_norm": 2.633910655975342,
"learning_rate": 1.3127460856951724e-06,
"loss": 0.0053,
"num_input_tokens_seen": 3395712,
"step": 513
},
{
"epoch": 3.308125502815768,
"grad_norm": 0.9618326425552368,
"learning_rate": 1.303721077246784e-06,
"loss": 0.006,
"num_input_tokens_seen": 3402384,
"step": 514
},
{
"epoch": 3.31456154465004,
"grad_norm": 0.22136647999286652,
"learning_rate": 1.2947162435741278e-06,
"loss": 0.0004,
"num_input_tokens_seen": 3409136,
"step": 515
},
{
"epoch": 3.320997586484312,
"grad_norm": 1.880077838897705,
"learning_rate": 1.2857317365401997e-06,
"loss": 0.0135,
"num_input_tokens_seen": 3415776,
"step": 516
},
{
"epoch": 3.327433628318584,
"grad_norm": 2.234178304672241,
"learning_rate": 1.2767677076651913e-06,
"loss": 0.0083,
"num_input_tokens_seen": 3422496,
"step": 517
},
{
"epoch": 3.333869670152856,
"grad_norm": 0.21132518351078033,
"learning_rate": 1.2678243081239421e-06,
"loss": 0.0004,
"num_input_tokens_seen": 3429312,
"step": 518
},
{
"epoch": 3.340305711987128,
"grad_norm": 1.0334022045135498,
"learning_rate": 1.2589016887433846e-06,
"loss": 0.0038,
"num_input_tokens_seen": 3435840,
"step": 519
},
{
"epoch": 3.3467417538213997,
"grad_norm": 2.751037359237671,
"learning_rate": 1.2500000000000007e-06,
"loss": 0.0269,
"num_input_tokens_seen": 3442176,
"step": 520
},
{
"epoch": 3.3531777956556716,
"grad_norm": 0.7970973253250122,
"learning_rate": 1.2411193920172866e-06,
"loss": 0.0062,
"num_input_tokens_seen": 3448784,
"step": 521
},
{
"epoch": 3.3596138374899436,
"grad_norm": 0.09952107071876526,
"learning_rate": 1.2322600145632204e-06,
"loss": 0.0004,
"num_input_tokens_seen": 3455184,
"step": 522
},
{
"epoch": 3.3660498793242155,
"grad_norm": 0.6218022108078003,
"learning_rate": 1.2234220170477332e-06,
"loss": 0.0011,
"num_input_tokens_seen": 3461792,
"step": 523
},
{
"epoch": 3.3724859211584874,
"grad_norm": 1.4417766332626343,
"learning_rate": 1.2146055485201943e-06,
"loss": 0.0026,
"num_input_tokens_seen": 3468624,
"step": 524
},
{
"epoch": 3.3789219629927594,
"grad_norm": 2.819247245788574,
"learning_rate": 1.205810757666894e-06,
"loss": 0.0183,
"num_input_tokens_seen": 3474976,
"step": 525
},
{
"epoch": 3.3853580048270313,
"grad_norm": 1.7066518068313599,
"learning_rate": 1.1970377928085372e-06,
"loss": 0.0079,
"num_input_tokens_seen": 3481360,
"step": 526
},
{
"epoch": 3.3917940466613032,
"grad_norm": 2.671914577484131,
"learning_rate": 1.188286801897743e-06,
"loss": 0.0123,
"num_input_tokens_seen": 3487904,
"step": 527
},
{
"epoch": 3.398230088495575,
"grad_norm": 0.37451621890068054,
"learning_rate": 1.1795579325165448e-06,
"loss": 0.0018,
"num_input_tokens_seen": 3494368,
"step": 528
},
{
"epoch": 3.404666130329847,
"grad_norm": 0.3565497398376465,
"learning_rate": 1.1708513318739096e-06,
"loss": 0.0014,
"num_input_tokens_seen": 3500704,
"step": 529
},
{
"epoch": 3.411102172164119,
"grad_norm": 0.22408631443977356,
"learning_rate": 1.1621671468032495e-06,
"loss": 0.0009,
"num_input_tokens_seen": 3507216,
"step": 530
},
{
"epoch": 3.417538213998391,
"grad_norm": 1.3339484930038452,
"learning_rate": 1.153505523759944e-06,
"loss": 0.0078,
"num_input_tokens_seen": 3513664,
"step": 531
},
{
"epoch": 3.423974255832663,
"grad_norm": 0.6494855880737305,
"learning_rate": 1.1448666088188766e-06,
"loss": 0.0027,
"num_input_tokens_seen": 3520096,
"step": 532
},
{
"epoch": 3.430410297666935,
"grad_norm": 0.10890411585569382,
"learning_rate": 1.1362505476719662e-06,
"loss": 0.0004,
"num_input_tokens_seen": 3526560,
"step": 533
},
{
"epoch": 3.4368463395012068,
"grad_norm": 13.174049377441406,
"learning_rate": 1.1276574856257097e-06,
"loss": 0.0064,
"num_input_tokens_seen": 3533536,
"step": 534
},
{
"epoch": 3.4432823813354787,
"grad_norm": 2.0001068115234375,
"learning_rate": 1.1190875675987355e-06,
"loss": 0.007,
"num_input_tokens_seen": 3540288,
"step": 535
},
{
"epoch": 3.4497184231697506,
"grad_norm": 0.15650025010108948,
"learning_rate": 1.1105409381193572e-06,
"loss": 0.0005,
"num_input_tokens_seen": 3546720,
"step": 536
},
{
"epoch": 3.4561544650040226,
"grad_norm": 0.13460475206375122,
"learning_rate": 1.1020177413231334e-06,
"loss": 0.0004,
"num_input_tokens_seen": 3553280,
"step": 537
},
{
"epoch": 3.4625905068382945,
"grad_norm": 2.165956735610962,
"learning_rate": 1.0935181209504422e-06,
"loss": 0.0294,
"num_input_tokens_seen": 3559776,
"step": 538
},
{
"epoch": 3.4690265486725664,
"grad_norm": 0.6856318712234497,
"learning_rate": 1.0850422203440555e-06,
"loss": 0.0036,
"num_input_tokens_seen": 3566848,
"step": 539
},
{
"epoch": 3.4754625905068384,
"grad_norm": 1.306766152381897,
"learning_rate": 1.0765901824467167e-06,
"loss": 0.0051,
"num_input_tokens_seen": 3573280,
"step": 540
},
{
"epoch": 3.4818986323411103,
"grad_norm": 0.3889179825782776,
"learning_rate": 1.068162149798737e-06,
"loss": 0.0012,
"num_input_tokens_seen": 3579712,
"step": 541
},
{
"epoch": 3.4883346741753822,
"grad_norm": 1.5245965719223022,
"learning_rate": 1.0597582645355891e-06,
"loss": 0.0231,
"num_input_tokens_seen": 3586480,
"step": 542
},
{
"epoch": 3.494770716009654,
"grad_norm": 0.6708037257194519,
"learning_rate": 1.0513786683855062e-06,
"loss": 0.0041,
"num_input_tokens_seen": 3593136,
"step": 543
},
{
"epoch": 3.501206757843926,
"grad_norm": 2.0138630867004395,
"learning_rate": 1.0430235026670979e-06,
"loss": 0.0124,
"num_input_tokens_seen": 3599968,
"step": 544
},
{
"epoch": 3.507642799678198,
"grad_norm": 7.274059295654297,
"learning_rate": 1.034692908286964e-06,
"loss": 0.0171,
"num_input_tokens_seen": 3606592,
"step": 545
},
{
"epoch": 3.51407884151247,
"grad_norm": 5.609940052032471,
"learning_rate": 1.0263870257373162e-06,
"loss": 0.008,
"num_input_tokens_seen": 3613072,
"step": 546
},
{
"epoch": 3.520514883346742,
"grad_norm": 1.4191588163375854,
"learning_rate": 1.0181059950936131e-06,
"loss": 0.0035,
"num_input_tokens_seen": 3619696,
"step": 547
},
{
"epoch": 3.526950925181014,
"grad_norm": 0.1580982804298401,
"learning_rate": 1.0098499560121943e-06,
"loss": 0.0006,
"num_input_tokens_seen": 3626240,
"step": 548
},
{
"epoch": 3.5333869670152858,
"grad_norm": 0.637765109539032,
"learning_rate": 1.0016190477279274e-06,
"loss": 0.002,
"num_input_tokens_seen": 3632704,
"step": 549
},
{
"epoch": 3.5398230088495577,
"grad_norm": 0.07971790432929993,
"learning_rate": 9.934134090518593e-07,
"loss": 0.0003,
"num_input_tokens_seen": 3639360,
"step": 550
},
{
"epoch": 3.5462590506838296,
"grad_norm": 0.15312433242797852,
"learning_rate": 9.852331783688722e-07,
"loss": 0.0004,
"num_input_tokens_seen": 3646112,
"step": 551
},
{
"epoch": 3.5526950925181016,
"grad_norm": 0.5353730916976929,
"learning_rate": 9.770784936353555e-07,
"loss": 0.0016,
"num_input_tokens_seen": 3652704,
"step": 552
},
{
"epoch": 3.5591311343523735,
"grad_norm": 0.3197666108608246,
"learning_rate": 9.689494923768756e-07,
"loss": 0.0012,
"num_input_tokens_seen": 3659696,
"step": 553
},
{
"epoch": 3.5655671761866454,
"grad_norm": 1.4529962539672852,
"learning_rate": 9.608463116858544e-07,
"loss": 0.0057,
"num_input_tokens_seen": 3666288,
"step": 554
},
{
"epoch": 3.5720032180209174,
"grad_norm": 2.7501587867736816,
"learning_rate": 9.527690882192636e-07,
"loss": 0.0168,
"num_input_tokens_seen": 3673104,
"step": 555
},
{
"epoch": 3.5784392598551893,
"grad_norm": 0.21036742627620697,
"learning_rate": 9.447179581963156e-07,
"loss": 0.0012,
"num_input_tokens_seen": 3679872,
"step": 556
},
{
"epoch": 3.5848753016894612,
"grad_norm": 0.03335335850715637,
"learning_rate": 9.366930573961649e-07,
"loss": 0.0002,
"num_input_tokens_seen": 3686288,
"step": 557
},
{
"epoch": 3.591311343523733,
"grad_norm": 1.3189131021499634,
"learning_rate": 9.286945211556231e-07,
"loss": 0.0025,
"num_input_tokens_seen": 3692976,
"step": 558
},
{
"epoch": 3.597747385358005,
"grad_norm": 1.1787400245666504,
"learning_rate": 9.207224843668733e-07,
"loss": 0.0194,
"num_input_tokens_seen": 3699312,
"step": 559
},
{
"epoch": 3.604183427192277,
"grad_norm": 0.9992094039916992,
"learning_rate": 9.127770814751933e-07,
"loss": 0.0055,
"num_input_tokens_seen": 3705888,
"step": 560
},
{
"epoch": 3.6106194690265485,
"grad_norm": 2.264843702316284,
"learning_rate": 9.048584464766938e-07,
"loss": 0.0215,
"num_input_tokens_seen": 3712688,
"step": 561
},
{
"epoch": 3.6170555108608204,
"grad_norm": 0.06527237594127655,
"learning_rate": 8.969667129160547e-07,
"loss": 0.0003,
"num_input_tokens_seen": 3719168,
"step": 562
},
{
"epoch": 3.6234915526950924,
"grad_norm": 3.79392409324646,
"learning_rate": 8.891020138842718e-07,
"loss": 0.0242,
"num_input_tokens_seen": 3726048,
"step": 563
},
{
"epoch": 3.6299275945293643,
"grad_norm": 0.9232211112976074,
"learning_rate": 8.81264482016416e-07,
"loss": 0.0206,
"num_input_tokens_seen": 3732672,
"step": 564
},
{
"epoch": 3.6363636363636362,
"grad_norm": 0.5276843309402466,
"learning_rate": 8.734542494893955e-07,
"loss": 0.0024,
"num_input_tokens_seen": 3739456,
"step": 565
},
{
"epoch": 3.642799678197908,
"grad_norm": 1.1676807403564453,
"learning_rate": 8.65671448019722e-07,
"loss": 0.0087,
"num_input_tokens_seen": 3746160,
"step": 566
},
{
"epoch": 3.64923572003218,
"grad_norm": 1.3703765869140625,
"learning_rate": 8.579162088612974e-07,
"loss": 0.0089,
"num_input_tokens_seen": 3752560,
"step": 567
},
{
"epoch": 3.655671761866452,
"grad_norm": 0.06538532674312592,
"learning_rate": 8.501886628031941e-07,
"loss": 0.0003,
"num_input_tokens_seen": 3759600,
"step": 568
},
{
"epoch": 3.662107803700724,
"grad_norm": 0.0386020764708519,
"learning_rate": 8.424889401674505e-07,
"loss": 0.0002,
"num_input_tokens_seen": 3766096,
"step": 569
},
{
"epoch": 3.668543845534996,
"grad_norm": 0.20554865896701813,
"learning_rate": 8.348171708068748e-07,
"loss": 0.0009,
"num_input_tokens_seen": 3772944,
"step": 570
},
{
"epoch": 3.674979887369268,
"grad_norm": 0.9973205327987671,
"learning_rate": 8.271734841028553e-07,
"loss": 0.0154,
"num_input_tokens_seen": 3779664,
"step": 571
},
{
"epoch": 3.6814159292035398,
"grad_norm": 0.30160781741142273,
"learning_rate": 8.195580089631733e-07,
"loss": 0.0012,
"num_input_tokens_seen": 3786080,
"step": 572
},
{
"epoch": 3.6878519710378117,
"grad_norm": 0.49049124121665955,
"learning_rate": 8.119708738198395e-07,
"loss": 0.0008,
"num_input_tokens_seen": 3792768,
"step": 573
},
{
"epoch": 3.6942880128720836,
"grad_norm": 1.6590077877044678,
"learning_rate": 8.04412206626915e-07,
"loss": 0.0081,
"num_input_tokens_seen": 3799472,
"step": 574
},
{
"epoch": 3.7007240547063556,
"grad_norm": 1.814943552017212,
"learning_rate": 7.968821348583644e-07,
"loss": 0.008,
"num_input_tokens_seen": 3805984,
"step": 575
},
{
"epoch": 3.7071600965406275,
"grad_norm": 1.6639471054077148,
"learning_rate": 7.89380785505901e-07,
"loss": 0.0073,
"num_input_tokens_seen": 3813088,
"step": 576
},
{
"epoch": 3.7135961383748994,
"grad_norm": 0.946050763130188,
"learning_rate": 7.819082850768433e-07,
"loss": 0.0062,
"num_input_tokens_seen": 3820208,
"step": 577
},
{
"epoch": 3.7200321802091714,
"grad_norm": 0.2189425230026245,
"learning_rate": 7.744647595919869e-07,
"loss": 0.0015,
"num_input_tokens_seen": 3826800,
"step": 578
},
{
"epoch": 3.7264682220434433,
"grad_norm": 2.796231985092163,
"learning_rate": 7.670503345834757e-07,
"loss": 0.0268,
"num_input_tokens_seen": 3833344,
"step": 579
},
{
"epoch": 3.7329042638777152,
"grad_norm": 0.13711552321910858,
"learning_rate": 7.596651350926837e-07,
"loss": 0.0007,
"num_input_tokens_seen": 3839920,
"step": 580
},
{
"epoch": 3.739340305711987,
"grad_norm": 0.3616367280483246,
"learning_rate": 7.523092856681099e-07,
"loss": 0.0016,
"num_input_tokens_seen": 3846432,
"step": 581
},
{
"epoch": 3.745776347546259,
"grad_norm": 2.3357245922088623,
"learning_rate": 7.44982910363276e-07,
"loss": 0.0631,
"num_input_tokens_seen": 3853216,
"step": 582
},
{
"epoch": 3.752212389380531,
"grad_norm": 1.7805283069610596,
"learning_rate": 7.376861327346325e-07,
"loss": 0.0128,
"num_input_tokens_seen": 3859664,
"step": 583
},
{
"epoch": 3.758648431214803,
"grad_norm": 0.5933414101600647,
"learning_rate": 7.304190758394775e-07,
"loss": 0.0034,
"num_input_tokens_seen": 3866208,
"step": 584
},
{
"epoch": 3.765084473049075,
"grad_norm": 3.1310431957244873,
"learning_rate": 7.231818622338824e-07,
"loss": 0.0633,
"num_input_tokens_seen": 3872736,
"step": 585
},
{
"epoch": 3.771520514883347,
"grad_norm": 0.2022082656621933,
"learning_rate": 7.159746139706194e-07,
"loss": 0.0007,
"num_input_tokens_seen": 3879264,
"step": 586
},
{
"epoch": 3.7779565567176188,
"grad_norm": 6.451120376586914,
"learning_rate": 7.087974525971103e-07,
"loss": 0.0211,
"num_input_tokens_seen": 3885744,
"step": 587
},
{
"epoch": 3.7843925985518907,
"grad_norm": 0.8931072354316711,
"learning_rate": 7.016504991533727e-07,
"loss": 0.009,
"num_input_tokens_seen": 3892304,
"step": 588
},
{
"epoch": 3.7908286403861626,
"grad_norm": 1.4347479343414307,
"learning_rate": 6.94533874169977e-07,
"loss": 0.0152,
"num_input_tokens_seen": 3898768,
"step": 589
},
{
"epoch": 3.7972646822204346,
"grad_norm": 0.5323463678359985,
"learning_rate": 6.874476976660185e-07,
"loss": 0.0022,
"num_input_tokens_seen": 3904976,
"step": 590
},
{
"epoch": 3.8037007240547065,
"grad_norm": 1.120011806488037,
"learning_rate": 6.803920891470905e-07,
"loss": 0.014,
"num_input_tokens_seen": 3911360,
"step": 591
},
{
"epoch": 3.8101367658889784,
"grad_norm": 0.6292040348052979,
"learning_rate": 6.733671676032674e-07,
"loss": 0.0085,
"num_input_tokens_seen": 3918224,
"step": 592
},
{
"epoch": 3.8165728077232504,
"grad_norm": 3.3647360801696777,
"learning_rate": 6.663730515071019e-07,
"loss": 0.0161,
"num_input_tokens_seen": 3924960,
"step": 593
},
{
"epoch": 3.823008849557522,
"grad_norm": 1.8465656042099,
"learning_rate": 6.594098588116243e-07,
"loss": 0.0234,
"num_input_tokens_seen": 3931712,
"step": 594
},
{
"epoch": 3.829444891391794,
"grad_norm": 1.0739251375198364,
"learning_rate": 6.524777069483526e-07,
"loss": 0.0186,
"num_input_tokens_seen": 3938304,
"step": 595
},
{
"epoch": 3.8358809332260657,
"grad_norm": 3.146777629852295,
"learning_rate": 6.455767128253148e-07,
"loss": 0.0199,
"num_input_tokens_seen": 3945200,
"step": 596
},
{
"epoch": 3.8423169750603376,
"grad_norm": 1.1694271564483643,
"learning_rate": 6.38706992825075e-07,
"loss": 0.0052,
"num_input_tokens_seen": 3951808,
"step": 597
},
{
"epoch": 3.8487530168946096,
"grad_norm": 1.1547743082046509,
"learning_rate": 6.318686628027723e-07,
"loss": 0.0165,
"num_input_tokens_seen": 3958480,
"step": 598
},
{
"epoch": 3.8551890587288815,
"grad_norm": 1.1595410108566284,
"learning_rate": 6.250618380841661e-07,
"loss": 0.01,
"num_input_tokens_seen": 3965072,
"step": 599
},
{
"epoch": 3.8616251005631534,
"grad_norm": 0.8452915549278259,
"learning_rate": 6.182866334636889e-07,
"loss": 0.0047,
"num_input_tokens_seen": 3971808,
"step": 600
},
{
"epoch": 3.8680611423974254,
"grad_norm": 2.201892375946045,
"learning_rate": 6.115431632025154e-07,
"loss": 0.0039,
"num_input_tokens_seen": 3978480,
"step": 601
},
{
"epoch": 3.8744971842316973,
"grad_norm": 0.24013373255729675,
"learning_rate": 6.048315410266326e-07,
"loss": 0.0022,
"num_input_tokens_seen": 3985216,
"step": 602
},
{
"epoch": 3.8809332260659692,
"grad_norm": 0.442757248878479,
"learning_rate": 5.981518801249192e-07,
"loss": 0.0042,
"num_input_tokens_seen": 3991792,
"step": 603
},
{
"epoch": 3.887369267900241,
"grad_norm": 2.5312795639038086,
"learning_rate": 5.915042931472426e-07,
"loss": 0.0076,
"num_input_tokens_seen": 3998224,
"step": 604
},
{
"epoch": 3.893805309734513,
"grad_norm": 0.3599741756916046,
"learning_rate": 5.848888922025553e-07,
"loss": 0.0019,
"num_input_tokens_seen": 4004960,
"step": 605
},
{
"epoch": 3.900241351568785,
"grad_norm": 0.33045250177383423,
"learning_rate": 5.783057888570034e-07,
"loss": 0.0014,
"num_input_tokens_seen": 4011984,
"step": 606
},
{
"epoch": 3.906677393403057,
"grad_norm": 0.540598452091217,
"learning_rate": 5.717550941320482e-07,
"loss": 0.0022,
"num_input_tokens_seen": 4018912,
"step": 607
},
{
"epoch": 3.913113435237329,
"grad_norm": 0.4901201128959656,
"learning_rate": 5.65236918502593e-07,
"loss": 0.0024,
"num_input_tokens_seen": 4025504,
"step": 608
},
{
"epoch": 3.919549477071601,
"grad_norm": 0.23451536893844604,
"learning_rate": 5.587513718951165e-07,
"loss": 0.0013,
"num_input_tokens_seen": 4031776,
"step": 609
},
{
"epoch": 3.9259855189058728,
"grad_norm": 0.9038437604904175,
"learning_rate": 5.522985636858238e-07,
"loss": 0.0064,
"num_input_tokens_seen": 4038208,
"step": 610
},
{
"epoch": 3.9324215607401447,
"grad_norm": 1.4877148866653442,
"learning_rate": 5.458786026988005e-07,
"loss": 0.0084,
"num_input_tokens_seen": 4044928,
"step": 611
},
{
"epoch": 3.9388576025744166,
"grad_norm": 0.12848466634750366,
"learning_rate": 5.394915972041739e-07,
"loss": 0.0009,
"num_input_tokens_seen": 4051552,
"step": 612
},
{
"epoch": 3.9452936444086886,
"grad_norm": 0.22914128005504608,
"learning_rate": 5.33137654916292e-07,
"loss": 0.001,
"num_input_tokens_seen": 4058304,
"step": 613
},
{
"epoch": 3.9517296862429605,
"grad_norm": 0.7593125700950623,
"learning_rate": 5.268168829919046e-07,
"loss": 0.0064,
"num_input_tokens_seen": 4064720,
"step": 614
},
{
"epoch": 3.9581657280772324,
"grad_norm": 0.6085631251335144,
"learning_rate": 5.205293880283552e-07,
"loss": 0.0033,
"num_input_tokens_seen": 4071216,
"step": 615
},
{
"epoch": 3.9646017699115044,
"grad_norm": 0.9351167678833008,
"learning_rate": 5.14275276061785e-07,
"loss": 0.0065,
"num_input_tokens_seen": 4077904,
"step": 616
},
{
"epoch": 3.9710378117457763,
"grad_norm": 2.1718461513519287,
"learning_rate": 5.080546525653448e-07,
"loss": 0.0272,
"num_input_tokens_seen": 4084656,
"step": 617
},
{
"epoch": 3.9774738535800482,
"grad_norm": 0.39000532031059265,
"learning_rate": 5.018676224474139e-07,
"loss": 0.0015,
"num_input_tokens_seen": 4091584,
"step": 618
},
{
"epoch": 3.98390989541432,
"grad_norm": 0.5723803639411926,
"learning_rate": 4.957142900498335e-07,
"loss": 0.0013,
"num_input_tokens_seen": 4098768,
"step": 619
},
{
"epoch": 3.990345937248592,
"grad_norm": 1.5266039371490479,
"learning_rate": 4.895947591461456e-07,
"loss": 0.0148,
"num_input_tokens_seen": 4105312,
"step": 620
},
{
"epoch": 3.996781979082864,
"grad_norm": 0.7928001880645752,
"learning_rate": 4.835091329398436e-07,
"loss": 0.0063,
"num_input_tokens_seen": 4112000,
"step": 621
},
{
"epoch": 4.003218020917136,
"grad_norm": 0.09017051756381989,
"learning_rate": 4.774575140626317e-07,
"loss": 0.0003,
"num_input_tokens_seen": 4118624,
"step": 622
},
{
"epoch": 4.009654062751408,
"grad_norm": 0.2493676394224167,
"learning_rate": 4.714400045726919e-07,
"loss": 0.001,
"num_input_tokens_seen": 4125408,
"step": 623
},
{
"epoch": 4.01609010458568,
"grad_norm": 0.03381378576159477,
"learning_rate": 4.6545670595296686e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4131936,
"step": 624
},
{
"epoch": 4.022526146419952,
"grad_norm": 2.845327854156494,
"learning_rate": 4.5950771910944603e-07,
"loss": 0.0193,
"num_input_tokens_seen": 4138352,
"step": 625
},
{
"epoch": 4.028962188254224,
"grad_norm": 0.6973279714584351,
"learning_rate": 4.5359314436946275e-07,
"loss": 0.0049,
"num_input_tokens_seen": 4144672,
"step": 626
},
{
"epoch": 4.035398230088496,
"grad_norm": 0.3552819788455963,
"learning_rate": 4.4771308148000487e-07,
"loss": 0.0015,
"num_input_tokens_seen": 4151296,
"step": 627
},
{
"epoch": 4.041834271922768,
"grad_norm": 0.2976234555244446,
"learning_rate": 4.418676296060323e-07,
"loss": 0.0019,
"num_input_tokens_seen": 4157696,
"step": 628
},
{
"epoch": 4.0482703137570395,
"grad_norm": 0.6438854932785034,
"learning_rate": 4.3605688732880097e-07,
"loss": 0.0034,
"num_input_tokens_seen": 4164352,
"step": 629
},
{
"epoch": 4.054706355591311,
"grad_norm": 0.055070556700229645,
"learning_rate": 4.302809526442053e-07,
"loss": 0.0003,
"num_input_tokens_seen": 4170992,
"step": 630
},
{
"epoch": 4.061142397425583,
"grad_norm": 0.5393857359886169,
"learning_rate": 4.2453992296112384e-07,
"loss": 0.0031,
"num_input_tokens_seen": 4177888,
"step": 631
},
{
"epoch": 4.067578439259855,
"grad_norm": 0.10041255503892899,
"learning_rate": 4.188338950997728e-07,
"loss": 0.0004,
"num_input_tokens_seen": 4184800,
"step": 632
},
{
"epoch": 4.074014481094127,
"grad_norm": 0.4824787378311157,
"learning_rate": 4.1316296529007955e-07,
"loss": 0.0027,
"num_input_tokens_seen": 4191136,
"step": 633
},
{
"epoch": 4.080450522928399,
"grad_norm": 0.8842573761940002,
"learning_rate": 4.075272291700558e-07,
"loss": 0.0047,
"num_input_tokens_seen": 4197984,
"step": 634
},
{
"epoch": 4.086886564762671,
"grad_norm": 0.0672411248087883,
"learning_rate": 4.019267817841835e-07,
"loss": 0.0003,
"num_input_tokens_seen": 4204688,
"step": 635
},
{
"epoch": 4.093322606596943,
"grad_norm": 1.144921898841858,
"learning_rate": 3.9636171758181657e-07,
"loss": 0.0204,
"num_input_tokens_seen": 4211360,
"step": 636
},
{
"epoch": 4.099758648431215,
"grad_norm": 1.0628600120544434,
"learning_rate": 3.908321304155846e-07,
"loss": 0.0043,
"num_input_tokens_seen": 4218000,
"step": 637
},
{
"epoch": 4.106194690265487,
"grad_norm": 0.03438463807106018,
"learning_rate": 3.853381135398093e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4224544,
"step": 638
},
{
"epoch": 4.112630732099759,
"grad_norm": 0.22854630649089813,
"learning_rate": 3.798797596089351e-07,
"loss": 0.0009,
"num_input_tokens_seen": 4230992,
"step": 639
},
{
"epoch": 4.119066773934031,
"grad_norm": 0.12790539860725403,
"learning_rate": 3.7445716067596506e-07,
"loss": 0.0004,
"num_input_tokens_seen": 4237808,
"step": 640
},
{
"epoch": 4.125502815768303,
"grad_norm": 0.040783047676086426,
"learning_rate": 3.6907040819090604e-07,
"loss": 0.0003,
"num_input_tokens_seen": 4244032,
"step": 641
},
{
"epoch": 4.131938857602575,
"grad_norm": 0.29912275075912476,
"learning_rate": 3.63719592999231e-07,
"loss": 0.0015,
"num_input_tokens_seen": 4250640,
"step": 642
},
{
"epoch": 4.1383748994368466,
"grad_norm": 0.1869562268257141,
"learning_rate": 3.5840480534034355e-07,
"loss": 0.0012,
"num_input_tokens_seen": 4257440,
"step": 643
},
{
"epoch": 4.1448109412711185,
"grad_norm": 0.6986035108566284,
"learning_rate": 3.5312613484605546e-07,
"loss": 0.0041,
"num_input_tokens_seen": 4263936,
"step": 644
},
{
"epoch": 4.15124698310539,
"grad_norm": 1.672957181930542,
"learning_rate": 3.4788367053908087e-07,
"loss": 0.0087,
"num_input_tokens_seen": 4270144,
"step": 645
},
{
"epoch": 4.157683024939662,
"grad_norm": 0.5007069110870361,
"learning_rate": 3.4267750083152587e-07,
"loss": 0.0025,
"num_input_tokens_seen": 4276944,
"step": 646
},
{
"epoch": 4.164119066773934,
"grad_norm": 0.2200661450624466,
"learning_rate": 3.375077135234051e-07,
"loss": 0.0012,
"num_input_tokens_seen": 4283488,
"step": 647
},
{
"epoch": 4.170555108608206,
"grad_norm": 1.018943190574646,
"learning_rate": 3.323743958011588e-07,
"loss": 0.0111,
"num_input_tokens_seen": 4290000,
"step": 648
},
{
"epoch": 4.176991150442478,
"grad_norm": 0.5488151907920837,
"learning_rate": 3.2727763423617915e-07,
"loss": 0.0026,
"num_input_tokens_seen": 4296544,
"step": 649
},
{
"epoch": 4.18342719227675,
"grad_norm": 0.12664268910884857,
"learning_rate": 3.222175147833556e-07,
"loss": 0.0004,
"num_input_tokens_seen": 4303056,
"step": 650
},
{
"epoch": 4.189863234111022,
"grad_norm": 0.04309312626719475,
"learning_rate": 3.171941227796227e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4309664,
"step": 651
},
{
"epoch": 4.196299275945294,
"grad_norm": 2.9486300945281982,
"learning_rate": 3.122075429425184e-07,
"loss": 0.0118,
"num_input_tokens_seen": 4316112,
"step": 652
},
{
"epoch": 4.202735317779566,
"grad_norm": 0.03176088631153107,
"learning_rate": 3.072578593687606e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4322800,
"step": 653
},
{
"epoch": 4.209171359613838,
"grad_norm": 4.464654445648193,
"learning_rate": 3.0234515553282523e-07,
"loss": 0.0151,
"num_input_tokens_seen": 4329408,
"step": 654
},
{
"epoch": 4.21560740144811,
"grad_norm": 0.017552955076098442,
"learning_rate": 2.9746951428553884e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4335648,
"step": 655
},
{
"epoch": 4.222043443282382,
"grad_norm": 0.385110467672348,
"learning_rate": 2.9263101785268253e-07,
"loss": 0.0019,
"num_input_tokens_seen": 4342256,
"step": 656
},
{
"epoch": 4.228479485116654,
"grad_norm": 0.3891147971153259,
"learning_rate": 2.8782974783360534e-07,
"loss": 0.0009,
"num_input_tokens_seen": 4349280,
"step": 657
},
{
"epoch": 4.2349155269509255,
"grad_norm": 0.687170147895813,
"learning_rate": 2.8306578519984526e-07,
"loss": 0.0051,
"num_input_tokens_seen": 4356128,
"step": 658
},
{
"epoch": 4.2413515687851975,
"grad_norm": 0.16641825437545776,
"learning_rate": 2.783392102937682e-07,
"loss": 0.0008,
"num_input_tokens_seen": 4362672,
"step": 659
},
{
"epoch": 4.247787610619469,
"grad_norm": 0.02807171456515789,
"learning_rate": 2.7365010282720954e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4369440,
"step": 660
},
{
"epoch": 4.254223652453741,
"grad_norm": 1.0298210382461548,
"learning_rate": 2.6899854188013054e-07,
"loss": 0.0043,
"num_input_tokens_seen": 4375760,
"step": 661
},
{
"epoch": 4.260659694288013,
"grad_norm": 0.15670017898082733,
"learning_rate": 2.643846058992866e-07,
"loss": 0.0004,
"num_input_tokens_seen": 4382768,
"step": 662
},
{
"epoch": 4.267095736122285,
"grad_norm": 1.2815680503845215,
"learning_rate": 2.5980837269690056e-07,
"loss": 0.0092,
"num_input_tokens_seen": 4389424,
"step": 663
},
{
"epoch": 4.273531777956556,
"grad_norm": 0.23917140066623688,
"learning_rate": 2.552699194493549e-07,
"loss": 0.0006,
"num_input_tokens_seen": 4395904,
"step": 664
},
{
"epoch": 4.279967819790828,
"grad_norm": 0.8005861043930054,
"learning_rate": 2.507693226958871e-07,
"loss": 0.0049,
"num_input_tokens_seen": 4402144,
"step": 665
},
{
"epoch": 4.2864038616251,
"grad_norm": 0.8631348609924316,
"learning_rate": 2.463066583372989e-07,
"loss": 0.0058,
"num_input_tokens_seen": 4408672,
"step": 666
},
{
"epoch": 4.292839903459372,
"grad_norm": 0.017498647794127464,
"learning_rate": 2.418820016346779e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4415040,
"step": 667
},
{
"epoch": 4.299275945293644,
"grad_norm": 0.8143237829208374,
"learning_rate": 2.3749542720812757e-07,
"loss": 0.0063,
"num_input_tokens_seen": 4421696,
"step": 668
},
{
"epoch": 4.305711987127916,
"grad_norm": 0.5140169262886047,
"learning_rate": 2.331470090355084e-07,
"loss": 0.0116,
"num_input_tokens_seen": 4428096,
"step": 669
},
{
"epoch": 4.312148028962188,
"grad_norm": 0.6097451448440552,
"learning_rate": 2.2883682045119066e-07,
"loss": 0.003,
"num_input_tokens_seen": 4434752,
"step": 670
},
{
"epoch": 4.31858407079646,
"grad_norm": 0.030739160254597664,
"learning_rate": 2.2456493414481778e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4441584,
"step": 671
},
{
"epoch": 4.325020112630732,
"grad_norm": 0.017091860994696617,
"learning_rate": 2.2033142216007913e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4448464,
"step": 672
},
{
"epoch": 4.331456154465004,
"grad_norm": 0.18128401041030884,
"learning_rate": 2.1613635589349756e-07,
"loss": 0.0008,
"num_input_tokens_seen": 4454976,
"step": 673
},
{
"epoch": 4.337892196299276,
"grad_norm": 0.0346699096262455,
"learning_rate": 2.1197980609322406e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4461440,
"step": 674
},
{
"epoch": 4.3443282381335475,
"grad_norm": 0.018729638308286667,
"learning_rate": 2.07861842857843e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4468080,
"step": 675
},
{
"epoch": 4.3507642799678194,
"grad_norm": 1.5200186967849731,
"learning_rate": 2.0378253563519247e-07,
"loss": 0.0105,
"num_input_tokens_seen": 4474944,
"step": 676
},
{
"epoch": 4.357200321802091,
"grad_norm": 0.03133641555905342,
"learning_rate": 1.997419532211925e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4481456,
"step": 677
},
{
"epoch": 4.363636363636363,
"grad_norm": 0.012541470117866993,
"learning_rate": 1.9574016375868282e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4487472,
"step": 678
},
{
"epoch": 4.370072405470635,
"grad_norm": 1.6271870136260986,
"learning_rate": 1.9177723473627647e-07,
"loss": 0.0076,
"num_input_tokens_seen": 4494320,
"step": 679
},
{
"epoch": 4.376508447304907,
"grad_norm": 0.03906352072954178,
"learning_rate": 1.8785323298722098e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4501152,
"step": 680
},
{
"epoch": 4.382944489139179,
"grad_norm": 0.024355776607990265,
"learning_rate": 1.839682246882682e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4507376,
"step": 681
},
{
"epoch": 4.389380530973451,
"grad_norm": 0.4699815511703491,
"learning_rate": 1.801222753585638e-07,
"loss": 0.002,
"num_input_tokens_seen": 4513904,
"step": 682
},
{
"epoch": 4.395816572807723,
"grad_norm": 1.0409318208694458,
"learning_rate": 1.7631544985853623e-07,
"loss": 0.0088,
"num_input_tokens_seen": 4520608,
"step": 683
},
{
"epoch": 4.402252614641995,
"grad_norm": 0.42572081089019775,
"learning_rate": 1.725478123888083e-07,
"loss": 0.0023,
"num_input_tokens_seen": 4527184,
"step": 684
},
{
"epoch": 4.408688656476267,
"grad_norm": 0.04756924882531166,
"learning_rate": 1.6881942648911077e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4533696,
"step": 685
},
{
"epoch": 4.415124698310539,
"grad_norm": 0.21432961523532867,
"learning_rate": 1.6513035503721213e-07,
"loss": 0.0017,
"num_input_tokens_seen": 4540624,
"step": 686
},
{
"epoch": 4.421560740144811,
"grad_norm": 0.035157278180122375,
"learning_rate": 1.614806602478583e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4547056,
"step": 687
},
{
"epoch": 4.427996781979083,
"grad_norm": 1.0397815704345703,
"learning_rate": 1.5787040367172379e-07,
"loss": 0.0073,
"num_input_tokens_seen": 4553712,
"step": 688
},
{
"epoch": 4.434432823813355,
"grad_norm": 0.8960546851158142,
"learning_rate": 1.542996461943716e-07,
"loss": 0.0013,
"num_input_tokens_seen": 4560080,
"step": 689
},
{
"epoch": 4.4408688656476265,
"grad_norm": 0.1009814515709877,
"learning_rate": 1.507684480352292e-07,
"loss": 0.0003,
"num_input_tokens_seen": 4566496,
"step": 690
},
{
"epoch": 4.447304907481898,
"grad_norm": 1.6075918674468994,
"learning_rate": 1.4727686874657143e-07,
"loss": 0.0149,
"num_input_tokens_seen": 4573152,
"step": 691
},
{
"epoch": 4.45374094931617,
"grad_norm": 0.5051795840263367,
"learning_rate": 1.4382496721251526e-07,
"loss": 0.0026,
"num_input_tokens_seen": 4580432,
"step": 692
},
{
"epoch": 4.460176991150442,
"grad_norm": 0.01903243362903595,
"learning_rate": 1.4041280164802967e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4587024,
"step": 693
},
{
"epoch": 4.466613032984714,
"grad_norm": 0.04848824068903923,
"learning_rate": 1.3704042959795132e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4593296,
"step": 694
},
{
"epoch": 4.473049074818986,
"grad_norm": 1.155561923980713,
"learning_rate": 1.3370790793601373e-07,
"loss": 0.006,
"num_input_tokens_seen": 4600000,
"step": 695
},
{
"epoch": 4.479485116653258,
"grad_norm": 0.06723422557115555,
"learning_rate": 1.3041529286389078e-07,
"loss": 0.0004,
"num_input_tokens_seen": 4606560,
"step": 696
},
{
"epoch": 4.48592115848753,
"grad_norm": 0.10645350813865662,
"learning_rate": 1.2716263991024712e-07,
"loss": 0.0004,
"num_input_tokens_seen": 4613424,
"step": 697
},
{
"epoch": 4.492357200321802,
"grad_norm": 0.09232950955629349,
"learning_rate": 1.2395000392980057e-07,
"loss": 0.0007,
"num_input_tokens_seen": 4620064,
"step": 698
},
{
"epoch": 4.498793242156074,
"grad_norm": 0.04445146396756172,
"learning_rate": 1.2077743910239998e-07,
"loss": 0.0002,
"num_input_tokens_seen": 4626608,
"step": 699
},
{
"epoch": 4.505229283990346,
"grad_norm": 0.03232429176568985,
"learning_rate": 1.1764499893210879e-07,
"loss": 0.0001,
"num_input_tokens_seen": 4633280,
"step": 700
},
{
"epoch": 4.511665325824618,
"grad_norm": 0.7786117792129517,
"learning_rate": 1.145527362463042e-07,
"loss": 0.0033,
"num_input_tokens_seen": 4639920,
"step": 701
},
{
"epoch": 4.51810136765889,
"grad_norm": 0.15632830560207367,
"learning_rate": 1.1150070319478679e-07,
"loss": 0.0009,
"num_input_tokens_seen": 4646736,
"step": 702
},
{
"epoch": 4.524537409493162,
"grad_norm": 2.964639186859131,
"learning_rate": 1.0848895124889819e-07,
"loss": 0.0216,
"num_input_tokens_seen": 4653328,
"step": 703
},
{
"epoch": 4.530973451327434,
"grad_norm": 0.8460017442703247,
"learning_rate": 1.0551753120065621e-07,
"loss": 0.0035,
"num_input_tokens_seen": 4660112,
"step": 704
},
{
"epoch": 4.5374094931617055,
"grad_norm": 1.1353970766067505,
"learning_rate": 1.0258649316189722e-07,
"loss": 0.0073,
"num_input_tokens_seen": 4666560,
"step": 705
},
{
"epoch": 4.543845534995977,
"grad_norm": 0.08409194648265839,
"learning_rate": 9.969588656342982e-08,
"loss": 0.0003,
"num_input_tokens_seen": 4673152,
"step": 706
},
{
"epoch": 4.550281576830249,
"grad_norm": 0.1840662956237793,
"learning_rate": 9.684576015420277e-08,
"loss": 0.0007,
"num_input_tokens_seen": 4679360,
"step": 707
},
{
"epoch": 4.556717618664521,
"grad_norm": 0.049431972205638885,
"learning_rate": 9.403616200048288e-08,
"loss": 0.0003,
"num_input_tokens_seen": 4685904,
"step": 708
},
{
"epoch": 4.563153660498793,
"grad_norm": 1.309277892112732,
"learning_rate": 9.12671394850423e-08,
"loss": 0.0168,
"num_input_tokens_seen": 4692320,
"step": 709
},
{
"epoch": 4.569589702333065,
"grad_norm": 0.3398638367652893,
"learning_rate": 8.85387393063622e-08,
"loss": 0.0015,
"num_input_tokens_seen": 4698928,
"step": 710
},
{
"epoch": 4.576025744167337,
"grad_norm": 0.016252102330327034,
"learning_rate": 8.585100747784376e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4706000,
"step": 711
},
{
"epoch": 4.582461786001609,
"grad_norm": 0.6447362303733826,
"learning_rate": 8.320398932703145e-08,
"loss": 0.0036,
"num_input_tokens_seen": 4712352,
"step": 712
},
{
"epoch": 4.588897827835881,
"grad_norm": 0.5575013756752014,
"learning_rate": 8.059772949485068e-08,
"loss": 0.0022,
"num_input_tokens_seen": 4718848,
"step": 713
},
{
"epoch": 4.595333869670153,
"grad_norm": 1.0512995719909668,
"learning_rate": 7.803227193485336e-08,
"loss": 0.0085,
"num_input_tokens_seen": 4725728,
"step": 714
},
{
"epoch": 4.601769911504425,
"grad_norm": 0.012490477412939072,
"learning_rate": 7.550765991247655e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4732352,
"step": 715
},
{
"epoch": 4.608205953338697,
"grad_norm": 0.025810543447732925,
"learning_rate": 7.30239360043139e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4738992,
"step": 716
},
{
"epoch": 4.614641995172969,
"grad_norm": 0.5250550508499146,
"learning_rate": 7.058114209739675e-08,
"loss": 0.007,
"num_input_tokens_seen": 4745872,
"step": 717
},
{
"epoch": 4.621078037007241,
"grad_norm": 0.05289880558848381,
"learning_rate": 6.817931938848805e-08,
"loss": 0.0002,
"num_input_tokens_seen": 4752544,
"step": 718
},
{
"epoch": 4.627514078841513,
"grad_norm": 0.0455067902803421,
"learning_rate": 6.581850838338816e-08,
"loss": 0.0002,
"num_input_tokens_seen": 4759360,
"step": 719
},
{
"epoch": 4.6339501206757845,
"grad_norm": 0.1569470316171646,
"learning_rate": 6.349874889624963e-08,
"loss": 0.0008,
"num_input_tokens_seen": 4766016,
"step": 720
},
{
"epoch": 4.640386162510056,
"grad_norm": 0.7371820211410522,
"learning_rate": 6.12200800489085e-08,
"loss": 0.0043,
"num_input_tokens_seen": 4772624,
"step": 721
},
{
"epoch": 4.646822204344328,
"grad_norm": 0.09805099666118622,
"learning_rate": 5.898254027022293e-08,
"loss": 0.0004,
"num_input_tokens_seen": 4778960,
"step": 722
},
{
"epoch": 4.6532582461786,
"grad_norm": 0.6390008926391602,
"learning_rate": 5.678616729542535e-08,
"loss": 0.0024,
"num_input_tokens_seen": 4785600,
"step": 723
},
{
"epoch": 4.659694288012872,
"grad_norm": 0.4965854585170746,
"learning_rate": 5.463099816548578e-08,
"loss": 0.0024,
"num_input_tokens_seen": 4792208,
"step": 724
},
{
"epoch": 4.666130329847144,
"grad_norm": 0.012886490672826767,
"learning_rate": 5.2517069226488694e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4798816,
"step": 725
},
{
"epoch": 4.672566371681416,
"grad_norm": 0.04072566702961922,
"learning_rate": 5.044441612901768e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4805408,
"step": 726
},
{
"epoch": 4.679002413515688,
"grad_norm": 0.07961362600326538,
"learning_rate": 4.841307382755567e-08,
"loss": 0.0002,
"num_input_tokens_seen": 4812480,
"step": 727
},
{
"epoch": 4.68543845534996,
"grad_norm": 0.006482974626123905,
"learning_rate": 4.6423076579895646e-08,
"loss": 0.0,
"num_input_tokens_seen": 4819360,
"step": 728
},
{
"epoch": 4.691874497184232,
"grad_norm": 0.06561946123838425,
"learning_rate": 4.4474457946562245e-08,
"loss": 0.0003,
"num_input_tokens_seen": 4825904,
"step": 729
},
{
"epoch": 4.698310539018504,
"grad_norm": 0.10200546681880951,
"learning_rate": 4.256725079024554e-08,
"loss": 0.0004,
"num_input_tokens_seen": 4832544,
"step": 730
},
{
"epoch": 4.704746580852776,
"grad_norm": 0.2905844449996948,
"learning_rate": 4.070148727524814e-08,
"loss": 0.001,
"num_input_tokens_seen": 4838960,
"step": 731
},
{
"epoch": 4.711182622687048,
"grad_norm": 0.013804874382913113,
"learning_rate": 3.887719886694091e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4845584,
"step": 732
},
{
"epoch": 4.71761866452132,
"grad_norm": 0.8108282685279846,
"learning_rate": 3.709441633123367e-08,
"loss": 0.0031,
"num_input_tokens_seen": 4852416,
"step": 733
},
{
"epoch": 4.7240547063555915,
"grad_norm": 0.11679325252771378,
"learning_rate": 3.535316973405672e-08,
"loss": 0.0006,
"num_input_tokens_seen": 4858864,
"step": 734
},
{
"epoch": 4.7304907481898635,
"grad_norm": 0.013299252837896347,
"learning_rate": 3.3653488440851255e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4865552,
"step": 735
},
{
"epoch": 4.736926790024135,
"grad_norm": 0.8139033913612366,
"learning_rate": 3.1995401116077516e-08,
"loss": 0.006,
"num_input_tokens_seen": 4871984,
"step": 736
},
{
"epoch": 4.743362831858407,
"grad_norm": 0.023770008236169815,
"learning_rate": 3.037893572272937e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4878688,
"step": 737
},
{
"epoch": 4.749798873692679,
"grad_norm": 0.1465252786874771,
"learning_rate": 2.8804119521862183e-08,
"loss": 0.001,
"num_input_tokens_seen": 4885504,
"step": 738
},
{
"epoch": 4.756234915526951,
"grad_norm": 0.7871643304824829,
"learning_rate": 2.7270979072135106e-08,
"loss": 0.0019,
"num_input_tokens_seen": 4892272,
"step": 739
},
{
"epoch": 4.762670957361223,
"grad_norm": 0.6069658994674683,
"learning_rate": 2.5779540229361744e-08,
"loss": 0.0047,
"num_input_tokens_seen": 4898848,
"step": 740
},
{
"epoch": 4.769106999195495,
"grad_norm": 0.07687046378850937,
"learning_rate": 2.4329828146074096e-08,
"loss": 0.0002,
"num_input_tokens_seen": 4905392,
"step": 741
},
{
"epoch": 4.775543041029767,
"grad_norm": 0.0760512426495552,
"learning_rate": 2.2921867271099296e-08,
"loss": 0.0002,
"num_input_tokens_seen": 4912672,
"step": 742
},
{
"epoch": 4.781979082864039,
"grad_norm": 0.14020369946956635,
"learning_rate": 2.155568134914604e-08,
"loss": 0.0006,
"num_input_tokens_seen": 4919472,
"step": 743
},
{
"epoch": 4.788415124698311,
"grad_norm": 0.020505385473370552,
"learning_rate": 2.0231293420405194e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4926448,
"step": 744
},
{
"epoch": 4.794851166532583,
"grad_norm": 0.6544126868247986,
"learning_rate": 1.8948725820160663e-08,
"loss": 0.0038,
"num_input_tokens_seen": 4933152,
"step": 745
},
{
"epoch": 4.801287208366855,
"grad_norm": 1.0113900899887085,
"learning_rate": 1.770800017841301e-08,
"loss": 0.0088,
"num_input_tokens_seen": 4939568,
"step": 746
},
{
"epoch": 4.807723250201127,
"grad_norm": 0.05956251546740532,
"learning_rate": 1.650913741951421e-08,
"loss": 0.0002,
"num_input_tokens_seen": 4946224,
"step": 747
},
{
"epoch": 4.814159292035399,
"grad_norm": 0.21396887302398682,
"learning_rate": 1.5352157761815978e-08,
"loss": 0.002,
"num_input_tokens_seen": 4952880,
"step": 748
},
{
"epoch": 4.8205953338696705,
"grad_norm": 0.3376445472240448,
"learning_rate": 1.4237080717326712e-08,
"loss": 0.0012,
"num_input_tokens_seen": 4960288,
"step": 749
},
{
"epoch": 4.8270313757039425,
"grad_norm": 0.47905248403549194,
"learning_rate": 1.3163925091384532e-08,
"loss": 0.0038,
"num_input_tokens_seen": 4966656,
"step": 750
},
{
"epoch": 4.833467417538214,
"grad_norm": 0.08333203196525574,
"learning_rate": 1.2132708982338925e-08,
"loss": 0.0003,
"num_input_tokens_seen": 4973184,
"step": 751
},
{
"epoch": 4.839903459372486,
"grad_norm": 0.45923787355422974,
"learning_rate": 1.1143449781245985e-08,
"loss": 0.0023,
"num_input_tokens_seen": 4980080,
"step": 752
},
{
"epoch": 4.846339501206758,
"grad_norm": 0.038865748792886734,
"learning_rate": 1.0196164171574762e-08,
"loss": 0.0001,
"num_input_tokens_seen": 4987104,
"step": 753
},
{
"epoch": 4.85277554304103,
"grad_norm": 0.2948670983314514,
"learning_rate": 9.290868128926378e-09,
"loss": 0.0044,
"num_input_tokens_seen": 4993728,
"step": 754
},
{
"epoch": 4.859211584875302,
"grad_norm": 0.5785093903541565,
"learning_rate": 8.427576920763957e-09,
"loss": 0.004,
"num_input_tokens_seen": 5000368,
"step": 755
},
{
"epoch": 4.865647626709574,
"grad_norm": 0.29199209809303284,
"learning_rate": 7.606305106155898e-09,
"loss": 0.0042,
"num_input_tokens_seen": 5006768,
"step": 756
},
{
"epoch": 4.872083668543846,
"grad_norm": 0.25938865542411804,
"learning_rate": 6.827066535529947e-09,
"loss": 0.0004,
"num_input_tokens_seen": 5013264,
"step": 757
},
{
"epoch": 4.878519710378118,
"grad_norm": 0.1507510095834732,
"learning_rate": 6.089874350439507e-09,
"loss": 0.0005,
"num_input_tokens_seen": 5019744,
"step": 758
},
{
"epoch": 4.88495575221239,
"grad_norm": 0.009750754572451115,
"learning_rate": 5.394740983341862e-09,
"loss": 0.0,
"num_input_tokens_seen": 5026080,
"step": 759
},
{
"epoch": 4.891391794046662,
"grad_norm": 0.049791790544986725,
"learning_rate": 4.74167815738974e-09,
"loss": 0.0001,
"num_input_tokens_seen": 5032544,
"step": 760
},
{
"epoch": 4.897827835880933,
"grad_norm": 0.971774160861969,
"learning_rate": 4.130696886231744e-09,
"loss": 0.0089,
"num_input_tokens_seen": 5039248,
"step": 761
},
{
"epoch": 4.904263877715205,
"grad_norm": 0.9287542700767517,
"learning_rate": 3.561807473827783e-09,
"loss": 0.0118,
"num_input_tokens_seen": 5045792,
"step": 762
},
{
"epoch": 4.910699919549477,
"grad_norm": 0.06942977011203766,
"learning_rate": 3.035019514275317e-09,
"loss": 0.0003,
"num_input_tokens_seen": 5052320,
"step": 763
},
{
"epoch": 4.917135961383749,
"grad_norm": 0.28999796509742737,
"learning_rate": 2.5503418916464352e-09,
"loss": 0.0013,
"num_input_tokens_seen": 5059200,
"step": 764
},
{
"epoch": 4.923572003218021,
"grad_norm": 0.8473367691040039,
"learning_rate": 2.1077827798404728e-09,
"loss": 0.0058,
"num_input_tokens_seen": 5065824,
"step": 765
},
{
"epoch": 4.9300080450522925,
"grad_norm": 0.14826533198356628,
"learning_rate": 1.707349642442735e-09,
"loss": 0.0002,
"num_input_tokens_seen": 5072080,
"step": 766
},
{
"epoch": 4.936444086886564,
"grad_norm": 0.9073830246925354,
"learning_rate": 1.349049232601818e-09,
"loss": 0.0059,
"num_input_tokens_seen": 5078336,
"step": 767
},
{
"epoch": 4.942880128720836,
"grad_norm": 1.048936367034912,
"learning_rate": 1.0328875929138671e-09,
"loss": 0.0068,
"num_input_tokens_seen": 5085152,
"step": 768
},
{
"epoch": 4.949316170555108,
"grad_norm": 0.06418836861848831,
"learning_rate": 7.588700553209926e-10,
"loss": 0.0003,
"num_input_tokens_seen": 5092080,
"step": 769
},
{
"epoch": 4.95575221238938,
"grad_norm": 2.129972457885742,
"learning_rate": 5.270012410216185e-10,
"loss": 0.0251,
"num_input_tokens_seen": 5098960,
"step": 770
},
{
"epoch": 4.962188254223652,
"grad_norm": 0.3871181309223175,
"learning_rate": 3.3728506039276686e-10,
"loss": 0.0054,
"num_input_tokens_seen": 5105648,
"step": 771
},
{
"epoch": 4.968624296057924,
"grad_norm": 0.185493603348732,
"learning_rate": 1.8972471292344474e-10,
"loss": 0.0007,
"num_input_tokens_seen": 5112096,
"step": 772
},
{
"epoch": 4.975060337892196,
"grad_norm": 0.05420377105474472,
"learning_rate": 8.432268716135338e-11,
"loss": 0.0002,
"num_input_tokens_seen": 5118960,
"step": 773
},
{
"epoch": 4.981496379726468,
"grad_norm": 0.5486555695533752,
"learning_rate": 2.108076067014464e-11,
"loss": 0.0032,
"num_input_tokens_seen": 5125872,
"step": 774
},
{
"epoch": 4.98793242156074,
"grad_norm": 1.0398619174957275,
"learning_rate": 0.0,
"loss": 0.0107,
"num_input_tokens_seen": 5132288,
"step": 775
},
{
"epoch": 4.98793242156074,
"num_input_tokens_seen": 5132288,
"step": 775,
"total_flos": 2.3110461174474342e+17,
"train_loss": 0.14078616270634925,
"train_runtime": 10317.9615,
"train_samples_per_second": 9.634,
"train_steps_per_second": 0.075
}
],
"logging_steps": 1,
"max_steps": 775,
"num_input_tokens_seen": 5132288,
"num_train_epochs": 5,
"save_steps": 5000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.3110461174474342e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}