llama-3-8b-patent-small-dataset / trainer_state.json
kimhyeongjun's picture
Upload 8 files
0644a28 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.9998846198223146,
"eval_steps": 500,
"global_step": 9750,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.02,
"grad_norm": 2.7104153633117676,
"learning_rate": 4.999532814343219e-05,
"loss": 1.5746,
"step": 60
},
{
"epoch": 0.04,
"grad_norm": 6.236631393432617,
"learning_rate": 4.998131431982826e-05,
"loss": 1.3364,
"step": 120
},
{
"epoch": 0.06,
"grad_norm": 3.3378114700317383,
"learning_rate": 4.99579637668341e-05,
"loss": 1.297,
"step": 180
},
{
"epoch": 0.07,
"grad_norm": 3.676436185836792,
"learning_rate": 4.992528521168449e-05,
"loss": 1.3101,
"step": 240
},
{
"epoch": 0.09,
"grad_norm": 2.182366132736206,
"learning_rate": 4.988329086794122e-05,
"loss": 1.3038,
"step": 300
},
{
"epoch": 0.11,
"grad_norm": 3.702284097671509,
"learning_rate": 4.9831996430928326e-05,
"loss": 1.2762,
"step": 360
},
{
"epoch": 0.13,
"grad_norm": 2.2854039669036865,
"learning_rate": 4.977142107186602e-05,
"loss": 1.2709,
"step": 420
},
{
"epoch": 0.15,
"grad_norm": 4.770021438598633,
"learning_rate": 4.9701587430705415e-05,
"loss": 1.2656,
"step": 480
},
{
"epoch": 0.17,
"grad_norm": 3.181934118270874,
"learning_rate": 4.962252160766693e-05,
"loss": 1.2751,
"step": 540
},
{
"epoch": 0.18,
"grad_norm": 2.0251986980438232,
"learning_rate": 4.953425315348534e-05,
"loss": 1.261,
"step": 600
},
{
"epoch": 0.2,
"grad_norm": 2.5135626792907715,
"learning_rate": 4.943681505836523e-05,
"loss": 1.2627,
"step": 660
},
{
"epoch": 0.22,
"grad_norm": 2.7317075729370117,
"learning_rate": 4.9330243739650964e-05,
"loss": 1.2619,
"step": 720
},
{
"epoch": 0.24,
"grad_norm": 4.191065311431885,
"learning_rate": 4.9214579028215776e-05,
"loss": 1.274,
"step": 780
},
{
"epoch": 0.26,
"grad_norm": 2.875460147857666,
"learning_rate": 4.9089864153575016e-05,
"loss": 1.2138,
"step": 840
},
{
"epoch": 0.28,
"grad_norm": 3.3708271980285645,
"learning_rate": 4.8956145727729156e-05,
"loss": 1.2295,
"step": 900
},
{
"epoch": 0.3,
"grad_norm": 2.718130111694336,
"learning_rate": 4.88134737277427e-05,
"loss": 1.2281,
"step": 960
},
{
"epoch": 0.31,
"grad_norm": 2.6917874813079834,
"learning_rate": 4.8661901477065244e-05,
"loss": 1.2429,
"step": 1020
},
{
"epoch": 0.33,
"grad_norm": 2.3287134170532227,
"learning_rate": 4.8501485625601996e-05,
"loss": 1.2524,
"step": 1080
},
{
"epoch": 0.35,
"grad_norm": 2.6867642402648926,
"learning_rate": 4.833228612854087e-05,
"loss": 1.2149,
"step": 1140
},
{
"epoch": 0.37,
"grad_norm": 3.3543832302093506,
"learning_rate": 4.815436622394441e-05,
"loss": 1.2188,
"step": 1200
},
{
"epoch": 0.39,
"grad_norm": 4.045291900634766,
"learning_rate": 4.7967792409114606e-05,
"loss": 1.2227,
"step": 1260
},
{
"epoch": 0.41,
"grad_norm": 2.516897439956665,
"learning_rate": 4.7772634415739624e-05,
"loss": 1.2365,
"step": 1320
},
{
"epoch": 0.42,
"grad_norm": 3.2049200534820557,
"learning_rate": 4.7568965183831726e-05,
"loss": 1.2102,
"step": 1380
},
{
"epoch": 0.44,
"grad_norm": 3.082902669906616,
"learning_rate": 4.735686083446599e-05,
"loss": 1.2465,
"step": 1440
},
{
"epoch": 0.46,
"grad_norm": 3.1766459941864014,
"learning_rate": 4.713640064133025e-05,
"loss": 1.2446,
"step": 1500
},
{
"epoch": 0.48,
"grad_norm": 3.313269853591919,
"learning_rate": 4.690766700109659e-05,
"loss": 1.2042,
"step": 1560
},
{
"epoch": 0.5,
"grad_norm": 3.3025588989257812,
"learning_rate": 4.667074540262577e-05,
"loss": 1.2229,
"step": 1620
},
{
"epoch": 0.52,
"grad_norm": 3.4626569747924805,
"learning_rate": 4.6425724395015865e-05,
"loss": 1.2064,
"step": 1680
},
{
"epoch": 0.54,
"grad_norm": 3.679563522338867,
"learning_rate": 4.617269555450715e-05,
"loss": 1.214,
"step": 1740
},
{
"epoch": 0.55,
"grad_norm": 2.328498363494873,
"learning_rate": 4.5911753450255665e-05,
"loss": 1.2208,
"step": 1800
},
{
"epoch": 0.57,
"grad_norm": 2.3970301151275635,
"learning_rate": 4.56429956089881e-05,
"loss": 1.1879,
"step": 1860
},
{
"epoch": 0.59,
"grad_norm": 3.7843685150146484,
"learning_rate": 4.5366522478551335e-05,
"loss": 1.2135,
"step": 1920
},
{
"epoch": 0.61,
"grad_norm": 3.4662420749664307,
"learning_rate": 4.508243739037016e-05,
"loss": 1.2137,
"step": 1980
},
{
"epoch": 0.63,
"grad_norm": 3.565755844116211,
"learning_rate": 4.47908465208274e-05,
"loss": 1.2167,
"step": 2040
},
{
"epoch": 0.65,
"grad_norm": 6.096432685852051,
"learning_rate": 4.449185885158056e-05,
"loss": 1.2095,
"step": 2100
},
{
"epoch": 0.66,
"grad_norm": 2.6640658378601074,
"learning_rate": 4.418558612883016e-05,
"loss": 1.2176,
"step": 2160
},
{
"epoch": 0.68,
"grad_norm": 3.7319626808166504,
"learning_rate": 4.387214282155469e-05,
"loss": 1.2092,
"step": 2220
},
{
"epoch": 0.7,
"grad_norm": 2.938650131225586,
"learning_rate": 4.355164607872806e-05,
"loss": 1.2288,
"step": 2280
},
{
"epoch": 0.72,
"grad_norm": 3.252643346786499,
"learning_rate": 4.3224215685535294e-05,
"loss": 1.2019,
"step": 2340
},
{
"epoch": 0.74,
"grad_norm": 4.297436714172363,
"learning_rate": 4.28955998543643e-05,
"loss": 1.2178,
"step": 2400
},
{
"epoch": 0.76,
"grad_norm": 2.8940541744232178,
"learning_rate": 4.255478223794045e-05,
"loss": 1.193,
"step": 2460
},
{
"epoch": 0.78,
"grad_norm": 2.981809377670288,
"learning_rate": 4.2207403547541e-05,
"loss": 1.2246,
"step": 2520
},
{
"epoch": 0.79,
"grad_norm": 2.6929879188537598,
"learning_rate": 4.185359361543927e-05,
"loss": 1.2228,
"step": 2580
},
{
"epoch": 0.81,
"grad_norm": 2.4627578258514404,
"learning_rate": 4.149348467757566e-05,
"loss": 1.2085,
"step": 2640
},
{
"epoch": 0.83,
"grad_norm": 2.8424439430236816,
"learning_rate": 4.112721132413467e-05,
"loss": 1.1943,
"step": 2700
},
{
"epoch": 0.85,
"grad_norm": 6.201737403869629,
"learning_rate": 4.075491044924209e-05,
"loss": 1.2187,
"step": 2760
},
{
"epoch": 0.87,
"grad_norm": 3.3698055744171143,
"learning_rate": 4.0376721199800896e-05,
"loss": 1.1978,
"step": 2820
},
{
"epoch": 0.89,
"grad_norm": 3.9884934425354004,
"learning_rate": 3.999278492348539e-05,
"loss": 1.2312,
"step": 2880
},
{
"epoch": 0.9,
"grad_norm": 2.8434813022613525,
"learning_rate": 3.9603245115912736e-05,
"loss": 1.1852,
"step": 2940
},
{
"epoch": 0.92,
"grad_norm": 3.6574010848999023,
"learning_rate": 3.9208247367011574e-05,
"loss": 1.162,
"step": 3000
},
{
"epoch": 0.94,
"grad_norm": 3.347731828689575,
"learning_rate": 3.880793930660813e-05,
"loss": 1.2028,
"step": 3060
},
{
"epoch": 0.96,
"grad_norm": 2.4751527309417725,
"learning_rate": 3.840247054924968e-05,
"loss": 1.2244,
"step": 3120
},
{
"epoch": 0.98,
"grad_norm": 2.099729299545288,
"learning_rate": 3.79919926382864e-05,
"loss": 1.2152,
"step": 3180
},
{
"epoch": 1.0,
"grad_norm": 3.765007972717285,
"learning_rate": 3.757665898923223e-05,
"loss": 1.1827,
"step": 3240
},
{
"epoch": 1.02,
"grad_norm": 2.1012563705444336,
"learning_rate": 3.715662483242605e-05,
"loss": 0.9704,
"step": 3300
},
{
"epoch": 1.03,
"grad_norm": 2.3256635665893555,
"learning_rate": 3.673204715501461e-05,
"loss": 0.9513,
"step": 3360
},
{
"epoch": 1.05,
"grad_norm": 3.2609546184539795,
"learning_rate": 3.630308464227877e-05,
"loss": 0.9259,
"step": 3420
},
{
"epoch": 1.07,
"grad_norm": 3.2246479988098145,
"learning_rate": 3.5869897618325126e-05,
"loss": 0.9585,
"step": 3480
},
{
"epoch": 1.09,
"grad_norm": 4.168481826782227,
"learning_rate": 3.54326479861651e-05,
"loss": 0.9518,
"step": 3540
},
{
"epoch": 1.11,
"grad_norm": 3.362656354904175,
"learning_rate": 3.499149916720398e-05,
"loss": 0.9482,
"step": 3600
},
{
"epoch": 1.13,
"grad_norm": 3.152392625808716,
"learning_rate": 3.4546616040162334e-05,
"loss": 0.9565,
"step": 3660
},
{
"epoch": 1.14,
"grad_norm": 2.6556153297424316,
"learning_rate": 3.409816487945286e-05,
"loss": 0.9448,
"step": 3720
},
{
"epoch": 1.16,
"grad_norm": 2.465056896209717,
"learning_rate": 3.364631329303564e-05,
"loss": 0.9558,
"step": 3780
},
{
"epoch": 1.18,
"grad_norm": 3.3087334632873535,
"learning_rate": 3.319123015977478e-05,
"loss": 0.9664,
"step": 3840
},
{
"epoch": 1.2,
"grad_norm": 2.0656557083129883,
"learning_rate": 3.2733085566320285e-05,
"loss": 0.9527,
"step": 3900
},
{
"epoch": 1.22,
"grad_norm": 3.9231972694396973,
"learning_rate": 3.2272050743538385e-05,
"loss": 0.9605,
"step": 3960
},
{
"epoch": 1.24,
"grad_norm": 1.9942598342895508,
"learning_rate": 3.180829800251428e-05,
"loss": 0.9232,
"step": 4020
},
{
"epoch": 1.26,
"grad_norm": 2.5485622882843018,
"learning_rate": 3.134200067015108e-05,
"loss": 0.9459,
"step": 4080
},
{
"epoch": 1.27,
"grad_norm": 3.6776840686798096,
"learning_rate": 3.087333302438916e-05,
"loss": 0.9388,
"step": 4140
},
{
"epoch": 1.29,
"grad_norm": 2.463665723800659,
"learning_rate": 3.0402470229070056e-05,
"loss": 0.9466,
"step": 4200
},
{
"epoch": 1.31,
"grad_norm": 2.8538758754730225,
"learning_rate": 2.992958826846918e-05,
"loss": 0.927,
"step": 4260
},
{
"epoch": 1.33,
"grad_norm": 3.169262409210205,
"learning_rate": 2.945486388152201e-05,
"loss": 0.9406,
"step": 4320
},
{
"epoch": 1.35,
"grad_norm": 2.700239896774292,
"learning_rate": 2.897847449576815e-05,
"loss": 0.9546,
"step": 4380
},
{
"epoch": 1.37,
"grad_norm": 2.7255053520202637,
"learning_rate": 2.8500598161038057e-05,
"loss": 0.9512,
"step": 4440
},
{
"epoch": 1.38,
"grad_norm": 2.7375125885009766,
"learning_rate": 2.8021413482907176e-05,
"loss": 0.9426,
"step": 4500
},
{
"epoch": 1.4,
"grad_norm": 2.461376667022705,
"learning_rate": 2.754109955594235e-05,
"loss": 0.9381,
"step": 4560
},
{
"epoch": 1.42,
"grad_norm": 3.127470016479492,
"learning_rate": 2.705983589676554e-05,
"loss": 0.9522,
"step": 4620
},
{
"epoch": 1.44,
"grad_norm": 3.07511830329895,
"learning_rate": 2.6577802376959698e-05,
"loss": 0.955,
"step": 4680
},
{
"epoch": 1.46,
"grad_norm": 2.239698648452759,
"learning_rate": 2.609517915584204e-05,
"loss": 0.9558,
"step": 4740
},
{
"epoch": 1.48,
"grad_norm": 3.7481932640075684,
"learning_rate": 2.5612146613129828e-05,
"loss": 0.9607,
"step": 4800
},
{
"epoch": 1.5,
"grad_norm": 2.2911975383758545,
"learning_rate": 2.5128885281523606e-05,
"loss": 0.9354,
"step": 4860
},
{
"epoch": 1.51,
"grad_norm": 3.977375030517578,
"learning_rate": 2.4645575779233464e-05,
"loss": 0.9593,
"step": 4920
},
{
"epoch": 1.53,
"grad_norm": 2.6258833408355713,
"learning_rate": 2.4162398742473214e-05,
"loss": 0.9403,
"step": 4980
},
{
"epoch": 1.55,
"grad_norm": 3.0968546867370605,
"learning_rate": 2.3679534757947862e-05,
"loss": 0.9555,
"step": 5040
},
{
"epoch": 1.57,
"grad_norm": 3.982872486114502,
"learning_rate": 2.3197164295359593e-05,
"loss": 0.9482,
"step": 5100
},
{
"epoch": 1.59,
"grad_norm": 2.4326817989349365,
"learning_rate": 2.271546763995752e-05,
"loss": 0.9209,
"step": 5160
},
{
"epoch": 1.61,
"grad_norm": 2.619610071182251,
"learning_rate": 2.2234624825156293e-05,
"loss": 0.9113,
"step": 5220
},
{
"epoch": 1.62,
"grad_norm": 2.6039981842041016,
"learning_rate": 2.175481556524892e-05,
"loss": 0.9225,
"step": 5280
},
{
"epoch": 1.64,
"grad_norm": 2.5391623973846436,
"learning_rate": 2.1276219188238768e-05,
"loss": 0.933,
"step": 5340
},
{
"epoch": 1.66,
"grad_norm": 2.85158109664917,
"learning_rate": 2.079901456881601e-05,
"loss": 0.9185,
"step": 5400
},
{
"epoch": 1.68,
"grad_norm": 2.5199923515319824,
"learning_rate": 2.0323380061503494e-05,
"loss": 0.9338,
"step": 5460
},
{
"epoch": 1.7,
"grad_norm": 3.02559757232666,
"learning_rate": 1.9849493433996963e-05,
"loss": 0.9386,
"step": 5520
},
{
"epoch": 1.72,
"grad_norm": 2.78646183013916,
"learning_rate": 1.937753180072466e-05,
"loss": 0.9361,
"step": 5580
},
{
"epoch": 1.74,
"grad_norm": 3.28619122505188,
"learning_rate": 1.8907671556651102e-05,
"loss": 0.9106,
"step": 5640
},
{
"epoch": 1.75,
"grad_norm": 2.7548022270202637,
"learning_rate": 1.8440088311349634e-05,
"loss": 0.9431,
"step": 5700
},
{
"epoch": 1.77,
"grad_norm": 2.226672649383545,
"learning_rate": 1.7974956823368727e-05,
"loss": 0.947,
"step": 5760
},
{
"epoch": 1.79,
"grad_norm": 2.304502010345459,
"learning_rate": 1.7512450934916128e-05,
"loss": 0.9505,
"step": 5820
},
{
"epoch": 1.81,
"grad_norm": 4.344946384429932,
"learning_rate": 1.7052743506885652e-05,
"loss": 0.9486,
"step": 5880
},
{
"epoch": 1.83,
"grad_norm": 3.1504874229431152,
"learning_rate": 1.659600635425066e-05,
"loss": 0.9548,
"step": 5940
},
{
"epoch": 1.85,
"grad_norm": 2.1094934940338135,
"learning_rate": 1.614241018184844e-05,
"loss": 0.9227,
"step": 6000
},
{
"epoch": 1.86,
"grad_norm": 3.3043620586395264,
"learning_rate": 1.5692124520579565e-05,
"loss": 0.9363,
"step": 6060
},
{
"epoch": 1.88,
"grad_norm": 2.956177234649658,
"learning_rate": 1.5245317664045971e-05,
"loss": 0.9332,
"step": 6120
},
{
"epoch": 1.9,
"grad_norm": 2.5686097145080566,
"learning_rate": 1.4802156605651432e-05,
"loss": 0.9186,
"step": 6180
},
{
"epoch": 1.92,
"grad_norm": 1.969947099685669,
"learning_rate": 1.4362806976188054e-05,
"loss": 0.9313,
"step": 6240
},
{
"epoch": 1.94,
"grad_norm": 2.0967071056365967,
"learning_rate": 1.392743298193197e-05,
"loss": 0.9473,
"step": 6300
},
{
"epoch": 1.96,
"grad_norm": 2.932309150695801,
"learning_rate": 1.349619734327156e-05,
"loss": 0.9312,
"step": 6360
},
{
"epoch": 1.98,
"grad_norm": 3.2197911739349365,
"learning_rate": 1.3069261233890891e-05,
"loss": 0.9315,
"step": 6420
},
{
"epoch": 1.99,
"grad_norm": Infinity,
"learning_rate": 1.2653788106514852e-05,
"loss": 0.9393,
"step": 6480
},
{
"epoch": 2.01,
"grad_norm": 1.9021306037902832,
"learning_rate": 1.2235849854764194e-05,
"loss": 0.865,
"step": 6540
},
{
"epoch": 2.03,
"grad_norm": 2.515909194946289,
"learning_rate": 1.1822682185308612e-05,
"loss": 0.8174,
"step": 6600
},
{
"epoch": 2.05,
"grad_norm": 2.4534690380096436,
"learning_rate": 1.1414439518955334e-05,
"loss": 0.8246,
"step": 6660
},
{
"epoch": 2.07,
"grad_norm": 3.6022934913635254,
"learning_rate": 1.101127443579891e-05,
"loss": 0.8026,
"step": 6720
},
{
"epoch": 2.09,
"grad_norm": 2.551551103591919,
"learning_rate": 1.0613337618194691e-05,
"loss": 0.8174,
"step": 6780
},
{
"epoch": 2.1,
"grad_norm": 2.128619432449341,
"learning_rate": 1.022077779444145e-05,
"loss": 0.8298,
"step": 6840
},
{
"epoch": 2.12,
"grad_norm": 2.290801763534546,
"learning_rate": 9.833741683194475e-06,
"loss": 0.7925,
"step": 6900
},
{
"epoch": 2.14,
"grad_norm": 2.4129796028137207,
"learning_rate": 9.452373938629619e-06,
"loss": 0.8262,
"step": 6960
},
{
"epoch": 2.16,
"grad_norm": 2.9145517349243164,
"learning_rate": 9.07681709637905e-06,
"loss": 0.8367,
"step": 7020
},
{
"epoch": 2.18,
"grad_norm": 1.9673601388931274,
"learning_rate": 8.707211520258782e-06,
"loss": 0.8133,
"step": 7080
},
{
"epoch": 2.2,
"grad_norm": 2.545464515686035,
"learning_rate": 8.34369534980789e-06,
"loss": 0.795,
"step": 7140
},
{
"epoch": 2.22,
"grad_norm": 2.039494752883911,
"learning_rate": 7.986404448659023e-06,
"loss": 0.8183,
"step": 7200
},
{
"epoch": 2.23,
"grad_norm": 2.6971044540405273,
"learning_rate": 7.63547235375966e-06,
"loss": 0.8219,
"step": 7260
},
{
"epoch": 2.25,
"grad_norm": 2.3640589714050293,
"learning_rate": 7.291030225462781e-06,
"loss": 0.8124,
"step": 7320
},
{
"epoch": 2.27,
"grad_norm": 2.031846761703491,
"learning_rate": 6.953206798505918e-06,
"loss": 0.8174,
"step": 7380
},
{
"epoch": 2.29,
"grad_norm": 2.62103533744812,
"learning_rate": 6.622128333896768e-06,
"loss": 0.806,
"step": 7440
},
{
"epoch": 2.31,
"grad_norm": 2.6120152473449707,
"learning_rate": 6.297918571723288e-06,
"loss": 0.8114,
"step": 7500
},
{
"epoch": 2.33,
"grad_norm": 2.473111391067505,
"learning_rate": 5.980698684905989e-06,
"loss": 0.8098,
"step": 7560
},
{
"epoch": 2.34,
"grad_norm": 2.5968923568725586,
"learning_rate": 5.6705872339098186e-06,
"loss": 0.8205,
"step": 7620
},
{
"epoch": 2.36,
"grad_norm": 2.617664098739624,
"learning_rate": 5.367700122432315e-06,
"loss": 0.8067,
"step": 7680
},
{
"epoch": 2.38,
"grad_norm": 1.9892934560775757,
"learning_rate": 5.072150554084745e-06,
"loss": 0.8287,
"step": 7740
},
{
"epoch": 2.4,
"grad_norm": 2.1314797401428223,
"learning_rate": 4.784048990082484e-06,
"loss": 0.8267,
"step": 7800
},
{
"epoch": 2.42,
"grad_norm": 3.0698599815368652,
"learning_rate": 4.5035031079602445e-06,
"loss": 0.8079,
"step": 7860
},
{
"epoch": 2.44,
"grad_norm": 2.567479133605957,
"learning_rate": 4.2306177613277765e-06,
"loss": 0.8262,
"step": 7920
},
{
"epoch": 2.46,
"grad_norm": 2.587759017944336,
"learning_rate": 3.9654949406809995e-06,
"loss": 0.8223,
"step": 7980
},
{
"epoch": 2.47,
"grad_norm": 2.034844398498535,
"learning_rate": 3.7082337352831923e-06,
"loss": 0.8114,
"step": 8040
},
{
"epoch": 2.49,
"grad_norm": 3.257702589035034,
"learning_rate": 3.458930296130519e-06,
"loss": 0.8267,
"step": 8100
},
{
"epoch": 2.51,
"grad_norm": 2.539806365966797,
"learning_rate": 3.2176778000157367e-06,
"loss": 0.8251,
"step": 8160
},
{
"epoch": 2.53,
"grad_norm": 2.7473506927490234,
"learning_rate": 2.9845664147035326e-06,
"loss": 0.7826,
"step": 8220
},
{
"epoch": 2.55,
"grad_norm": 2.324519395828247,
"learning_rate": 2.7596832652304283e-06,
"loss": 0.8178,
"step": 8280
},
{
"epoch": 2.57,
"grad_norm": 2.691235065460205,
"learning_rate": 2.5431124013419237e-06,
"loss": 0.7982,
"step": 8340
},
{
"epoch": 2.58,
"grad_norm": 3.084143877029419,
"learning_rate": 2.338335189634505e-06,
"loss": 0.8007,
"step": 8400
},
{
"epoch": 2.6,
"grad_norm": 2.8486242294311523,
"learning_rate": 2.138486784704746e-06,
"loss": 0.8192,
"step": 8460
},
{
"epoch": 2.62,
"grad_norm": 2.290395975112915,
"learning_rate": 1.947182836628192e-06,
"loss": 0.8303,
"step": 8520
},
{
"epoch": 2.64,
"grad_norm": 2.120853900909424,
"learning_rate": 1.7644948449733384e-06,
"loss": 0.8317,
"step": 8580
},
{
"epoch": 2.66,
"grad_norm": 2.403571605682373,
"learning_rate": 1.590491089107679e-06,
"loss": 0.8012,
"step": 8640
},
{
"epoch": 2.68,
"grad_norm": 1.6690127849578857,
"learning_rate": 1.425236602678387e-06,
"loss": 0.8241,
"step": 8700
},
{
"epoch": 2.7,
"grad_norm": 3.3557851314544678,
"learning_rate": 1.268793149306091e-06,
"loss": 0.8234,
"step": 8760
},
{
"epoch": 2.71,
"grad_norm": 3.1359851360321045,
"learning_rate": 1.1212191995007975e-06,
"loss": 0.802,
"step": 8820
},
{
"epoch": 2.73,
"grad_norm": 2.1641125679016113,
"learning_rate": 9.825699088086338e-07,
"loss": 0.8138,
"step": 8880
},
{
"epoch": 2.75,
"grad_norm": 2.6551270484924316,
"learning_rate": 8.528970971975553e-07,
"loss": 0.8375,
"step": 8940
},
{
"epoch": 2.77,
"grad_norm": 1.8573861122131348,
"learning_rate": 7.322492296896799e-07,
"loss": 0.8166,
"step": 9000
},
{
"epoch": 2.79,
"grad_norm": 2.5303752422332764,
"learning_rate": 6.206713982475909e-07,
"loss": 0.8346,
"step": 9060
},
{
"epoch": 2.81,
"grad_norm": 2.4010512828826904,
"learning_rate": 5.182053049212626e-07,
"loss": 0.8077,
"step": 9120
},
{
"epoch": 2.82,
"grad_norm": 2.183361053466797,
"learning_rate": 4.248892462619725e-07,
"loss": 0.799,
"step": 9180
},
{
"epoch": 2.84,
"grad_norm": 2.452626943588257,
"learning_rate": 3.4075809900904756e-07,
"loss": 0.8143,
"step": 9240
},
{
"epoch": 2.86,
"grad_norm": 2.959763526916504,
"learning_rate": 2.658433070547195e-07,
"loss": 0.8163,
"step": 9300
},
{
"epoch": 2.88,
"grad_norm": 2.3562393188476562,
"learning_rate": 2.0017286969200578e-07,
"loss": 0.8134,
"step": 9360
},
{
"epoch": 2.9,
"grad_norm": 1.8572157621383667,
"learning_rate": 1.4377133115004438e-07,
"loss": 0.8188,
"step": 9420
},
{
"epoch": 2.92,
"grad_norm": 3.5713188648223877,
"learning_rate": 9.665977142068738e-08,
"loss": 0.8248,
"step": 9480
},
{
"epoch": 2.94,
"grad_norm": 2.22605562210083,
"learning_rate": 5.885579837992261e-08,
"loss": 0.7936,
"step": 9540
},
{
"epoch": 2.95,
"grad_norm": 2.264847993850708,
"learning_rate": 3.037354120692393e-08,
"loss": 0.8273,
"step": 9600
},
{
"epoch": 2.97,
"grad_norm": 2.1906256675720215,
"learning_rate": 1.1223645103311531e-08,
"loss": 0.8199,
"step": 9660
},
{
"epoch": 2.99,
"grad_norm": 1.983068823814392,
"learning_rate": 1.413267314517852e-09,
"loss": 0.813,
"step": 9720
},
{
"epoch": 3.0,
"step": 9750,
"total_flos": 1.6697107709546988e+18,
"train_loss": 0.9979035949707031,
"train_runtime": 4718.1281,
"train_samples_per_second": 33.065,
"train_steps_per_second": 2.066
}
],
"logging_steps": 60,
"max_steps": 9750,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 5000,
"total_flos": 1.6697107709546988e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}