|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.9998846198223146, |
|
"eval_steps": 500, |
|
"global_step": 9750, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 2.7104153633117676, |
|
"learning_rate": 4.999532814343219e-05, |
|
"loss": 1.5746, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 6.236631393432617, |
|
"learning_rate": 4.998131431982826e-05, |
|
"loss": 1.3364, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 3.3378114700317383, |
|
"learning_rate": 4.99579637668341e-05, |
|
"loss": 1.297, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 3.676436185836792, |
|
"learning_rate": 4.992528521168449e-05, |
|
"loss": 1.3101, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 2.182366132736206, |
|
"learning_rate": 4.988329086794122e-05, |
|
"loss": 1.3038, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 3.702284097671509, |
|
"learning_rate": 4.9831996430928326e-05, |
|
"loss": 1.2762, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 2.2854039669036865, |
|
"learning_rate": 4.977142107186602e-05, |
|
"loss": 1.2709, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 4.770021438598633, |
|
"learning_rate": 4.9701587430705415e-05, |
|
"loss": 1.2656, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 3.181934118270874, |
|
"learning_rate": 4.962252160766693e-05, |
|
"loss": 1.2751, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 2.0251986980438232, |
|
"learning_rate": 4.953425315348534e-05, |
|
"loss": 1.261, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 2.5135626792907715, |
|
"learning_rate": 4.943681505836523e-05, |
|
"loss": 1.2627, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 2.7317075729370117, |
|
"learning_rate": 4.9330243739650964e-05, |
|
"loss": 1.2619, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 4.191065311431885, |
|
"learning_rate": 4.9214579028215776e-05, |
|
"loss": 1.274, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 2.875460147857666, |
|
"learning_rate": 4.9089864153575016e-05, |
|
"loss": 1.2138, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 3.3708271980285645, |
|
"learning_rate": 4.8956145727729156e-05, |
|
"loss": 1.2295, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 2.718130111694336, |
|
"learning_rate": 4.88134737277427e-05, |
|
"loss": 1.2281, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 2.6917874813079834, |
|
"learning_rate": 4.8661901477065244e-05, |
|
"loss": 1.2429, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 2.3287134170532227, |
|
"learning_rate": 4.8501485625601996e-05, |
|
"loss": 1.2524, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 2.6867642402648926, |
|
"learning_rate": 4.833228612854087e-05, |
|
"loss": 1.2149, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 3.3543832302093506, |
|
"learning_rate": 4.815436622394441e-05, |
|
"loss": 1.2188, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 4.045291900634766, |
|
"learning_rate": 4.7967792409114606e-05, |
|
"loss": 1.2227, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 2.516897439956665, |
|
"learning_rate": 4.7772634415739624e-05, |
|
"loss": 1.2365, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 3.2049200534820557, |
|
"learning_rate": 4.7568965183831726e-05, |
|
"loss": 1.2102, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 3.082902669906616, |
|
"learning_rate": 4.735686083446599e-05, |
|
"loss": 1.2465, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 3.1766459941864014, |
|
"learning_rate": 4.713640064133025e-05, |
|
"loss": 1.2446, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.48, |
|
"grad_norm": 3.313269853591919, |
|
"learning_rate": 4.690766700109659e-05, |
|
"loss": 1.2042, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.5, |
|
"grad_norm": 3.3025588989257812, |
|
"learning_rate": 4.667074540262577e-05, |
|
"loss": 1.2229, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.52, |
|
"grad_norm": 3.4626569747924805, |
|
"learning_rate": 4.6425724395015865e-05, |
|
"loss": 1.2064, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.54, |
|
"grad_norm": 3.679563522338867, |
|
"learning_rate": 4.617269555450715e-05, |
|
"loss": 1.214, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.55, |
|
"grad_norm": 2.328498363494873, |
|
"learning_rate": 4.5911753450255665e-05, |
|
"loss": 1.2208, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.57, |
|
"grad_norm": 2.3970301151275635, |
|
"learning_rate": 4.56429956089881e-05, |
|
"loss": 1.1879, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.59, |
|
"grad_norm": 3.7843685150146484, |
|
"learning_rate": 4.5366522478551335e-05, |
|
"loss": 1.2135, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.61, |
|
"grad_norm": 3.4662420749664307, |
|
"learning_rate": 4.508243739037016e-05, |
|
"loss": 1.2137, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.63, |
|
"grad_norm": 3.565755844116211, |
|
"learning_rate": 4.47908465208274e-05, |
|
"loss": 1.2167, |
|
"step": 2040 |
|
}, |
|
{ |
|
"epoch": 0.65, |
|
"grad_norm": 6.096432685852051, |
|
"learning_rate": 4.449185885158056e-05, |
|
"loss": 1.2095, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.66, |
|
"grad_norm": 2.6640658378601074, |
|
"learning_rate": 4.418558612883016e-05, |
|
"loss": 1.2176, |
|
"step": 2160 |
|
}, |
|
{ |
|
"epoch": 0.68, |
|
"grad_norm": 3.7319626808166504, |
|
"learning_rate": 4.387214282155469e-05, |
|
"loss": 1.2092, |
|
"step": 2220 |
|
}, |
|
{ |
|
"epoch": 0.7, |
|
"grad_norm": 2.938650131225586, |
|
"learning_rate": 4.355164607872806e-05, |
|
"loss": 1.2288, |
|
"step": 2280 |
|
}, |
|
{ |
|
"epoch": 0.72, |
|
"grad_norm": 3.252643346786499, |
|
"learning_rate": 4.3224215685535294e-05, |
|
"loss": 1.2019, |
|
"step": 2340 |
|
}, |
|
{ |
|
"epoch": 0.74, |
|
"grad_norm": 4.297436714172363, |
|
"learning_rate": 4.28955998543643e-05, |
|
"loss": 1.2178, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.76, |
|
"grad_norm": 2.8940541744232178, |
|
"learning_rate": 4.255478223794045e-05, |
|
"loss": 1.193, |
|
"step": 2460 |
|
}, |
|
{ |
|
"epoch": 0.78, |
|
"grad_norm": 2.981809377670288, |
|
"learning_rate": 4.2207403547541e-05, |
|
"loss": 1.2246, |
|
"step": 2520 |
|
}, |
|
{ |
|
"epoch": 0.79, |
|
"grad_norm": 2.6929879188537598, |
|
"learning_rate": 4.185359361543927e-05, |
|
"loss": 1.2228, |
|
"step": 2580 |
|
}, |
|
{ |
|
"epoch": 0.81, |
|
"grad_norm": 2.4627578258514404, |
|
"learning_rate": 4.149348467757566e-05, |
|
"loss": 1.2085, |
|
"step": 2640 |
|
}, |
|
{ |
|
"epoch": 0.83, |
|
"grad_norm": 2.8424439430236816, |
|
"learning_rate": 4.112721132413467e-05, |
|
"loss": 1.1943, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.85, |
|
"grad_norm": 6.201737403869629, |
|
"learning_rate": 4.075491044924209e-05, |
|
"loss": 1.2187, |
|
"step": 2760 |
|
}, |
|
{ |
|
"epoch": 0.87, |
|
"grad_norm": 3.3698055744171143, |
|
"learning_rate": 4.0376721199800896e-05, |
|
"loss": 1.1978, |
|
"step": 2820 |
|
}, |
|
{ |
|
"epoch": 0.89, |
|
"grad_norm": 3.9884934425354004, |
|
"learning_rate": 3.999278492348539e-05, |
|
"loss": 1.2312, |
|
"step": 2880 |
|
}, |
|
{ |
|
"epoch": 0.9, |
|
"grad_norm": 2.8434813022613525, |
|
"learning_rate": 3.9603245115912736e-05, |
|
"loss": 1.1852, |
|
"step": 2940 |
|
}, |
|
{ |
|
"epoch": 0.92, |
|
"grad_norm": 3.6574010848999023, |
|
"learning_rate": 3.9208247367011574e-05, |
|
"loss": 1.162, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.94, |
|
"grad_norm": 3.347731828689575, |
|
"learning_rate": 3.880793930660813e-05, |
|
"loss": 1.2028, |
|
"step": 3060 |
|
}, |
|
{ |
|
"epoch": 0.96, |
|
"grad_norm": 2.4751527309417725, |
|
"learning_rate": 3.840247054924968e-05, |
|
"loss": 1.2244, |
|
"step": 3120 |
|
}, |
|
{ |
|
"epoch": 0.98, |
|
"grad_norm": 2.099729299545288, |
|
"learning_rate": 3.79919926382864e-05, |
|
"loss": 1.2152, |
|
"step": 3180 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 3.765007972717285, |
|
"learning_rate": 3.757665898923223e-05, |
|
"loss": 1.1827, |
|
"step": 3240 |
|
}, |
|
{ |
|
"epoch": 1.02, |
|
"grad_norm": 2.1012563705444336, |
|
"learning_rate": 3.715662483242605e-05, |
|
"loss": 0.9704, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 1.03, |
|
"grad_norm": 2.3256635665893555, |
|
"learning_rate": 3.673204715501461e-05, |
|
"loss": 0.9513, |
|
"step": 3360 |
|
}, |
|
{ |
|
"epoch": 1.05, |
|
"grad_norm": 3.2609546184539795, |
|
"learning_rate": 3.630308464227877e-05, |
|
"loss": 0.9259, |
|
"step": 3420 |
|
}, |
|
{ |
|
"epoch": 1.07, |
|
"grad_norm": 3.2246479988098145, |
|
"learning_rate": 3.5869897618325126e-05, |
|
"loss": 0.9585, |
|
"step": 3480 |
|
}, |
|
{ |
|
"epoch": 1.09, |
|
"grad_norm": 4.168481826782227, |
|
"learning_rate": 3.54326479861651e-05, |
|
"loss": 0.9518, |
|
"step": 3540 |
|
}, |
|
{ |
|
"epoch": 1.11, |
|
"grad_norm": 3.362656354904175, |
|
"learning_rate": 3.499149916720398e-05, |
|
"loss": 0.9482, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 1.13, |
|
"grad_norm": 3.152392625808716, |
|
"learning_rate": 3.4546616040162334e-05, |
|
"loss": 0.9565, |
|
"step": 3660 |
|
}, |
|
{ |
|
"epoch": 1.14, |
|
"grad_norm": 2.6556153297424316, |
|
"learning_rate": 3.409816487945286e-05, |
|
"loss": 0.9448, |
|
"step": 3720 |
|
}, |
|
{ |
|
"epoch": 1.16, |
|
"grad_norm": 2.465056896209717, |
|
"learning_rate": 3.364631329303564e-05, |
|
"loss": 0.9558, |
|
"step": 3780 |
|
}, |
|
{ |
|
"epoch": 1.18, |
|
"grad_norm": 3.3087334632873535, |
|
"learning_rate": 3.319123015977478e-05, |
|
"loss": 0.9664, |
|
"step": 3840 |
|
}, |
|
{ |
|
"epoch": 1.2, |
|
"grad_norm": 2.0656557083129883, |
|
"learning_rate": 3.2733085566320285e-05, |
|
"loss": 0.9527, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 1.22, |
|
"grad_norm": 3.9231972694396973, |
|
"learning_rate": 3.2272050743538385e-05, |
|
"loss": 0.9605, |
|
"step": 3960 |
|
}, |
|
{ |
|
"epoch": 1.24, |
|
"grad_norm": 1.9942598342895508, |
|
"learning_rate": 3.180829800251428e-05, |
|
"loss": 0.9232, |
|
"step": 4020 |
|
}, |
|
{ |
|
"epoch": 1.26, |
|
"grad_norm": 2.5485622882843018, |
|
"learning_rate": 3.134200067015108e-05, |
|
"loss": 0.9459, |
|
"step": 4080 |
|
}, |
|
{ |
|
"epoch": 1.27, |
|
"grad_norm": 3.6776840686798096, |
|
"learning_rate": 3.087333302438916e-05, |
|
"loss": 0.9388, |
|
"step": 4140 |
|
}, |
|
{ |
|
"epoch": 1.29, |
|
"grad_norm": 2.463665723800659, |
|
"learning_rate": 3.0402470229070056e-05, |
|
"loss": 0.9466, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 1.31, |
|
"grad_norm": 2.8538758754730225, |
|
"learning_rate": 2.992958826846918e-05, |
|
"loss": 0.927, |
|
"step": 4260 |
|
}, |
|
{ |
|
"epoch": 1.33, |
|
"grad_norm": 3.169262409210205, |
|
"learning_rate": 2.945486388152201e-05, |
|
"loss": 0.9406, |
|
"step": 4320 |
|
}, |
|
{ |
|
"epoch": 1.35, |
|
"grad_norm": 2.700239896774292, |
|
"learning_rate": 2.897847449576815e-05, |
|
"loss": 0.9546, |
|
"step": 4380 |
|
}, |
|
{ |
|
"epoch": 1.37, |
|
"grad_norm": 2.7255053520202637, |
|
"learning_rate": 2.8500598161038057e-05, |
|
"loss": 0.9512, |
|
"step": 4440 |
|
}, |
|
{ |
|
"epoch": 1.38, |
|
"grad_norm": 2.7375125885009766, |
|
"learning_rate": 2.8021413482907176e-05, |
|
"loss": 0.9426, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 1.4, |
|
"grad_norm": 2.461376667022705, |
|
"learning_rate": 2.754109955594235e-05, |
|
"loss": 0.9381, |
|
"step": 4560 |
|
}, |
|
{ |
|
"epoch": 1.42, |
|
"grad_norm": 3.127470016479492, |
|
"learning_rate": 2.705983589676554e-05, |
|
"loss": 0.9522, |
|
"step": 4620 |
|
}, |
|
{ |
|
"epoch": 1.44, |
|
"grad_norm": 3.07511830329895, |
|
"learning_rate": 2.6577802376959698e-05, |
|
"loss": 0.955, |
|
"step": 4680 |
|
}, |
|
{ |
|
"epoch": 1.46, |
|
"grad_norm": 2.239698648452759, |
|
"learning_rate": 2.609517915584204e-05, |
|
"loss": 0.9558, |
|
"step": 4740 |
|
}, |
|
{ |
|
"epoch": 1.48, |
|
"grad_norm": 3.7481932640075684, |
|
"learning_rate": 2.5612146613129828e-05, |
|
"loss": 0.9607, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 1.5, |
|
"grad_norm": 2.2911975383758545, |
|
"learning_rate": 2.5128885281523606e-05, |
|
"loss": 0.9354, |
|
"step": 4860 |
|
}, |
|
{ |
|
"epoch": 1.51, |
|
"grad_norm": 3.977375030517578, |
|
"learning_rate": 2.4645575779233464e-05, |
|
"loss": 0.9593, |
|
"step": 4920 |
|
}, |
|
{ |
|
"epoch": 1.53, |
|
"grad_norm": 2.6258833408355713, |
|
"learning_rate": 2.4162398742473214e-05, |
|
"loss": 0.9403, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 1.55, |
|
"grad_norm": 3.0968546867370605, |
|
"learning_rate": 2.3679534757947862e-05, |
|
"loss": 0.9555, |
|
"step": 5040 |
|
}, |
|
{ |
|
"epoch": 1.57, |
|
"grad_norm": 3.982872486114502, |
|
"learning_rate": 2.3197164295359593e-05, |
|
"loss": 0.9482, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 1.59, |
|
"grad_norm": 2.4326817989349365, |
|
"learning_rate": 2.271546763995752e-05, |
|
"loss": 0.9209, |
|
"step": 5160 |
|
}, |
|
{ |
|
"epoch": 1.61, |
|
"grad_norm": 2.619610071182251, |
|
"learning_rate": 2.2234624825156293e-05, |
|
"loss": 0.9113, |
|
"step": 5220 |
|
}, |
|
{ |
|
"epoch": 1.62, |
|
"grad_norm": 2.6039981842041016, |
|
"learning_rate": 2.175481556524892e-05, |
|
"loss": 0.9225, |
|
"step": 5280 |
|
}, |
|
{ |
|
"epoch": 1.64, |
|
"grad_norm": 2.5391623973846436, |
|
"learning_rate": 2.1276219188238768e-05, |
|
"loss": 0.933, |
|
"step": 5340 |
|
}, |
|
{ |
|
"epoch": 1.66, |
|
"grad_norm": 2.85158109664917, |
|
"learning_rate": 2.079901456881601e-05, |
|
"loss": 0.9185, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 1.68, |
|
"grad_norm": 2.5199923515319824, |
|
"learning_rate": 2.0323380061503494e-05, |
|
"loss": 0.9338, |
|
"step": 5460 |
|
}, |
|
{ |
|
"epoch": 1.7, |
|
"grad_norm": 3.02559757232666, |
|
"learning_rate": 1.9849493433996963e-05, |
|
"loss": 0.9386, |
|
"step": 5520 |
|
}, |
|
{ |
|
"epoch": 1.72, |
|
"grad_norm": 2.78646183013916, |
|
"learning_rate": 1.937753180072466e-05, |
|
"loss": 0.9361, |
|
"step": 5580 |
|
}, |
|
{ |
|
"epoch": 1.74, |
|
"grad_norm": 3.28619122505188, |
|
"learning_rate": 1.8907671556651102e-05, |
|
"loss": 0.9106, |
|
"step": 5640 |
|
}, |
|
{ |
|
"epoch": 1.75, |
|
"grad_norm": 2.7548022270202637, |
|
"learning_rate": 1.8440088311349634e-05, |
|
"loss": 0.9431, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 1.77, |
|
"grad_norm": 2.226672649383545, |
|
"learning_rate": 1.7974956823368727e-05, |
|
"loss": 0.947, |
|
"step": 5760 |
|
}, |
|
{ |
|
"epoch": 1.79, |
|
"grad_norm": 2.304502010345459, |
|
"learning_rate": 1.7512450934916128e-05, |
|
"loss": 0.9505, |
|
"step": 5820 |
|
}, |
|
{ |
|
"epoch": 1.81, |
|
"grad_norm": 4.344946384429932, |
|
"learning_rate": 1.7052743506885652e-05, |
|
"loss": 0.9486, |
|
"step": 5880 |
|
}, |
|
{ |
|
"epoch": 1.83, |
|
"grad_norm": 3.1504874229431152, |
|
"learning_rate": 1.659600635425066e-05, |
|
"loss": 0.9548, |
|
"step": 5940 |
|
}, |
|
{ |
|
"epoch": 1.85, |
|
"grad_norm": 2.1094934940338135, |
|
"learning_rate": 1.614241018184844e-05, |
|
"loss": 0.9227, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 1.86, |
|
"grad_norm": 3.3043620586395264, |
|
"learning_rate": 1.5692124520579565e-05, |
|
"loss": 0.9363, |
|
"step": 6060 |
|
}, |
|
{ |
|
"epoch": 1.88, |
|
"grad_norm": 2.956177234649658, |
|
"learning_rate": 1.5245317664045971e-05, |
|
"loss": 0.9332, |
|
"step": 6120 |
|
}, |
|
{ |
|
"epoch": 1.9, |
|
"grad_norm": 2.5686097145080566, |
|
"learning_rate": 1.4802156605651432e-05, |
|
"loss": 0.9186, |
|
"step": 6180 |
|
}, |
|
{ |
|
"epoch": 1.92, |
|
"grad_norm": 1.969947099685669, |
|
"learning_rate": 1.4362806976188054e-05, |
|
"loss": 0.9313, |
|
"step": 6240 |
|
}, |
|
{ |
|
"epoch": 1.94, |
|
"grad_norm": 2.0967071056365967, |
|
"learning_rate": 1.392743298193197e-05, |
|
"loss": 0.9473, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 1.96, |
|
"grad_norm": 2.932309150695801, |
|
"learning_rate": 1.349619734327156e-05, |
|
"loss": 0.9312, |
|
"step": 6360 |
|
}, |
|
{ |
|
"epoch": 1.98, |
|
"grad_norm": 3.2197911739349365, |
|
"learning_rate": 1.3069261233890891e-05, |
|
"loss": 0.9315, |
|
"step": 6420 |
|
}, |
|
{ |
|
"epoch": 1.99, |
|
"grad_norm": Infinity, |
|
"learning_rate": 1.2653788106514852e-05, |
|
"loss": 0.9393, |
|
"step": 6480 |
|
}, |
|
{ |
|
"epoch": 2.01, |
|
"grad_norm": 1.9021306037902832, |
|
"learning_rate": 1.2235849854764194e-05, |
|
"loss": 0.865, |
|
"step": 6540 |
|
}, |
|
{ |
|
"epoch": 2.03, |
|
"grad_norm": 2.515909194946289, |
|
"learning_rate": 1.1822682185308612e-05, |
|
"loss": 0.8174, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 2.05, |
|
"grad_norm": 2.4534690380096436, |
|
"learning_rate": 1.1414439518955334e-05, |
|
"loss": 0.8246, |
|
"step": 6660 |
|
}, |
|
{ |
|
"epoch": 2.07, |
|
"grad_norm": 3.6022934913635254, |
|
"learning_rate": 1.101127443579891e-05, |
|
"loss": 0.8026, |
|
"step": 6720 |
|
}, |
|
{ |
|
"epoch": 2.09, |
|
"grad_norm": 2.551551103591919, |
|
"learning_rate": 1.0613337618194691e-05, |
|
"loss": 0.8174, |
|
"step": 6780 |
|
}, |
|
{ |
|
"epoch": 2.1, |
|
"grad_norm": 2.128619432449341, |
|
"learning_rate": 1.022077779444145e-05, |
|
"loss": 0.8298, |
|
"step": 6840 |
|
}, |
|
{ |
|
"epoch": 2.12, |
|
"grad_norm": 2.290801763534546, |
|
"learning_rate": 9.833741683194475e-06, |
|
"loss": 0.7925, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 2.14, |
|
"grad_norm": 2.4129796028137207, |
|
"learning_rate": 9.452373938629619e-06, |
|
"loss": 0.8262, |
|
"step": 6960 |
|
}, |
|
{ |
|
"epoch": 2.16, |
|
"grad_norm": 2.9145517349243164, |
|
"learning_rate": 9.07681709637905e-06, |
|
"loss": 0.8367, |
|
"step": 7020 |
|
}, |
|
{ |
|
"epoch": 2.18, |
|
"grad_norm": 1.9673601388931274, |
|
"learning_rate": 8.707211520258782e-06, |
|
"loss": 0.8133, |
|
"step": 7080 |
|
}, |
|
{ |
|
"epoch": 2.2, |
|
"grad_norm": 2.545464515686035, |
|
"learning_rate": 8.34369534980789e-06, |
|
"loss": 0.795, |
|
"step": 7140 |
|
}, |
|
{ |
|
"epoch": 2.22, |
|
"grad_norm": 2.039494752883911, |
|
"learning_rate": 7.986404448659023e-06, |
|
"loss": 0.8183, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 2.23, |
|
"grad_norm": 2.6971044540405273, |
|
"learning_rate": 7.63547235375966e-06, |
|
"loss": 0.8219, |
|
"step": 7260 |
|
}, |
|
{ |
|
"epoch": 2.25, |
|
"grad_norm": 2.3640589714050293, |
|
"learning_rate": 7.291030225462781e-06, |
|
"loss": 0.8124, |
|
"step": 7320 |
|
}, |
|
{ |
|
"epoch": 2.27, |
|
"grad_norm": 2.031846761703491, |
|
"learning_rate": 6.953206798505918e-06, |
|
"loss": 0.8174, |
|
"step": 7380 |
|
}, |
|
{ |
|
"epoch": 2.29, |
|
"grad_norm": 2.62103533744812, |
|
"learning_rate": 6.622128333896768e-06, |
|
"loss": 0.806, |
|
"step": 7440 |
|
}, |
|
{ |
|
"epoch": 2.31, |
|
"grad_norm": 2.6120152473449707, |
|
"learning_rate": 6.297918571723288e-06, |
|
"loss": 0.8114, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 2.33, |
|
"grad_norm": 2.473111391067505, |
|
"learning_rate": 5.980698684905989e-06, |
|
"loss": 0.8098, |
|
"step": 7560 |
|
}, |
|
{ |
|
"epoch": 2.34, |
|
"grad_norm": 2.5968923568725586, |
|
"learning_rate": 5.6705872339098186e-06, |
|
"loss": 0.8205, |
|
"step": 7620 |
|
}, |
|
{ |
|
"epoch": 2.36, |
|
"grad_norm": 2.617664098739624, |
|
"learning_rate": 5.367700122432315e-06, |
|
"loss": 0.8067, |
|
"step": 7680 |
|
}, |
|
{ |
|
"epoch": 2.38, |
|
"grad_norm": 1.9892934560775757, |
|
"learning_rate": 5.072150554084745e-06, |
|
"loss": 0.8287, |
|
"step": 7740 |
|
}, |
|
{ |
|
"epoch": 2.4, |
|
"grad_norm": 2.1314797401428223, |
|
"learning_rate": 4.784048990082484e-06, |
|
"loss": 0.8267, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 2.42, |
|
"grad_norm": 3.0698599815368652, |
|
"learning_rate": 4.5035031079602445e-06, |
|
"loss": 0.8079, |
|
"step": 7860 |
|
}, |
|
{ |
|
"epoch": 2.44, |
|
"grad_norm": 2.567479133605957, |
|
"learning_rate": 4.2306177613277765e-06, |
|
"loss": 0.8262, |
|
"step": 7920 |
|
}, |
|
{ |
|
"epoch": 2.46, |
|
"grad_norm": 2.587759017944336, |
|
"learning_rate": 3.9654949406809995e-06, |
|
"loss": 0.8223, |
|
"step": 7980 |
|
}, |
|
{ |
|
"epoch": 2.47, |
|
"grad_norm": 2.034844398498535, |
|
"learning_rate": 3.7082337352831923e-06, |
|
"loss": 0.8114, |
|
"step": 8040 |
|
}, |
|
{ |
|
"epoch": 2.49, |
|
"grad_norm": 3.257702589035034, |
|
"learning_rate": 3.458930296130519e-06, |
|
"loss": 0.8267, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 2.51, |
|
"grad_norm": 2.539806365966797, |
|
"learning_rate": 3.2176778000157367e-06, |
|
"loss": 0.8251, |
|
"step": 8160 |
|
}, |
|
{ |
|
"epoch": 2.53, |
|
"grad_norm": 2.7473506927490234, |
|
"learning_rate": 2.9845664147035326e-06, |
|
"loss": 0.7826, |
|
"step": 8220 |
|
}, |
|
{ |
|
"epoch": 2.55, |
|
"grad_norm": 2.324519395828247, |
|
"learning_rate": 2.7596832652304283e-06, |
|
"loss": 0.8178, |
|
"step": 8280 |
|
}, |
|
{ |
|
"epoch": 2.57, |
|
"grad_norm": 2.691235065460205, |
|
"learning_rate": 2.5431124013419237e-06, |
|
"loss": 0.7982, |
|
"step": 8340 |
|
}, |
|
{ |
|
"epoch": 2.58, |
|
"grad_norm": 3.084143877029419, |
|
"learning_rate": 2.338335189634505e-06, |
|
"loss": 0.8007, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 2.6, |
|
"grad_norm": 2.8486242294311523, |
|
"learning_rate": 2.138486784704746e-06, |
|
"loss": 0.8192, |
|
"step": 8460 |
|
}, |
|
{ |
|
"epoch": 2.62, |
|
"grad_norm": 2.290395975112915, |
|
"learning_rate": 1.947182836628192e-06, |
|
"loss": 0.8303, |
|
"step": 8520 |
|
}, |
|
{ |
|
"epoch": 2.64, |
|
"grad_norm": 2.120853900909424, |
|
"learning_rate": 1.7644948449733384e-06, |
|
"loss": 0.8317, |
|
"step": 8580 |
|
}, |
|
{ |
|
"epoch": 2.66, |
|
"grad_norm": 2.403571605682373, |
|
"learning_rate": 1.590491089107679e-06, |
|
"loss": 0.8012, |
|
"step": 8640 |
|
}, |
|
{ |
|
"epoch": 2.68, |
|
"grad_norm": 1.6690127849578857, |
|
"learning_rate": 1.425236602678387e-06, |
|
"loss": 0.8241, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 2.7, |
|
"grad_norm": 3.3557851314544678, |
|
"learning_rate": 1.268793149306091e-06, |
|
"loss": 0.8234, |
|
"step": 8760 |
|
}, |
|
{ |
|
"epoch": 2.71, |
|
"grad_norm": 3.1359851360321045, |
|
"learning_rate": 1.1212191995007975e-06, |
|
"loss": 0.802, |
|
"step": 8820 |
|
}, |
|
{ |
|
"epoch": 2.73, |
|
"grad_norm": 2.1641125679016113, |
|
"learning_rate": 9.825699088086338e-07, |
|
"loss": 0.8138, |
|
"step": 8880 |
|
}, |
|
{ |
|
"epoch": 2.75, |
|
"grad_norm": 2.6551270484924316, |
|
"learning_rate": 8.528970971975553e-07, |
|
"loss": 0.8375, |
|
"step": 8940 |
|
}, |
|
{ |
|
"epoch": 2.77, |
|
"grad_norm": 1.8573861122131348, |
|
"learning_rate": 7.322492296896799e-07, |
|
"loss": 0.8166, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 2.79, |
|
"grad_norm": 2.5303752422332764, |
|
"learning_rate": 6.206713982475909e-07, |
|
"loss": 0.8346, |
|
"step": 9060 |
|
}, |
|
{ |
|
"epoch": 2.81, |
|
"grad_norm": 2.4010512828826904, |
|
"learning_rate": 5.182053049212626e-07, |
|
"loss": 0.8077, |
|
"step": 9120 |
|
}, |
|
{ |
|
"epoch": 2.82, |
|
"grad_norm": 2.183361053466797, |
|
"learning_rate": 4.248892462619725e-07, |
|
"loss": 0.799, |
|
"step": 9180 |
|
}, |
|
{ |
|
"epoch": 2.84, |
|
"grad_norm": 2.452626943588257, |
|
"learning_rate": 3.4075809900904756e-07, |
|
"loss": 0.8143, |
|
"step": 9240 |
|
}, |
|
{ |
|
"epoch": 2.86, |
|
"grad_norm": 2.959763526916504, |
|
"learning_rate": 2.658433070547195e-07, |
|
"loss": 0.8163, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 2.88, |
|
"grad_norm": 2.3562393188476562, |
|
"learning_rate": 2.0017286969200578e-07, |
|
"loss": 0.8134, |
|
"step": 9360 |
|
}, |
|
{ |
|
"epoch": 2.9, |
|
"grad_norm": 1.8572157621383667, |
|
"learning_rate": 1.4377133115004438e-07, |
|
"loss": 0.8188, |
|
"step": 9420 |
|
}, |
|
{ |
|
"epoch": 2.92, |
|
"grad_norm": 3.5713188648223877, |
|
"learning_rate": 9.665977142068738e-08, |
|
"loss": 0.8248, |
|
"step": 9480 |
|
}, |
|
{ |
|
"epoch": 2.94, |
|
"grad_norm": 2.22605562210083, |
|
"learning_rate": 5.885579837992261e-08, |
|
"loss": 0.7936, |
|
"step": 9540 |
|
}, |
|
{ |
|
"epoch": 2.95, |
|
"grad_norm": 2.264847993850708, |
|
"learning_rate": 3.037354120692393e-08, |
|
"loss": 0.8273, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 2.97, |
|
"grad_norm": 2.1906256675720215, |
|
"learning_rate": 1.1223645103311531e-08, |
|
"loss": 0.8199, |
|
"step": 9660 |
|
}, |
|
{ |
|
"epoch": 2.99, |
|
"grad_norm": 1.983068823814392, |
|
"learning_rate": 1.413267314517852e-09, |
|
"loss": 0.813, |
|
"step": 9720 |
|
}, |
|
{ |
|
"epoch": 3.0, |
|
"step": 9750, |
|
"total_flos": 1.6697107709546988e+18, |
|
"train_loss": 0.9979035949707031, |
|
"train_runtime": 4718.1281, |
|
"train_samples_per_second": 33.065, |
|
"train_steps_per_second": 2.066 |
|
} |
|
], |
|
"logging_steps": 60, |
|
"max_steps": 9750, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 3, |
|
"save_steps": 5000, |
|
"total_flos": 1.6697107709546988e+18, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|