| { |
| "best_metric": null, |
| "best_model_checkpoint": null, |
| "epoch": 2.0, |
| "eval_steps": 500, |
| "global_step": 118, |
| "is_hyper_param_search": false, |
| "is_local_process_zero": true, |
| "is_world_process_zero": true, |
| "log_history": [ |
| { |
| "epoch": 0.01694915254237288, |
| "grad_norm": 13.582984428143623, |
| "learning_rate": 8.333333333333333e-07, |
| "loss": 4.6947, |
| "step": 1 |
| }, |
| { |
| "epoch": 0.03389830508474576, |
| "grad_norm": 14.123407176783754, |
| "learning_rate": 1.6666666666666667e-06, |
| "loss": 4.7048, |
| "step": 2 |
| }, |
| { |
| "epoch": 0.05084745762711865, |
| "grad_norm": 14.538515649925046, |
| "learning_rate": 2.5e-06, |
| "loss": 4.4774, |
| "step": 3 |
| }, |
| { |
| "epoch": 0.06779661016949153, |
| "grad_norm": 12.917201583651316, |
| "learning_rate": 3.3333333333333333e-06, |
| "loss": 4.5093, |
| "step": 4 |
| }, |
| { |
| "epoch": 0.0847457627118644, |
| "grad_norm": 12.878282813705553, |
| "learning_rate": 4.166666666666667e-06, |
| "loss": 4.5946, |
| "step": 5 |
| }, |
| { |
| "epoch": 0.1016949152542373, |
| "grad_norm": 13.424010409794782, |
| "learning_rate": 5e-06, |
| "loss": 4.6969, |
| "step": 6 |
| }, |
| { |
| "epoch": 0.11864406779661017, |
| "grad_norm": 12.247691943280598, |
| "learning_rate": 5.833333333333334e-06, |
| "loss": 4.4876, |
| "step": 7 |
| }, |
| { |
| "epoch": 0.13559322033898305, |
| "grad_norm": 13.219835780776348, |
| "learning_rate": 6.666666666666667e-06, |
| "loss": 4.551, |
| "step": 8 |
| }, |
| { |
| "epoch": 0.15254237288135594, |
| "grad_norm": 13.377987371106677, |
| "learning_rate": 7.500000000000001e-06, |
| "loss": 4.1263, |
| "step": 9 |
| }, |
| { |
| "epoch": 0.1694915254237288, |
| "grad_norm": 12.792383156845725, |
| "learning_rate": 8.333333333333334e-06, |
| "loss": 4.1093, |
| "step": 10 |
| }, |
| { |
| "epoch": 0.1864406779661017, |
| "grad_norm": 11.58927724154859, |
| "learning_rate": 9.166666666666666e-06, |
| "loss": 4.0903, |
| "step": 11 |
| }, |
| { |
| "epoch": 0.2033898305084746, |
| "grad_norm": 12.857260545779804, |
| "learning_rate": 1e-05, |
| "loss": 3.7907, |
| "step": 12 |
| }, |
| { |
| "epoch": 0.22033898305084745, |
| "grad_norm": 11.263428499431882, |
| "learning_rate": 9.997804182543973e-06, |
| "loss": 3.4157, |
| "step": 13 |
| }, |
| { |
| "epoch": 0.23728813559322035, |
| "grad_norm": 9.096821276875527, |
| "learning_rate": 9.991218658821609e-06, |
| "loss": 3.248, |
| "step": 14 |
| }, |
| { |
| "epoch": 0.2542372881355932, |
| "grad_norm": 8.683777340261885, |
| "learning_rate": 9.980249213076085e-06, |
| "loss": 3.0936, |
| "step": 15 |
| }, |
| { |
| "epoch": 0.2711864406779661, |
| "grad_norm": 7.426067753465437, |
| "learning_rate": 9.964905480067585e-06, |
| "loss": 2.9003, |
| "step": 16 |
| }, |
| { |
| "epoch": 0.288135593220339, |
| "grad_norm": 6.801819161049987, |
| "learning_rate": 9.945200936610821e-06, |
| "loss": 2.7429, |
| "step": 17 |
| }, |
| { |
| "epoch": 0.3050847457627119, |
| "grad_norm": 5.517891034162887, |
| "learning_rate": 9.921152889737985e-06, |
| "loss": 2.4546, |
| "step": 18 |
| }, |
| { |
| "epoch": 0.3220338983050847, |
| "grad_norm": 4.607884862007527, |
| "learning_rate": 9.892782461497521e-06, |
| "loss": 2.2101, |
| "step": 19 |
| }, |
| { |
| "epoch": 0.3389830508474576, |
| "grad_norm": 4.198018015086853, |
| "learning_rate": 9.860114570402055e-06, |
| "loss": 2.002, |
| "step": 20 |
| }, |
| { |
| "epoch": 0.3559322033898305, |
| "grad_norm": 3.8937197111619337, |
| "learning_rate": 9.823177909541795e-06, |
| "loss": 1.9571, |
| "step": 21 |
| }, |
| { |
| "epoch": 0.3728813559322034, |
| "grad_norm": 3.116953631354838, |
| "learning_rate": 9.782004921382612e-06, |
| "loss": 1.7892, |
| "step": 22 |
| }, |
| { |
| "epoch": 0.3898305084745763, |
| "grad_norm": 2.88170444003024, |
| "learning_rate": 9.736631769270958e-06, |
| "loss": 1.8111, |
| "step": 23 |
| }, |
| { |
| "epoch": 0.4067796610169492, |
| "grad_norm": 2.360981852701721, |
| "learning_rate": 9.687098305670606e-06, |
| "loss": 1.6501, |
| "step": 24 |
| }, |
| { |
| "epoch": 0.423728813559322, |
| "grad_norm": 1.9535896880852222, |
| "learning_rate": 9.633448037159167e-06, |
| "loss": 1.4975, |
| "step": 25 |
| }, |
| { |
| "epoch": 0.4406779661016949, |
| "grad_norm": 1.742928903020929, |
| "learning_rate": 9.575728086215093e-06, |
| "loss": 1.6203, |
| "step": 26 |
| }, |
| { |
| "epoch": 0.4576271186440678, |
| "grad_norm": 1.6631917739747308, |
| "learning_rate": 9.513989149828718e-06, |
| "loss": 1.5796, |
| "step": 27 |
| }, |
| { |
| "epoch": 0.4745762711864407, |
| "grad_norm": 1.1808991053273623, |
| "learning_rate": 9.448285454973739e-06, |
| "loss": 1.4377, |
| "step": 28 |
| }, |
| { |
| "epoch": 0.4915254237288136, |
| "grad_norm": 1.0227483418783903, |
| "learning_rate": 9.378674710978185e-06, |
| "loss": 1.4009, |
| "step": 29 |
| }, |
| { |
| "epoch": 0.5084745762711864, |
| "grad_norm": 0.9157571127323277, |
| "learning_rate": 9.305218058836778e-06, |
| "loss": 1.2478, |
| "step": 30 |
| }, |
| { |
| "epoch": 0.5254237288135594, |
| "grad_norm": 0.9338477321471075, |
| "learning_rate": 9.22798001750913e-06, |
| "loss": 1.3688, |
| "step": 31 |
| }, |
| { |
| "epoch": 0.5423728813559322, |
| "grad_norm": 0.8796568592495111, |
| "learning_rate": 9.14702842725101e-06, |
| "loss": 1.2535, |
| "step": 32 |
| }, |
| { |
| "epoch": 0.559322033898305, |
| "grad_norm": 0.9219993618943527, |
| "learning_rate": 9.062434390028407e-06, |
| "loss": 1.2974, |
| "step": 33 |
| }, |
| { |
| "epoch": 0.576271186440678, |
| "grad_norm": 0.9825650845819016, |
| "learning_rate": 8.974272207066767e-06, |
| "loss": 1.3383, |
| "step": 34 |
| }, |
| { |
| "epoch": 0.5932203389830508, |
| "grad_norm": 0.9055827980183389, |
| "learning_rate": 8.882619313590212e-06, |
| "loss": 1.2601, |
| "step": 35 |
| }, |
| { |
| "epoch": 0.6101694915254238, |
| "grad_norm": 0.9167811987216382, |
| "learning_rate": 8.787556210808101e-06, |
| "loss": 1.226, |
| "step": 36 |
| }, |
| { |
| "epoch": 0.6271186440677966, |
| "grad_norm": 0.8488058692801289, |
| "learning_rate": 8.689166395208638e-06, |
| "loss": 1.1668, |
| "step": 37 |
| }, |
| { |
| "epoch": 0.6440677966101694, |
| "grad_norm": 0.8462653312925347, |
| "learning_rate": 8.587536285221656e-06, |
| "loss": 1.2131, |
| "step": 38 |
| }, |
| { |
| "epoch": 0.6610169491525424, |
| "grad_norm": 0.741157564370513, |
| "learning_rate": 8.482755145314987e-06, |
| "loss": 1.1322, |
| "step": 39 |
| }, |
| { |
| "epoch": 0.6779661016949152, |
| "grad_norm": 0.7694007284550851, |
| "learning_rate": 8.374915007591053e-06, |
| "loss": 1.1526, |
| "step": 40 |
| }, |
| { |
| "epoch": 0.6949152542372882, |
| "grad_norm": 0.7658488631154642, |
| "learning_rate": 8.264110590952609e-06, |
| "loss": 1.1856, |
| "step": 41 |
| }, |
| { |
| "epoch": 0.711864406779661, |
| "grad_norm": 0.6405702241168747, |
| "learning_rate": 8.150439217908557e-06, |
| "loss": 1.0037, |
| "step": 42 |
| }, |
| { |
| "epoch": 0.7288135593220338, |
| "grad_norm": 0.616014862055841, |
| "learning_rate": 8.034000729092967e-06, |
| "loss": 1.0329, |
| "step": 43 |
| }, |
| { |
| "epoch": 0.7457627118644068, |
| "grad_norm": 0.6578920077905241, |
| "learning_rate": 7.914897395572362e-06, |
| "loss": 1.0895, |
| "step": 44 |
| }, |
| { |
| "epoch": 0.7627118644067796, |
| "grad_norm": 0.5912450960157642, |
| "learning_rate": 7.793233829018263e-06, |
| "loss": 1.0218, |
| "step": 45 |
| }, |
| { |
| "epoch": 0.7796610169491526, |
| "grad_norm": 0.5582880461976706, |
| "learning_rate": 7.669116889823955e-06, |
| "loss": 1.013, |
| "step": 46 |
| }, |
| { |
| "epoch": 0.7966101694915254, |
| "grad_norm": 0.524466989613803, |
| "learning_rate": 7.542655593246103e-06, |
| "loss": 0.9566, |
| "step": 47 |
| }, |
| { |
| "epoch": 0.8135593220338984, |
| "grad_norm": 0.48335542595907816, |
| "learning_rate": 7.413961013653725e-06, |
| "loss": 0.8977, |
| "step": 48 |
| }, |
| { |
| "epoch": 0.8305084745762712, |
| "grad_norm": 0.5365637758122486, |
| "learning_rate": 7.283146186968566e-06, |
| "loss": 1.0325, |
| "step": 49 |
| }, |
| { |
| "epoch": 0.847457627118644, |
| "grad_norm": 0.5099489526642094, |
| "learning_rate": 7.1503260113826035e-06, |
| "loss": 1.0182, |
| "step": 50 |
| }, |
| { |
| "epoch": 0.864406779661017, |
| "grad_norm": 0.4657708339456314, |
| "learning_rate": 7.015617146439863e-06, |
| "loss": 0.9675, |
| "step": 51 |
| }, |
| { |
| "epoch": 0.8813559322033898, |
| "grad_norm": 0.48914070170734636, |
| "learning_rate": 6.879137910571191e-06, |
| "loss": 0.9954, |
| "step": 52 |
| }, |
| { |
| "epoch": 0.8983050847457628, |
| "grad_norm": 0.46882412163559467, |
| "learning_rate": 6.741008177171995e-06, |
| "loss": 0.9976, |
| "step": 53 |
| }, |
| { |
| "epoch": 0.9152542372881356, |
| "grad_norm": 0.3936331113504765, |
| "learning_rate": 6.601349269314188e-06, |
| "loss": 0.8422, |
| "step": 54 |
| }, |
| { |
| "epoch": 0.9322033898305084, |
| "grad_norm": 0.437572210845862, |
| "learning_rate": 6.46028385318488e-06, |
| "loss": 0.9785, |
| "step": 55 |
| }, |
| { |
| "epoch": 0.9491525423728814, |
| "grad_norm": 0.42239632796928905, |
| "learning_rate": 6.3179358303453386e-06, |
| "loss": 0.9773, |
| "step": 56 |
| }, |
| { |
| "epoch": 0.9661016949152542, |
| "grad_norm": 0.3849602768372098, |
| "learning_rate": 6.17443022890492e-06, |
| "loss": 0.9132, |
| "step": 57 |
| }, |
| { |
| "epoch": 0.9830508474576272, |
| "grad_norm": 0.35887608129604537, |
| "learning_rate": 6.029893093705492e-06, |
| "loss": 0.8932, |
| "step": 58 |
| }, |
| { |
| "epoch": 1.0, |
| "grad_norm": 0.3775382679783028, |
| "learning_rate": 5.884451375612865e-06, |
| "loss": 0.9605, |
| "step": 59 |
| }, |
| { |
| "epoch": 1.0169491525423728, |
| "grad_norm": 0.35775761239990117, |
| "learning_rate": 5.738232820012407e-06, |
| "loss": 0.9254, |
| "step": 60 |
| }, |
| { |
| "epoch": 1.0338983050847457, |
| "grad_norm": 0.34923814072726866, |
| "learning_rate": 5.591365854606829e-06, |
| "loss": 0.9174, |
| "step": 61 |
| }, |
| { |
| "epoch": 1.0508474576271187, |
| "grad_norm": 0.34017872341374605, |
| "learning_rate": 5.443979476614674e-06, |
| "loss": 0.9484, |
| "step": 62 |
| }, |
| { |
| "epoch": 1.0677966101694916, |
| "grad_norm": 0.3089243172711035, |
| "learning_rate": 5.296203139468572e-06, |
| "loss": 0.8374, |
| "step": 63 |
| }, |
| { |
| "epoch": 1.0847457627118644, |
| "grad_norm": 0.2871261805872519, |
| "learning_rate": 5.148166639112799e-06, |
| "loss": 0.8462, |
| "step": 64 |
| }, |
| { |
| "epoch": 1.1016949152542372, |
| "grad_norm": 0.30706922907791745, |
| "learning_rate": 5e-06, |
| "loss": 0.8865, |
| "step": 65 |
| }, |
| { |
| "epoch": 1.11864406779661, |
| "grad_norm": 0.2938750275036386, |
| "learning_rate": 4.8518333608872015e-06, |
| "loss": 0.8858, |
| "step": 66 |
| }, |
| { |
| "epoch": 1.1355932203389831, |
| "grad_norm": 0.2549040934942167, |
| "learning_rate": 4.703796860531429e-06, |
| "loss": 0.7952, |
| "step": 67 |
| }, |
| { |
| "epoch": 1.152542372881356, |
| "grad_norm": 0.28685443691965085, |
| "learning_rate": 4.556020523385326e-06, |
| "loss": 0.8437, |
| "step": 68 |
| }, |
| { |
| "epoch": 1.1694915254237288, |
| "grad_norm": 0.2867782597037466, |
| "learning_rate": 4.408634145393172e-06, |
| "loss": 0.8934, |
| "step": 69 |
| }, |
| { |
| "epoch": 1.1864406779661016, |
| "grad_norm": 0.31047696033713723, |
| "learning_rate": 4.261767179987595e-06, |
| "loss": 0.9716, |
| "step": 70 |
| }, |
| { |
| "epoch": 1.2033898305084745, |
| "grad_norm": 0.25941316846282275, |
| "learning_rate": 4.115548624387136e-06, |
| "loss": 0.8287, |
| "step": 71 |
| }, |
| { |
| "epoch": 1.2203389830508475, |
| "grad_norm": 0.2699492319649961, |
| "learning_rate": 3.970106906294509e-06, |
| "loss": 0.8472, |
| "step": 72 |
| }, |
| { |
| "epoch": 1.2372881355932204, |
| "grad_norm": 0.25234340024796287, |
| "learning_rate": 3.825569771095082e-06, |
| "loss": 0.8287, |
| "step": 73 |
| }, |
| { |
| "epoch": 1.2542372881355932, |
| "grad_norm": 0.2660601540051665, |
| "learning_rate": 3.682064169654663e-06, |
| "loss": 0.8793, |
| "step": 74 |
| }, |
| { |
| "epoch": 1.271186440677966, |
| "grad_norm": 0.242634531727012, |
| "learning_rate": 3.539716146815122e-06, |
| "loss": 0.7988, |
| "step": 75 |
| }, |
| { |
| "epoch": 1.288135593220339, |
| "grad_norm": 0.2465357214488806, |
| "learning_rate": 3.398650730685813e-06, |
| "loss": 0.8259, |
| "step": 76 |
| }, |
| { |
| "epoch": 1.305084745762712, |
| "grad_norm": 0.22575413510974532, |
| "learning_rate": 3.258991822828007e-06, |
| "loss": 0.7525, |
| "step": 77 |
| }, |
| { |
| "epoch": 1.3220338983050848, |
| "grad_norm": 0.23693526616000352, |
| "learning_rate": 3.1208620894288105e-06, |
| "loss": 0.8184, |
| "step": 78 |
| }, |
| { |
| "epoch": 1.3389830508474576, |
| "grad_norm": 0.23000459251399064, |
| "learning_rate": 2.98438285356014e-06, |
| "loss": 0.7968, |
| "step": 79 |
| }, |
| { |
| "epoch": 1.3559322033898304, |
| "grad_norm": 0.1958693615856507, |
| "learning_rate": 2.8496739886173994e-06, |
| "loss": 0.6473, |
| "step": 80 |
| }, |
| { |
| "epoch": 1.3728813559322033, |
| "grad_norm": 0.24742346042002725, |
| "learning_rate": 2.716853813031435e-06, |
| "loss": 0.8747, |
| "step": 81 |
| }, |
| { |
| "epoch": 1.3898305084745763, |
| "grad_norm": 0.2506984767340435, |
| "learning_rate": 2.5860389863462765e-06, |
| "loss": 0.8807, |
| "step": 82 |
| }, |
| { |
| "epoch": 1.4067796610169492, |
| "grad_norm": 0.22963412267254538, |
| "learning_rate": 2.457344406753899e-06, |
| "loss": 0.8101, |
| "step": 83 |
| }, |
| { |
| "epoch": 1.423728813559322, |
| "grad_norm": 0.2317036951827151, |
| "learning_rate": 2.330883110176049e-06, |
| "loss": 0.8256, |
| "step": 84 |
| }, |
| { |
| "epoch": 1.4406779661016949, |
| "grad_norm": 0.19606132470172136, |
| "learning_rate": 2.2067661709817384e-06, |
| "loss": 0.6938, |
| "step": 85 |
| }, |
| { |
| "epoch": 1.457627118644068, |
| "grad_norm": 0.219441511162039, |
| "learning_rate": 2.0851026044276405e-06, |
| "loss": 0.7687, |
| "step": 86 |
| }, |
| { |
| "epoch": 1.4745762711864407, |
| "grad_norm": 0.20310562062320556, |
| "learning_rate": 1.9659992709070346e-06, |
| "loss": 0.76, |
| "step": 87 |
| }, |
| { |
| "epoch": 1.4915254237288136, |
| "grad_norm": 0.21286522814139364, |
| "learning_rate": 1.8495607820914451e-06, |
| "loss": 0.7625, |
| "step": 88 |
| }, |
| { |
| "epoch": 1.5084745762711864, |
| "grad_norm": 0.22107972808860454, |
| "learning_rate": 1.7358894090473928e-06, |
| "loss": 0.8307, |
| "step": 89 |
| }, |
| { |
| "epoch": 1.5254237288135593, |
| "grad_norm": 0.21471387542490836, |
| "learning_rate": 1.6250849924089485e-06, |
| "loss": 0.7501, |
| "step": 90 |
| }, |
| { |
| "epoch": 1.542372881355932, |
| "grad_norm": 0.22047778238568697, |
| "learning_rate": 1.5172448546850166e-06, |
| "loss": 0.8155, |
| "step": 91 |
| }, |
| { |
| "epoch": 1.559322033898305, |
| "grad_norm": 0.21411524032874296, |
| "learning_rate": 1.4124637147783431e-06, |
| "loss": 0.7937, |
| "step": 92 |
| }, |
| { |
| "epoch": 1.576271186440678, |
| "grad_norm": 0.21327217337137594, |
| "learning_rate": 1.3108336047913633e-06, |
| "loss": 0.7728, |
| "step": 93 |
| }, |
| { |
| "epoch": 1.5932203389830508, |
| "grad_norm": 0.1920350890527698, |
| "learning_rate": 1.2124437891918995e-06, |
| "loss": 0.6718, |
| "step": 94 |
| }, |
| { |
| "epoch": 1.6101694915254239, |
| "grad_norm": 0.21902635675798737, |
| "learning_rate": 1.1173806864097885e-06, |
| "loss": 0.804, |
| "step": 95 |
| }, |
| { |
| "epoch": 1.6271186440677967, |
| "grad_norm": 0.22207901313261283, |
| "learning_rate": 1.0257277929332332e-06, |
| "loss": 0.8154, |
| "step": 96 |
| }, |
| { |
| "epoch": 1.6440677966101696, |
| "grad_norm": 0.21571048693428027, |
| "learning_rate": 9.375656099715935e-07, |
| "loss": 0.7889, |
| "step": 97 |
| }, |
| { |
| "epoch": 1.6610169491525424, |
| "grad_norm": 0.21357881995728437, |
| "learning_rate": 8.529715727489912e-07, |
| "loss": 0.8078, |
| "step": 98 |
| }, |
| { |
| "epoch": 1.6779661016949152, |
| "grad_norm": 0.2235192978949996, |
| "learning_rate": 7.720199824908692e-07, |
| "loss": 0.8096, |
| "step": 99 |
| }, |
| { |
| "epoch": 1.694915254237288, |
| "grad_norm": 0.22622640413844738, |
| "learning_rate": 6.947819411632223e-07, |
| "loss": 0.8223, |
| "step": 100 |
| }, |
| { |
| "epoch": 1.711864406779661, |
| "grad_norm": 0.22601175493797776, |
| "learning_rate": 6.213252890218163e-07, |
| "loss": 0.8135, |
| "step": 101 |
| }, |
| { |
| "epoch": 1.7288135593220337, |
| "grad_norm": 0.1911211920002579, |
| "learning_rate": 5.517145450262639e-07, |
| "loss": 0.7068, |
| "step": 102 |
| }, |
| { |
| "epoch": 1.7457627118644068, |
| "grad_norm": 0.21175848010983564, |
| "learning_rate": 4.860108501712824e-07, |
| "loss": 0.7783, |
| "step": 103 |
| }, |
| { |
| "epoch": 1.7627118644067796, |
| "grad_norm": 0.21053330888859195, |
| "learning_rate": 4.242719137849077e-07, |
| "loss": 0.8187, |
| "step": 104 |
| }, |
| { |
| "epoch": 1.7796610169491527, |
| "grad_norm": 0.2130541935136037, |
| "learning_rate": 3.665519628408332e-07, |
| "loss": 0.7656, |
| "step": 105 |
| }, |
| { |
| "epoch": 1.7966101694915255, |
| "grad_norm": 0.22526303577323026, |
| "learning_rate": 3.1290169432939556e-07, |
| "loss": 0.8245, |
| "step": 106 |
| }, |
| { |
| "epoch": 1.8135593220338984, |
| "grad_norm": 0.2072005309322467, |
| "learning_rate": 2.6336823072904305e-07, |
| "loss": 0.743, |
| "step": 107 |
| }, |
| { |
| "epoch": 1.8305084745762712, |
| "grad_norm": 0.22448423254493008, |
| "learning_rate": 2.179950786173879e-07, |
| "loss": 0.7898, |
| "step": 108 |
| }, |
| { |
| "epoch": 1.847457627118644, |
| "grad_norm": 0.2138555012447717, |
| "learning_rate": 1.7682209045820687e-07, |
| "loss": 0.7782, |
| "step": 109 |
| }, |
| { |
| "epoch": 1.8644067796610169, |
| "grad_norm": 0.1993249828378952, |
| "learning_rate": 1.3988542959794627e-07, |
| "loss": 0.7393, |
| "step": 110 |
| }, |
| { |
| "epoch": 1.8813559322033897, |
| "grad_norm": 0.21708860295387616, |
| "learning_rate": 1.0721753850247984e-07, |
| "loss": 0.7906, |
| "step": 111 |
| }, |
| { |
| "epoch": 1.8983050847457628, |
| "grad_norm": 0.18874363435593658, |
| "learning_rate": 7.884711026201586e-08, |
| "loss": 0.6812, |
| "step": 112 |
| }, |
| { |
| "epoch": 1.9152542372881356, |
| "grad_norm": 0.25545899012578377, |
| "learning_rate": 5.479906338917984e-08, |
| "loss": 0.9486, |
| "step": 113 |
| }, |
| { |
| "epoch": 1.9322033898305084, |
| "grad_norm": 0.1936498628431566, |
| "learning_rate": 3.5094519932415417e-08, |
| "loss": 0.745, |
| "step": 114 |
| }, |
| { |
| "epoch": 1.9491525423728815, |
| "grad_norm": 0.20802679119960885, |
| "learning_rate": 1.975078692391552e-08, |
| "loss": 0.7714, |
| "step": 115 |
| }, |
| { |
| "epoch": 1.9661016949152543, |
| "grad_norm": 0.22619971076796636, |
| "learning_rate": 8.781341178393244e-09, |
| "loss": 0.8194, |
| "step": 116 |
| }, |
| { |
| "epoch": 1.9830508474576272, |
| "grad_norm": 0.2211171658614822, |
| "learning_rate": 2.19581745602826e-09, |
| "loss": 0.7564, |
| "step": 117 |
| }, |
| { |
| "epoch": 2.0, |
| "grad_norm": 0.22426334671251008, |
| "learning_rate": 0.0, |
| "loss": 0.8302, |
| "step": 118 |
| }, |
| { |
| "epoch": 2.0, |
| "step": 118, |
| "total_flos": 1854342635192320.0, |
| "train_loss": 1.4390717199293233, |
| "train_runtime": 1287.2643, |
| "train_samples_per_second": 5.867, |
| "train_steps_per_second": 0.092 |
| } |
| ], |
| "logging_steps": 1, |
| "max_steps": 118, |
| "num_input_tokens_seen": 0, |
| "num_train_epochs": 2, |
| "save_steps": 500, |
| "stateful_callbacks": { |
| "TrainerControl": { |
| "args": { |
| "should_epoch_stop": false, |
| "should_evaluate": false, |
| "should_log": false, |
| "should_save": true, |
| "should_training_stop": true |
| }, |
| "attributes": {} |
| } |
| }, |
| "total_flos": 1854342635192320.0, |
| "train_batch_size": 8, |
| "trial_name": null, |
| "trial_params": null |
| } |
|
|