{ "best_metric": null, "best_model_checkpoint": null, "epoch": 19.995409685563462, "eval_steps": 500, "global_step": 21780, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.09180628873077806, "grad_norm": 0.229265034198761, "learning_rate": 0.0003, "loss": 1.6148, "step": 100 }, { "epoch": 0.18361257746155613, "grad_norm": 0.2753294110298157, "learning_rate": 0.0003, "loss": 1.5846, "step": 200 }, { "epoch": 0.2754188661923342, "grad_norm": 0.2562897801399231, "learning_rate": 0.0003, "loss": 1.5631, "step": 300 }, { "epoch": 0.36722515492311225, "grad_norm": 0.24621829390525818, "learning_rate": 0.0003, "loss": 1.5662, "step": 400 }, { "epoch": 0.4590314436538903, "grad_norm": 0.2583877742290497, "learning_rate": 0.0003, "loss": 1.561, "step": 500 }, { "epoch": 0.5508377323846684, "grad_norm": 0.275136262178421, "learning_rate": 0.0003, "loss": 1.5561, "step": 600 }, { "epoch": 0.6426440211154464, "grad_norm": 0.2927466928958893, "learning_rate": 0.0003, "loss": 1.5552, "step": 700 }, { "epoch": 0.7344503098462245, "grad_norm": 0.3323851525783539, "learning_rate": 0.0003, "loss": 1.5608, "step": 800 }, { "epoch": 0.8262565985770025, "grad_norm": 0.33202576637268066, "learning_rate": 0.0003, "loss": 1.5254, "step": 900 }, { "epoch": 0.9180628873077806, "grad_norm": 0.3486601412296295, "learning_rate": 0.0003, "loss": 1.5356, "step": 1000 }, { "epoch": 0.999770484278173, "eval_accuracy": 0.6863755458515284, "eval_loss": 1.3710707426071167, "eval_runtime": 8.9317, "eval_samples_per_second": 55.98, "eval_steps_per_second": 7.054, "step": 1089 }, { "epoch": 1.0098691760385587, "grad_norm": 0.34695035219192505, "learning_rate": 0.0003, "loss": 1.5044, "step": 1100 }, { "epoch": 1.1016754647693368, "grad_norm": 0.3605581820011139, "learning_rate": 0.0003, "loss": 1.3328, "step": 1200 }, { "epoch": 1.1934817535001148, "grad_norm": 0.37978020310401917, "learning_rate": 0.0003, "loss": 1.3246, "step": 1300 }, { "epoch": 1.2852880422308928, "grad_norm": 0.40200668573379517, "learning_rate": 0.0003, "loss": 1.3183, "step": 1400 }, { "epoch": 1.377094330961671, "grad_norm": 0.49451494216918945, "learning_rate": 0.0003, "loss": 1.3153, "step": 1500 }, { "epoch": 1.468900619692449, "grad_norm": 0.4396137297153473, "learning_rate": 0.0003, "loss": 1.3301, "step": 1600 }, { "epoch": 1.560706908423227, "grad_norm": 0.4403189718723297, "learning_rate": 0.0003, "loss": 1.2971, "step": 1700 }, { "epoch": 1.652513197154005, "grad_norm": 0.4628995656967163, "learning_rate": 0.0003, "loss": 1.3159, "step": 1800 }, { "epoch": 1.744319485884783, "grad_norm": 0.4826644957065582, "learning_rate": 0.0003, "loss": 1.3027, "step": 1900 }, { "epoch": 1.836125774615561, "grad_norm": 0.48213547468185425, "learning_rate": 0.0003, "loss": 1.2977, "step": 2000 }, { "epoch": 1.9279320633463393, "grad_norm": 0.5228666067123413, "learning_rate": 0.0003, "loss": 1.3102, "step": 2100 }, { "epoch": 1.999540968556346, "eval_accuracy": 0.7020116448326056, "eval_loss": 1.1752803325653076, "eval_runtime": 8.9457, "eval_samples_per_second": 55.892, "eval_steps_per_second": 7.042, "step": 2178 }, { "epoch": 2.0197383520771175, "grad_norm": 0.4973185360431671, "learning_rate": 0.0003, "loss": 1.2434, "step": 2200 }, { "epoch": 2.1115446408078955, "grad_norm": 0.5200226902961731, "learning_rate": 0.0003, "loss": 1.0191, "step": 2300 }, { "epoch": 2.2033509295386735, "grad_norm": 0.6249756217002869, "learning_rate": 0.0003, "loss": 1.0435, "step": 2400 }, { "epoch": 2.2951572182694515, "grad_norm": 0.5348188877105713, "learning_rate": 0.0003, "loss": 1.0434, "step": 2500 }, { "epoch": 2.3869635070002295, "grad_norm": 0.6106544733047485, "learning_rate": 0.0003, "loss": 1.0546, "step": 2600 }, { "epoch": 2.4787697957310075, "grad_norm": 0.5895304679870605, "learning_rate": 0.0003, "loss": 1.0437, "step": 2700 }, { "epoch": 2.5705760844617855, "grad_norm": 0.5809823274612427, "learning_rate": 0.0003, "loss": 1.0628, "step": 2800 }, { "epoch": 2.6623823731925635, "grad_norm": 0.6009641289710999, "learning_rate": 0.0003, "loss": 1.0478, "step": 2900 }, { "epoch": 2.754188661923342, "grad_norm": 0.679328978061676, "learning_rate": 0.0003, "loss": 1.0674, "step": 3000 }, { "epoch": 2.84599495065412, "grad_norm": 0.5471652746200562, "learning_rate": 0.0003, "loss": 1.0462, "step": 3100 }, { "epoch": 2.937801239384898, "grad_norm": 0.6065019965171814, "learning_rate": 0.0003, "loss": 1.0549, "step": 3200 }, { "epoch": 2.9993114528345193, "eval_accuracy": 0.716448326055313, "eval_loss": 1.0094784498214722, "eval_runtime": 8.9265, "eval_samples_per_second": 56.013, "eval_steps_per_second": 7.058, "step": 3267 }, { "epoch": 3.029607528115676, "grad_norm": 0.6480036973953247, "learning_rate": 0.0003, "loss": 0.9648, "step": 3300 }, { "epoch": 3.121413816846454, "grad_norm": 0.7049514651298523, "learning_rate": 0.0003, "loss": 0.7919, "step": 3400 }, { "epoch": 3.213220105577232, "grad_norm": 0.7723010182380676, "learning_rate": 0.0003, "loss": 0.804, "step": 3500 }, { "epoch": 3.30502639430801, "grad_norm": 0.7994470596313477, "learning_rate": 0.0003, "loss": 0.8323, "step": 3600 }, { "epoch": 3.396832683038788, "grad_norm": 0.7567194700241089, "learning_rate": 0.0003, "loss": 0.8261, "step": 3700 }, { "epoch": 3.488638971769566, "grad_norm": 0.7832695245742798, "learning_rate": 0.0003, "loss": 0.8337, "step": 3800 }, { "epoch": 3.580445260500344, "grad_norm": 0.7570985555648804, "learning_rate": 0.0003, "loss": 0.8334, "step": 3900 }, { "epoch": 3.672251549231122, "grad_norm": 0.7717698216438293, "learning_rate": 0.0003, "loss": 0.8455, "step": 4000 }, { "epoch": 3.7640578379619005, "grad_norm": 0.8522344827651978, "learning_rate": 0.0003, "loss": 0.8448, "step": 4100 }, { "epoch": 3.8558641266926785, "grad_norm": 0.8296842575073242, "learning_rate": 0.0003, "loss": 0.8534, "step": 4200 }, { "epoch": 3.9476704154234565, "grad_norm": 0.7863412499427795, "learning_rate": 0.0003, "loss": 0.8461, "step": 4300 }, { "epoch": 4.0, "eval_accuracy": 0.7297409024745269, "eval_loss": 0.8722099661827087, "eval_runtime": 8.8823, "eval_samples_per_second": 56.292, "eval_steps_per_second": 7.093, "step": 4357 }, { "epoch": 4.039476704154235, "grad_norm": 0.7983632683753967, "learning_rate": 0.0003, "loss": 0.7547, "step": 4400 }, { "epoch": 4.131282992885013, "grad_norm": 0.699720561504364, "learning_rate": 0.0003, "loss": 0.618, "step": 4500 }, { "epoch": 4.223089281615791, "grad_norm": 0.8726249933242798, "learning_rate": 0.0003, "loss": 0.6373, "step": 4600 }, { "epoch": 4.314895570346569, "grad_norm": 0.8412553071975708, "learning_rate": 0.0003, "loss": 0.6582, "step": 4700 }, { "epoch": 4.406701859077347, "grad_norm": 0.86082923412323, "learning_rate": 0.0003, "loss": 0.6612, "step": 4800 }, { "epoch": 4.498508147808125, "grad_norm": 0.8449448943138123, "learning_rate": 0.0003, "loss": 0.6704, "step": 4900 }, { "epoch": 4.590314436538903, "grad_norm": 0.9901677370071411, "learning_rate": 0.0003, "loss": 0.683, "step": 5000 }, { "epoch": 4.682120725269681, "grad_norm": 0.884429395198822, "learning_rate": 0.0003, "loss": 0.6903, "step": 5100 }, { "epoch": 4.773927014000459, "grad_norm": 0.8438558578491211, "learning_rate": 0.0003, "loss": 0.693, "step": 5200 }, { "epoch": 4.865733302731237, "grad_norm": 0.8200265169143677, "learning_rate": 0.0003, "loss": 0.6986, "step": 5300 }, { "epoch": 4.957539591462015, "grad_norm": 0.9410732984542847, "learning_rate": 0.0003, "loss": 0.701, "step": 5400 }, { "epoch": 4.999770484278173, "eval_accuracy": 0.7405647743813683, "eval_loss": 0.7641452550888062, "eval_runtime": 8.8775, "eval_samples_per_second": 56.322, "eval_steps_per_second": 7.097, "step": 5446 }, { "epoch": 5.049345880192793, "grad_norm": 0.8364540934562683, "learning_rate": 0.0003, "loss": 0.5945, "step": 5500 }, { "epoch": 5.141152168923571, "grad_norm": 0.8985306620597839, "learning_rate": 0.0003, "loss": 0.5072, "step": 5600 }, { "epoch": 5.232958457654349, "grad_norm": 0.9036381244659424, "learning_rate": 0.0003, "loss": 0.5187, "step": 5700 }, { "epoch": 5.324764746385127, "grad_norm": 0.9836907386779785, "learning_rate": 0.0003, "loss": 0.5468, "step": 5800 }, { "epoch": 5.416571035115905, "grad_norm": 0.8932607769966125, "learning_rate": 0.0003, "loss": 0.546, "step": 5900 }, { "epoch": 5.508377323846684, "grad_norm": 0.9328716397285461, "learning_rate": 0.0003, "loss": 0.5636, "step": 6000 }, { "epoch": 5.600183612577462, "grad_norm": 1.0993613004684448, "learning_rate": 0.0003, "loss": 0.567, "step": 6100 }, { "epoch": 5.69198990130824, "grad_norm": 0.9644467830657959, "learning_rate": 0.0003, "loss": 0.5853, "step": 6200 }, { "epoch": 5.783796190039018, "grad_norm": 1.0126985311508179, "learning_rate": 0.0003, "loss": 0.5883, "step": 6300 }, { "epoch": 5.875602478769796, "grad_norm": 0.9119430184364319, "learning_rate": 0.0003, "loss": 0.5853, "step": 6400 }, { "epoch": 5.967408767500574, "grad_norm": 0.88072669506073, "learning_rate": 0.0003, "loss": 0.5977, "step": 6500 }, { "epoch": 5.999540968556346, "eval_accuracy": 0.7490451237263465, "eval_loss": 0.6796724796295166, "eval_runtime": 8.9191, "eval_samples_per_second": 56.059, "eval_steps_per_second": 7.063, "step": 6535 }, { "epoch": 6.059215056231352, "grad_norm": 0.9065115451812744, "learning_rate": 0.0003, "loss": 0.4864, "step": 6600 }, { "epoch": 6.15102134496213, "grad_norm": 0.8660620450973511, "learning_rate": 0.0003, "loss": 0.4337, "step": 6700 }, { "epoch": 6.242827633692908, "grad_norm": 0.9163237810134888, "learning_rate": 0.0003, "loss": 0.4496, "step": 6800 }, { "epoch": 6.334633922423686, "grad_norm": 0.8863268494606018, "learning_rate": 0.0003, "loss": 0.4626, "step": 6900 }, { "epoch": 6.426440211154464, "grad_norm": 0.9574721455574036, "learning_rate": 0.0003, "loss": 0.4722, "step": 7000 }, { "epoch": 6.518246499885242, "grad_norm": 1.096871256828308, "learning_rate": 0.0003, "loss": 0.4887, "step": 7100 }, { "epoch": 6.61005278861602, "grad_norm": 1.0900388956069946, "learning_rate": 0.0003, "loss": 0.4958, "step": 7200 }, { "epoch": 6.701859077346798, "grad_norm": 0.9678546190261841, "learning_rate": 0.0003, "loss": 0.4991, "step": 7300 }, { "epoch": 6.793665366077576, "grad_norm": 1.0782684087753296, "learning_rate": 0.0003, "loss": 0.5068, "step": 7400 }, { "epoch": 6.885471654808354, "grad_norm": 1.0213241577148438, "learning_rate": 0.0003, "loss": 0.512, "step": 7500 }, { "epoch": 6.977277943539132, "grad_norm": 1.049910545349121, "learning_rate": 0.0003, "loss": 0.5238, "step": 7600 }, { "epoch": 6.999311452834519, "eval_accuracy": 0.7558719068413392, "eval_loss": 0.620939314365387, "eval_runtime": 9.0834, "eval_samples_per_second": 55.045, "eval_steps_per_second": 6.936, "step": 7624 }, { "epoch": 7.06908423226991, "grad_norm": 0.9470316767692566, "learning_rate": 0.0003, "loss": 0.4131, "step": 7700 }, { "epoch": 7.160890521000688, "grad_norm": 0.8299586772918701, "learning_rate": 0.0003, "loss": 0.39, "step": 7800 }, { "epoch": 7.252696809731467, "grad_norm": 0.9369521141052246, "learning_rate": 0.0003, "loss": 0.3953, "step": 7900 }, { "epoch": 7.344503098462245, "grad_norm": 1.0843796730041504, "learning_rate": 0.0003, "loss": 0.4144, "step": 8000 }, { "epoch": 7.436309387193023, "grad_norm": 0.9615817666053772, "learning_rate": 0.0003, "loss": 0.4266, "step": 8100 }, { "epoch": 7.528115675923801, "grad_norm": 0.9017691612243652, "learning_rate": 0.0003, "loss": 0.4411, "step": 8200 }, { "epoch": 7.619921964654579, "grad_norm": 1.0230755805969238, "learning_rate": 0.0003, "loss": 0.4392, "step": 8300 }, { "epoch": 7.711728253385357, "grad_norm": 1.0185121297836304, "learning_rate": 0.0003, "loss": 0.4489, "step": 8400 }, { "epoch": 7.803534542116135, "grad_norm": 1.003201961517334, "learning_rate": 0.0003, "loss": 0.4475, "step": 8500 }, { "epoch": 7.895340830846913, "grad_norm": 0.8930547833442688, "learning_rate": 0.0003, "loss": 0.4569, "step": 8600 }, { "epoch": 7.987147119577691, "grad_norm": 1.1096137762069702, "learning_rate": 0.0003, "loss": 0.4742, "step": 8700 }, { "epoch": 8.0, "eval_accuracy": 0.7599941775836973, "eval_loss": 0.5836606025695801, "eval_runtime": 8.8891, "eval_samples_per_second": 56.249, "eval_steps_per_second": 7.087, "step": 8714 }, { "epoch": 8.07895340830847, "grad_norm": 0.9343357086181641, "learning_rate": 0.0003, "loss": 0.3623, "step": 8800 }, { "epoch": 8.170759697039248, "grad_norm": 1.0631523132324219, "learning_rate": 0.0003, "loss": 0.3456, "step": 8900 }, { "epoch": 8.262565985770026, "grad_norm": 0.9419318437576294, "learning_rate": 0.0003, "loss": 0.3648, "step": 9000 }, { "epoch": 8.354372274500804, "grad_norm": 0.8649080395698547, "learning_rate": 0.0003, "loss": 0.3758, "step": 9100 }, { "epoch": 8.446178563231582, "grad_norm": 0.9168979525566101, "learning_rate": 0.0003, "loss": 0.3843, "step": 9200 }, { "epoch": 8.53798485196236, "grad_norm": 0.9176032543182373, "learning_rate": 0.0003, "loss": 0.3942, "step": 9300 }, { "epoch": 8.629791140693138, "grad_norm": 0.9991121292114258, "learning_rate": 0.0003, "loss": 0.4084, "step": 9400 }, { "epoch": 8.721597429423916, "grad_norm": 1.0064568519592285, "learning_rate": 0.0003, "loss": 0.4115, "step": 9500 }, { "epoch": 8.813403718154694, "grad_norm": 1.0827409029006958, "learning_rate": 0.0003, "loss": 0.4254, "step": 9600 }, { "epoch": 8.905210006885472, "grad_norm": 1.1357035636901855, "learning_rate": 0.0003, "loss": 0.425, "step": 9700 }, { "epoch": 8.99701629561625, "grad_norm": 1.0116100311279297, "learning_rate": 0.0003, "loss": 0.438, "step": 9800 }, { "epoch": 8.999770484278173, "eval_accuracy": 0.7638136826783115, "eval_loss": 0.5532092452049255, "eval_runtime": 8.896, "eval_samples_per_second": 56.205, "eval_steps_per_second": 7.082, "step": 9803 }, { "epoch": 9.088822584347028, "grad_norm": 0.943336546421051, "learning_rate": 0.0003, "loss": 0.3196, "step": 9900 }, { "epoch": 9.180628873077806, "grad_norm": 0.966768205165863, "learning_rate": 0.0003, "loss": 0.3288, "step": 10000 }, { "epoch": 9.272435161808584, "grad_norm": 1.0522419214248657, "learning_rate": 0.0003, "loss": 0.3416, "step": 10100 }, { "epoch": 9.364241450539362, "grad_norm": 0.9877403378486633, "learning_rate": 0.0003, "loss": 0.3522, "step": 10200 }, { "epoch": 9.45604773927014, "grad_norm": 1.1037845611572266, "learning_rate": 0.0003, "loss": 0.3595, "step": 10300 }, { "epoch": 9.547854028000918, "grad_norm": 1.1503223180770874, "learning_rate": 0.0003, "loss": 0.3734, "step": 10400 }, { "epoch": 9.639660316731696, "grad_norm": 1.0321091413497925, "learning_rate": 0.0003, "loss": 0.3808, "step": 10500 }, { "epoch": 9.731466605462474, "grad_norm": 0.9578890204429626, "learning_rate": 0.0003, "loss": 0.3878, "step": 10600 }, { "epoch": 9.823272894193252, "grad_norm": 1.0930492877960205, "learning_rate": 0.0003, "loss": 0.3943, "step": 10700 }, { "epoch": 9.91507918292403, "grad_norm": 1.2598655223846436, "learning_rate": 0.0003, "loss": 0.402, "step": 10800 }, { "epoch": 9.999540968556346, "eval_accuracy": 0.7663522561863173, "eval_loss": 0.5331353545188904, "eval_runtime": 8.8958, "eval_samples_per_second": 56.206, "eval_steps_per_second": 7.082, "step": 10892 }, { "epoch": 10.006885471654808, "grad_norm": 0.9872549176216125, "learning_rate": 0.0003, "loss": 0.4016, "step": 10900 }, { "epoch": 10.098691760385586, "grad_norm": 0.9196072220802307, "learning_rate": 0.0003, "loss": 0.2985, "step": 11000 }, { "epoch": 10.190498049116364, "grad_norm": 0.9028146266937256, "learning_rate": 0.0003, "loss": 0.311, "step": 11100 }, { "epoch": 10.282304337847142, "grad_norm": 0.9019381999969482, "learning_rate": 0.0003, "loss": 0.3199, "step": 11200 }, { "epoch": 10.37411062657792, "grad_norm": 1.0821175575256348, "learning_rate": 0.0003, "loss": 0.3366, "step": 11300 }, { "epoch": 10.465916915308698, "grad_norm": 0.9856448769569397, "learning_rate": 0.0003, "loss": 0.3391, "step": 11400 }, { "epoch": 10.557723204039476, "grad_norm": 1.0605835914611816, "learning_rate": 0.0003, "loss": 0.3512, "step": 11500 }, { "epoch": 10.649529492770254, "grad_norm": 0.9575846791267395, "learning_rate": 0.0003, "loss": 0.3546, "step": 11600 }, { "epoch": 10.741335781501032, "grad_norm": 1.0335845947265625, "learning_rate": 0.0003, "loss": 0.3692, "step": 11700 }, { "epoch": 10.83314207023181, "grad_norm": 1.131901502609253, "learning_rate": 0.0003, "loss": 0.3747, "step": 11800 }, { "epoch": 10.924948358962588, "grad_norm": 0.9360109567642212, "learning_rate": 0.0003, "loss": 0.383, "step": 11900 }, { "epoch": 10.999311452834519, "eval_accuracy": 0.7685181950509461, "eval_loss": 0.5156339406967163, "eval_runtime": 8.925, "eval_samples_per_second": 56.022, "eval_steps_per_second": 7.059, "step": 11981 }, { "epoch": 11.016754647693366, "grad_norm": 0.8987607359886169, "learning_rate": 0.0003, "loss": 0.3677, "step": 12000 }, { "epoch": 11.108560936424144, "grad_norm": 0.882723867893219, "learning_rate": 0.0003, "loss": 0.2865, "step": 12100 }, { "epoch": 11.200367225154924, "grad_norm": 0.9493284821510315, "learning_rate": 0.0003, "loss": 0.2998, "step": 12200 }, { "epoch": 11.292173513885702, "grad_norm": 0.9848419427871704, "learning_rate": 0.0003, "loss": 0.3062, "step": 12300 }, { "epoch": 11.38397980261648, "grad_norm": 0.9860394597053528, "learning_rate": 0.0003, "loss": 0.3197, "step": 12400 }, { "epoch": 11.475786091347258, "grad_norm": 1.246336579322815, "learning_rate": 0.0003, "loss": 0.3318, "step": 12500 }, { "epoch": 11.567592380078036, "grad_norm": 1.0225517749786377, "learning_rate": 0.0003, "loss": 0.3333, "step": 12600 }, { "epoch": 11.659398668808814, "grad_norm": 1.043228268623352, "learning_rate": 0.0003, "loss": 0.3496, "step": 12700 }, { "epoch": 11.751204957539592, "grad_norm": 1.0331432819366455, "learning_rate": 0.0003, "loss": 0.3561, "step": 12800 }, { "epoch": 11.84301124627037, "grad_norm": 1.009142279624939, "learning_rate": 0.0003, "loss": 0.3591, "step": 12900 }, { "epoch": 11.934817535001148, "grad_norm": 0.9662990570068359, "learning_rate": 0.0003, "loss": 0.3627, "step": 13000 }, { "epoch": 12.0, "eval_accuracy": 0.7702008733624454, "eval_loss": 0.5070006251335144, "eval_runtime": 8.8961, "eval_samples_per_second": 56.204, "eval_steps_per_second": 7.082, "step": 13071 }, { "epoch": 12.026623823731926, "grad_norm": 0.8747033476829529, "learning_rate": 0.0003, "loss": 0.3406, "step": 13100 }, { "epoch": 12.118430112462704, "grad_norm": 1.044632911682129, "learning_rate": 0.0003, "loss": 0.2725, "step": 13200 }, { "epoch": 12.210236401193482, "grad_norm": 1.118489384651184, "learning_rate": 0.0003, "loss": 0.2867, "step": 13300 }, { "epoch": 12.30204268992426, "grad_norm": 1.1297811269760132, "learning_rate": 0.0003, "loss": 0.3031, "step": 13400 }, { "epoch": 12.393848978655038, "grad_norm": 1.005750298500061, "learning_rate": 0.0003, "loss": 0.308, "step": 13500 }, { "epoch": 12.485655267385816, "grad_norm": 1.1464134454727173, "learning_rate": 0.0003, "loss": 0.3173, "step": 13600 }, { "epoch": 12.577461556116594, "grad_norm": 1.0724166631698608, "learning_rate": 0.0003, "loss": 0.3263, "step": 13700 }, { "epoch": 12.669267844847372, "grad_norm": 0.9200640320777893, "learning_rate": 0.0003, "loss": 0.3348, "step": 13800 }, { "epoch": 12.76107413357815, "grad_norm": 1.012441635131836, "learning_rate": 0.0003, "loss": 0.34, "step": 13900 }, { "epoch": 12.852880422308928, "grad_norm": 1.1097270250320435, "learning_rate": 0.0003, "loss": 0.3515, "step": 14000 }, { "epoch": 12.944686711039706, "grad_norm": 1.0985119342803955, "learning_rate": 0.0003, "loss": 0.3521, "step": 14100 }, { "epoch": 12.999770484278173, "eval_accuracy": 0.7713799126637555, "eval_loss": 0.4984254539012909, "eval_runtime": 7.8448, "eval_samples_per_second": 63.737, "eval_steps_per_second": 8.031, "step": 14160 }, { "epoch": 13.036492999770484, "grad_norm": 0.8010920882225037, "learning_rate": 0.0003, "loss": 0.3178, "step": 14200 }, { "epoch": 13.128299288501262, "grad_norm": 0.7816724181175232, "learning_rate": 0.0003, "loss": 0.2667, "step": 14300 }, { "epoch": 13.22010557723204, "grad_norm": 0.9443672895431519, "learning_rate": 0.0003, "loss": 0.2779, "step": 14400 }, { "epoch": 13.311911865962818, "grad_norm": 0.9471251964569092, "learning_rate": 0.0003, "loss": 0.2889, "step": 14500 }, { "epoch": 13.403718154693596, "grad_norm": 1.0766810178756714, "learning_rate": 0.0003, "loss": 0.301, "step": 14600 }, { "epoch": 13.495524443424374, "grad_norm": 1.0341782569885254, "learning_rate": 0.0003, "loss": 0.3079, "step": 14700 }, { "epoch": 13.587330732155152, "grad_norm": 0.9588432312011719, "learning_rate": 0.0003, "loss": 0.3184, "step": 14800 }, { "epoch": 13.67913702088593, "grad_norm": 1.0522781610488892, "learning_rate": 0.0003, "loss": 0.3228, "step": 14900 }, { "epoch": 13.770943309616708, "grad_norm": 1.0346490144729614, "learning_rate": 0.0003, "loss": 0.33, "step": 15000 }, { "epoch": 13.862749598347486, "grad_norm": 1.129906177520752, "learning_rate": 0.0003, "loss": 0.3378, "step": 15100 }, { "epoch": 13.954555887078264, "grad_norm": 0.9843702912330627, "learning_rate": 0.0003, "loss": 0.344, "step": 15200 }, { "epoch": 13.999540968556346, "eval_accuracy": 0.7721863173216885, "eval_loss": 0.4925004243850708, "eval_runtime": 8.9168, "eval_samples_per_second": 56.074, "eval_steps_per_second": 7.065, "step": 15249 }, { "epoch": 14.046362175809042, "grad_norm": 0.9825355410575867, "learning_rate": 0.0003, "loss": 0.2998, "step": 15300 }, { "epoch": 14.13816846453982, "grad_norm": 1.0309946537017822, "learning_rate": 0.0003, "loss": 0.2621, "step": 15400 }, { "epoch": 14.229974753270598, "grad_norm": 0.8775806427001953, "learning_rate": 0.0003, "loss": 0.2744, "step": 15500 }, { "epoch": 14.321781042001376, "grad_norm": 0.9635316133499146, "learning_rate": 0.0003, "loss": 0.2828, "step": 15600 }, { "epoch": 14.413587330732156, "grad_norm": 1.0143553018569946, "learning_rate": 0.0003, "loss": 0.29, "step": 15700 }, { "epoch": 14.505393619462934, "grad_norm": 1.0207178592681885, "learning_rate": 0.0003, "loss": 0.3022, "step": 15800 }, { "epoch": 14.597199908193712, "grad_norm": 1.07411527633667, "learning_rate": 0.0003, "loss": 0.3113, "step": 15900 }, { "epoch": 14.68900619692449, "grad_norm": 1.0961089134216309, "learning_rate": 0.0003, "loss": 0.3153, "step": 16000 }, { "epoch": 14.780812485655268, "grad_norm": 0.9983770847320557, "learning_rate": 0.0003, "loss": 0.323, "step": 16100 }, { "epoch": 14.872618774386046, "grad_norm": 1.0680583715438843, "learning_rate": 0.0003, "loss": 0.3275, "step": 16200 }, { "epoch": 14.964425063116824, "grad_norm": 1.0475704669952393, "learning_rate": 0.0003, "loss": 0.3341, "step": 16300 }, { "epoch": 14.999311452834519, "eval_accuracy": 0.7735604075691412, "eval_loss": 0.4847230315208435, "eval_runtime": 9.1634, "eval_samples_per_second": 54.565, "eval_steps_per_second": 6.875, "step": 16338 }, { "epoch": 15.056231351847602, "grad_norm": 0.7305497527122498, "learning_rate": 0.0003, "loss": 0.2829, "step": 16400 }, { "epoch": 15.14803764057838, "grad_norm": 0.936439037322998, "learning_rate": 0.0003, "loss": 0.2585, "step": 16500 }, { "epoch": 15.239843929309158, "grad_norm": 0.9285767078399658, "learning_rate": 0.0003, "loss": 0.262, "step": 16600 }, { "epoch": 15.331650218039936, "grad_norm": 1.0587061643600464, "learning_rate": 0.0003, "loss": 0.2795, "step": 16700 }, { "epoch": 15.423456506770714, "grad_norm": 1.0315121412277222, "learning_rate": 0.0003, "loss": 0.2858, "step": 16800 }, { "epoch": 15.515262795501492, "grad_norm": 1.007581114768982, "learning_rate": 0.0003, "loss": 0.2941, "step": 16900 }, { "epoch": 15.60706908423227, "grad_norm": 1.123679280281067, "learning_rate": 0.0003, "loss": 0.2998, "step": 17000 }, { "epoch": 15.698875372963048, "grad_norm": 1.245398759841919, "learning_rate": 0.0003, "loss": 0.3086, "step": 17100 }, { "epoch": 15.790681661693826, "grad_norm": 0.9799935817718506, "learning_rate": 0.0003, "loss": 0.3133, "step": 17200 }, { "epoch": 15.882487950424604, "grad_norm": 1.1230483055114746, "learning_rate": 0.0003, "loss": 0.3251, "step": 17300 }, { "epoch": 15.974294239155382, "grad_norm": 1.2086437940597534, "learning_rate": 0.0003, "loss": 0.3275, "step": 17400 }, { "epoch": 16.0, "eval_accuracy": 0.7748180494905386, "eval_loss": 0.4808007478713989, "eval_runtime": 8.8958, "eval_samples_per_second": 56.206, "eval_steps_per_second": 7.082, "step": 17428 }, { "epoch": 16.06610052788616, "grad_norm": 0.8386597633361816, "learning_rate": 0.0003, "loss": 0.2679, "step": 17500 }, { "epoch": 16.15790681661694, "grad_norm": 0.865949273109436, "learning_rate": 0.0003, "loss": 0.2542, "step": 17600 }, { "epoch": 16.249713105347716, "grad_norm": 1.052322506904602, "learning_rate": 0.0003, "loss": 0.2646, "step": 17700 }, { "epoch": 16.341519394078496, "grad_norm": 0.8787432312965393, "learning_rate": 0.0003, "loss": 0.2671, "step": 17800 }, { "epoch": 16.433325682809272, "grad_norm": 0.9127426147460938, "learning_rate": 0.0003, "loss": 0.2772, "step": 17900 }, { "epoch": 16.525131971540052, "grad_norm": 0.9554562568664551, "learning_rate": 0.0003, "loss": 0.286, "step": 18000 }, { "epoch": 16.616938260270828, "grad_norm": 0.9382279515266418, "learning_rate": 0.0003, "loss": 0.2972, "step": 18100 }, { "epoch": 16.708744549001608, "grad_norm": 0.928916335105896, "learning_rate": 0.0003, "loss": 0.3031, "step": 18200 }, { "epoch": 16.800550837732384, "grad_norm": 0.9512932300567627, "learning_rate": 0.0003, "loss": 0.3058, "step": 18300 }, { "epoch": 16.892357126463164, "grad_norm": 1.2069571018218994, "learning_rate": 0.0003, "loss": 0.3168, "step": 18400 }, { "epoch": 16.98416341519394, "grad_norm": 1.1233407258987427, "learning_rate": 0.0003, "loss": 0.3223, "step": 18500 }, { "epoch": 16.999770484278173, "eval_accuracy": 0.7750887918486172, "eval_loss": 0.4775759279727936, "eval_runtime": 8.8897, "eval_samples_per_second": 56.245, "eval_steps_per_second": 7.087, "step": 18517 }, { "epoch": 17.07596970392472, "grad_norm": 0.9757447838783264, "learning_rate": 0.0003, "loss": 0.2534, "step": 18600 }, { "epoch": 17.167775992655496, "grad_norm": 0.7774830460548401, "learning_rate": 0.0003, "loss": 0.2487, "step": 18700 }, { "epoch": 17.259582281386276, "grad_norm": 0.9797073602676392, "learning_rate": 0.0003, "loss": 0.2564, "step": 18800 }, { "epoch": 17.351388570117052, "grad_norm": 0.9221961498260498, "learning_rate": 0.0003, "loss": 0.2675, "step": 18900 }, { "epoch": 17.443194858847832, "grad_norm": 0.9980019927024841, "learning_rate": 0.0003, "loss": 0.2759, "step": 19000 }, { "epoch": 17.53500114757861, "grad_norm": 1.1230412721633911, "learning_rate": 0.0003, "loss": 0.2884, "step": 19100 }, { "epoch": 17.626807436309388, "grad_norm": 1.0681424140930176, "learning_rate": 0.0003, "loss": 0.2894, "step": 19200 }, { "epoch": 17.718613725040164, "grad_norm": 1.0071791410446167, "learning_rate": 0.0003, "loss": 0.294, "step": 19300 }, { "epoch": 17.810420013770944, "grad_norm": 1.0229710340499878, "learning_rate": 0.0003, "loss": 0.305, "step": 19400 }, { "epoch": 17.90222630250172, "grad_norm": 1.1899082660675049, "learning_rate": 0.0003, "loss": 0.3114, "step": 19500 }, { "epoch": 17.9940325912325, "grad_norm": 1.1344189643859863, "learning_rate": 0.0003, "loss": 0.3155, "step": 19600 }, { "epoch": 17.999540968556346, "eval_accuracy": 0.7757671033478893, "eval_loss": 0.4804111421108246, "eval_runtime": 8.8913, "eval_samples_per_second": 56.234, "eval_steps_per_second": 7.086, "step": 19606 }, { "epoch": 18.085838879963276, "grad_norm": 0.9845724105834961, "learning_rate": 0.0003, "loss": 0.2418, "step": 19700 }, { "epoch": 18.177645168694056, "grad_norm": 0.879173994064331, "learning_rate": 0.0003, "loss": 0.2453, "step": 19800 }, { "epoch": 18.269451457424832, "grad_norm": 0.8902878761291504, "learning_rate": 0.0003, "loss": 0.2545, "step": 19900 }, { "epoch": 18.361257746155612, "grad_norm": 1.160914421081543, "learning_rate": 0.0003, "loss": 0.2648, "step": 20000 }, { "epoch": 18.45306403488639, "grad_norm": 0.9650329947471619, "learning_rate": 0.0003, "loss": 0.271, "step": 20100 }, { "epoch": 18.544870323617168, "grad_norm": 1.0495669841766357, "learning_rate": 0.0003, "loss": 0.2782, "step": 20200 }, { "epoch": 18.636676612347944, "grad_norm": 0.9071341156959534, "learning_rate": 0.0003, "loss": 0.2886, "step": 20300 }, { "epoch": 18.728482901078724, "grad_norm": 1.0429463386535645, "learning_rate": 0.0003, "loss": 0.2934, "step": 20400 }, { "epoch": 18.8202891898095, "grad_norm": 1.096994400024414, "learning_rate": 0.0003, "loss": 0.2992, "step": 20500 }, { "epoch": 18.91209547854028, "grad_norm": 1.0189508199691772, "learning_rate": 0.0003, "loss": 0.3033, "step": 20600 }, { "epoch": 18.99931145283452, "eval_accuracy": 0.7760727802037846, "eval_loss": 0.4787416458129883, "eval_runtime": 9.278, "eval_samples_per_second": 53.891, "eval_steps_per_second": 6.79, "step": 20695 }, { "epoch": 19.003901767271056, "grad_norm": 0.8995434045791626, "learning_rate": 0.0003, "loss": 0.3075, "step": 20700 }, { "epoch": 19.095708056001836, "grad_norm": 0.8744321465492249, "learning_rate": 0.0003, "loss": 0.2342, "step": 20800 }, { "epoch": 19.187514344732612, "grad_norm": 0.9471040964126587, "learning_rate": 0.0003, "loss": 0.2422, "step": 20900 }, { "epoch": 19.279320633463392, "grad_norm": 1.0133309364318848, "learning_rate": 0.0003, "loss": 0.2533, "step": 21000 }, { "epoch": 19.371126922194172, "grad_norm": 1.0121233463287354, "learning_rate": 0.0003, "loss": 0.2607, "step": 21100 }, { "epoch": 19.462933210924948, "grad_norm": 1.1256521940231323, "learning_rate": 0.0003, "loss": 0.2706, "step": 21200 }, { "epoch": 19.554739499655728, "grad_norm": 1.0843769311904907, "learning_rate": 0.0003, "loss": 0.2735, "step": 21300 }, { "epoch": 19.646545788386504, "grad_norm": 1.1019864082336426, "learning_rate": 0.0003, "loss": 0.286, "step": 21400 }, { "epoch": 19.738352077117284, "grad_norm": 1.135688066482544, "learning_rate": 0.0003, "loss": 0.2886, "step": 21500 }, { "epoch": 19.83015836584806, "grad_norm": 1.0445754528045654, "learning_rate": 0.0003, "loss": 0.2953, "step": 21600 }, { "epoch": 19.92196465457884, "grad_norm": 1.035718560218811, "learning_rate": 0.0003, "loss": 0.2989, "step": 21700 }, { "epoch": 19.995409685563462, "eval_accuracy": 0.7763580786026201, "eval_loss": 0.475557804107666, "eval_runtime": 9.1559, "eval_samples_per_second": 54.61, "eval_steps_per_second": 6.881, "step": 21780 }, { "epoch": 19.995409685563462, "step": 21780, "total_flos": 2.2953223726028554e+18, "train_loss": 0.5338389528903313, "train_runtime": 46593.736, "train_samples_per_second": 14.961, "train_steps_per_second": 0.467 } ], "logging_steps": 100, "max_steps": 21780, "num_input_tokens_seen": 0, "num_train_epochs": 20, "save_steps": 500, "total_flos": 2.2953223726028554e+18, "train_batch_size": 1, "trial_name": null, "trial_params": null }