|
{ |
|
"best_metric": 1.9382596015930176, |
|
"best_model_checkpoint": "./codegen-350M-mono-QLoRa-flytech/checkpoint-2000", |
|
"epoch": 0.4605907075824745, |
|
"eval_steps": 5, |
|
"global_step": 2000, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 2.743941068649292, |
|
"eval_runtime": 749.578, |
|
"eval_samples_per_second": 19.862, |
|
"eval_steps_per_second": 2.483, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.3329066038131714, |
|
"learning_rate": 0.000199, |
|
"loss": 2.9969, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 2.6796510219573975, |
|
"eval_runtime": 751.2092, |
|
"eval_samples_per_second": 19.819, |
|
"eval_steps_per_second": 2.477, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 2.5984177589416504, |
|
"eval_runtime": 764.9736, |
|
"eval_samples_per_second": 19.462, |
|
"eval_steps_per_second": 2.433, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"grad_norm": 0.283966600894928, |
|
"learning_rate": 0.00019800000000000002, |
|
"loss": 2.5613, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.0, |
|
"eval_loss": 2.510923147201538, |
|
"eval_runtime": 775.4648, |
|
"eval_samples_per_second": 19.199, |
|
"eval_steps_per_second": 2.4, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.4384284019470215, |
|
"eval_runtime": 807.2199, |
|
"eval_samples_per_second": 18.444, |
|
"eval_steps_per_second": 2.305, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.3212601840496063, |
|
"learning_rate": 0.00019700000000000002, |
|
"loss": 2.4609, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.398897886276245, |
|
"eval_runtime": 806.9842, |
|
"eval_samples_per_second": 18.449, |
|
"eval_steps_per_second": 2.306, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.3749544620513916, |
|
"eval_runtime": 807.3296, |
|
"eval_samples_per_second": 18.441, |
|
"eval_steps_per_second": 2.305, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.5253788232803345, |
|
"learning_rate": 0.000196, |
|
"loss": 2.4028, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.3531863689422607, |
|
"eval_runtime": 807.2812, |
|
"eval_samples_per_second": 18.442, |
|
"eval_steps_per_second": 2.305, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.3311145305633545, |
|
"eval_runtime": 807.5901, |
|
"eval_samples_per_second": 18.435, |
|
"eval_steps_per_second": 2.304, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.4847645163536072, |
|
"learning_rate": 0.000195, |
|
"loss": 2.4339, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.31105637550354, |
|
"eval_runtime": 807.447, |
|
"eval_samples_per_second": 18.438, |
|
"eval_steps_per_second": 2.305, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.294403314590454, |
|
"eval_runtime": 804.5301, |
|
"eval_samples_per_second": 18.505, |
|
"eval_steps_per_second": 2.313, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"grad_norm": 0.5184943675994873, |
|
"learning_rate": 0.000194, |
|
"loss": 2.3291, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.28222393989563, |
|
"eval_runtime": 744.3103, |
|
"eval_samples_per_second": 20.002, |
|
"eval_steps_per_second": 2.5, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.01, |
|
"eval_loss": 2.271087169647217, |
|
"eval_runtime": 744.2352, |
|
"eval_samples_per_second": 20.004, |
|
"eval_steps_per_second": 2.501, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.42745935916900635, |
|
"learning_rate": 0.000193, |
|
"loss": 2.3128, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.2600862979888916, |
|
"eval_runtime": 747.8797, |
|
"eval_samples_per_second": 19.907, |
|
"eval_steps_per_second": 2.488, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.251253128051758, |
|
"eval_runtime": 755.1321, |
|
"eval_samples_per_second": 19.716, |
|
"eval_steps_per_second": 2.464, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.5841810703277588, |
|
"learning_rate": 0.000192, |
|
"loss": 2.6201, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.2421493530273438, |
|
"eval_runtime": 750.5937, |
|
"eval_samples_per_second": 19.835, |
|
"eval_steps_per_second": 2.479, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.2358789443969727, |
|
"eval_runtime": 750.3335, |
|
"eval_samples_per_second": 19.842, |
|
"eval_steps_per_second": 2.48, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.34617361426353455, |
|
"learning_rate": 0.000191, |
|
"loss": 2.2322, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.2303481101989746, |
|
"eval_runtime": 748.12, |
|
"eval_samples_per_second": 19.901, |
|
"eval_steps_per_second": 2.488, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.2245254516601562, |
|
"eval_runtime": 763.3784, |
|
"eval_samples_per_second": 19.503, |
|
"eval_steps_per_second": 2.438, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"grad_norm": 0.5058037638664246, |
|
"learning_rate": 0.00019, |
|
"loss": 2.3576, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.217428207397461, |
|
"eval_runtime": 800.3219, |
|
"eval_samples_per_second": 18.603, |
|
"eval_steps_per_second": 2.325, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.02, |
|
"eval_loss": 2.2102160453796387, |
|
"eval_runtime": 806.6841, |
|
"eval_samples_per_second": 18.456, |
|
"eval_steps_per_second": 2.307, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.5872398018836975, |
|
"learning_rate": 0.00018899999999999999, |
|
"loss": 2.3798, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.203975200653076, |
|
"eval_runtime": 778.7436, |
|
"eval_samples_per_second": 19.118, |
|
"eval_steps_per_second": 2.39, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.198529005050659, |
|
"eval_runtime": 754.0178, |
|
"eval_samples_per_second": 19.745, |
|
"eval_steps_per_second": 2.468, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.4760008752346039, |
|
"learning_rate": 0.000188, |
|
"loss": 2.2499, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.1929078102111816, |
|
"eval_runtime": 770.8151, |
|
"eval_samples_per_second": 19.315, |
|
"eval_steps_per_second": 2.414, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.1861350536346436, |
|
"eval_runtime": 749.444, |
|
"eval_samples_per_second": 19.865, |
|
"eval_steps_per_second": 2.483, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.6719084978103638, |
|
"learning_rate": 0.00018700000000000002, |
|
"loss": 2.2002, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.1807847023010254, |
|
"eval_runtime": 790.3227, |
|
"eval_samples_per_second": 18.838, |
|
"eval_steps_per_second": 2.355, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.177309274673462, |
|
"eval_runtime": 767.7802, |
|
"eval_samples_per_second": 19.391, |
|
"eval_steps_per_second": 2.424, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.6400887966156006, |
|
"learning_rate": 0.00018600000000000002, |
|
"loss": 2.3217, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.1720025539398193, |
|
"eval_runtime": 794.6729, |
|
"eval_samples_per_second": 18.735, |
|
"eval_steps_per_second": 2.342, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.1658318042755127, |
|
"eval_runtime": 774.1526, |
|
"eval_samples_per_second": 19.231, |
|
"eval_steps_per_second": 2.404, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"grad_norm": 0.48335766792297363, |
|
"learning_rate": 0.00018500000000000002, |
|
"loss": 2.1297, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.03, |
|
"eval_loss": 2.1612164974212646, |
|
"eval_runtime": 778.6453, |
|
"eval_samples_per_second": 19.12, |
|
"eval_steps_per_second": 2.39, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1569318771362305, |
|
"eval_runtime": 761.275, |
|
"eval_samples_per_second": 19.557, |
|
"eval_steps_per_second": 2.445, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.8584076166152954, |
|
"learning_rate": 0.00018400000000000003, |
|
"loss": 2.1718, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1521499156951904, |
|
"eval_runtime": 746.5904, |
|
"eval_samples_per_second": 19.941, |
|
"eval_steps_per_second": 2.493, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1496639251708984, |
|
"eval_runtime": 751.6198, |
|
"eval_samples_per_second": 19.808, |
|
"eval_steps_per_second": 2.476, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.7087550163269043, |
|
"learning_rate": 0.000183, |
|
"loss": 2.3469, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.146427631378174, |
|
"eval_runtime": 768.8707, |
|
"eval_samples_per_second": 19.363, |
|
"eval_steps_per_second": 2.42, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1417484283447266, |
|
"eval_runtime": 775.3535, |
|
"eval_samples_per_second": 19.202, |
|
"eval_steps_per_second": 2.4, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.6968621611595154, |
|
"learning_rate": 0.000182, |
|
"loss": 2.2498, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.138108015060425, |
|
"eval_runtime": 773.4912, |
|
"eval_samples_per_second": 19.248, |
|
"eval_steps_per_second": 2.406, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1360790729522705, |
|
"eval_runtime": 773.9293, |
|
"eval_samples_per_second": 19.237, |
|
"eval_steps_per_second": 2.405, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"grad_norm": 0.7121121287345886, |
|
"learning_rate": 0.000181, |
|
"loss": 2.3611, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1335365772247314, |
|
"eval_runtime": 763.0423, |
|
"eval_samples_per_second": 19.511, |
|
"eval_steps_per_second": 2.439, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.04, |
|
"eval_loss": 2.1294620037078857, |
|
"eval_runtime": 755.8471, |
|
"eval_samples_per_second": 19.697, |
|
"eval_steps_per_second": 2.462, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.6406265497207642, |
|
"learning_rate": 0.00018, |
|
"loss": 2.1718, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.125330924987793, |
|
"eval_runtime": 764.112, |
|
"eval_samples_per_second": 19.484, |
|
"eval_steps_per_second": 2.436, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.121659278869629, |
|
"eval_runtime": 784.0069, |
|
"eval_samples_per_second": 18.99, |
|
"eval_steps_per_second": 2.374, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.94074547290802, |
|
"learning_rate": 0.00017900000000000001, |
|
"loss": 2.2531, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.1184146404266357, |
|
"eval_runtime": 795.2056, |
|
"eval_samples_per_second": 18.722, |
|
"eval_steps_per_second": 2.34, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.11686635017395, |
|
"eval_runtime": 782.4427, |
|
"eval_samples_per_second": 19.028, |
|
"eval_steps_per_second": 2.378, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.4468848407268524, |
|
"learning_rate": 0.00017800000000000002, |
|
"loss": 2.276, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.1175310611724854, |
|
"eval_runtime": 755.2557, |
|
"eval_samples_per_second": 19.713, |
|
"eval_steps_per_second": 2.464, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.1144332885742188, |
|
"eval_runtime": 777.8858, |
|
"eval_samples_per_second": 19.139, |
|
"eval_steps_per_second": 2.392, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"grad_norm": 0.697229266166687, |
|
"learning_rate": 0.00017700000000000002, |
|
"loss": 2.2138, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.111482620239258, |
|
"eval_runtime": 751.8011, |
|
"eval_samples_per_second": 19.803, |
|
"eval_steps_per_second": 2.475, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.05, |
|
"eval_loss": 2.1105000972747803, |
|
"eval_runtime": 751.7683, |
|
"eval_samples_per_second": 19.804, |
|
"eval_steps_per_second": 2.475, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.9181985855102539, |
|
"learning_rate": 0.00017600000000000002, |
|
"loss": 2.3692, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.109123468399048, |
|
"eval_runtime": 749.086, |
|
"eval_samples_per_second": 19.875, |
|
"eval_steps_per_second": 2.484, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.10528564453125, |
|
"eval_runtime": 750.8258, |
|
"eval_samples_per_second": 19.829, |
|
"eval_steps_per_second": 2.479, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.48959431052207947, |
|
"learning_rate": 0.000175, |
|
"loss": 2.2648, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.103628396987915, |
|
"eval_runtime": 785.1888, |
|
"eval_samples_per_second": 18.961, |
|
"eval_steps_per_second": 2.37, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.1015634536743164, |
|
"eval_runtime": 771.9757, |
|
"eval_samples_per_second": 19.286, |
|
"eval_steps_per_second": 2.411, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.6428173780441284, |
|
"learning_rate": 0.000174, |
|
"loss": 2.1139, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.097482919692993, |
|
"eval_runtime": 763.409, |
|
"eval_samples_per_second": 19.502, |
|
"eval_steps_per_second": 2.438, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.095787286758423, |
|
"eval_runtime": 760.3894, |
|
"eval_samples_per_second": 19.579, |
|
"eval_steps_per_second": 2.447, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.5767454504966736, |
|
"learning_rate": 0.000173, |
|
"loss": 2.1961, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.0945658683776855, |
|
"eval_runtime": 758.8729, |
|
"eval_samples_per_second": 19.619, |
|
"eval_steps_per_second": 2.452, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.0915117263793945, |
|
"eval_runtime": 748.4004, |
|
"eval_samples_per_second": 19.893, |
|
"eval_steps_per_second": 2.487, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"grad_norm": 0.6312183737754822, |
|
"learning_rate": 0.000172, |
|
"loss": 2.2072, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.06, |
|
"eval_loss": 2.0887210369110107, |
|
"eval_runtime": 762.6911, |
|
"eval_samples_per_second": 19.52, |
|
"eval_steps_per_second": 2.44, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.086923360824585, |
|
"eval_runtime": 806.6105, |
|
"eval_samples_per_second": 18.457, |
|
"eval_steps_per_second": 2.307, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.6048781275749207, |
|
"learning_rate": 0.000171, |
|
"loss": 2.0356, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.085240125656128, |
|
"eval_runtime": 806.3511, |
|
"eval_samples_per_second": 18.463, |
|
"eval_steps_per_second": 2.308, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.084153175354004, |
|
"eval_runtime": 806.3163, |
|
"eval_samples_per_second": 18.464, |
|
"eval_steps_per_second": 2.308, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.7954710125923157, |
|
"learning_rate": 0.00017, |
|
"loss": 2.0189, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.082386016845703, |
|
"eval_runtime": 806.423, |
|
"eval_samples_per_second": 18.462, |
|
"eval_steps_per_second": 2.308, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.0789432525634766, |
|
"eval_runtime": 806.4464, |
|
"eval_samples_per_second": 18.461, |
|
"eval_steps_per_second": 2.308, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.7538740038871765, |
|
"learning_rate": 0.00016900000000000002, |
|
"loss": 1.8978, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.076179027557373, |
|
"eval_runtime": 806.5403, |
|
"eval_samples_per_second": 18.459, |
|
"eval_steps_per_second": 2.307, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.075080156326294, |
|
"eval_runtime": 784.5471, |
|
"eval_samples_per_second": 18.977, |
|
"eval_steps_per_second": 2.372, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"grad_norm": 0.2843266427516937, |
|
"learning_rate": 0.000168, |
|
"loss": 2.1163, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.074650764465332, |
|
"eval_runtime": 752.9254, |
|
"eval_samples_per_second": 19.774, |
|
"eval_steps_per_second": 2.472, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.07, |
|
"eval_loss": 2.0736653804779053, |
|
"eval_runtime": 757.9611, |
|
"eval_samples_per_second": 19.642, |
|
"eval_steps_per_second": 2.455, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.7260650396347046, |
|
"learning_rate": 0.000167, |
|
"loss": 2.2219, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.073575496673584, |
|
"eval_runtime": 806.2254, |
|
"eval_samples_per_second": 18.466, |
|
"eval_steps_per_second": 2.308, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.0729920864105225, |
|
"eval_runtime": 779.9527, |
|
"eval_samples_per_second": 19.088, |
|
"eval_steps_per_second": 2.386, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.7835765480995178, |
|
"learning_rate": 0.000166, |
|
"loss": 2.2111, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.070727825164795, |
|
"eval_runtime": 758.1738, |
|
"eval_samples_per_second": 19.637, |
|
"eval_steps_per_second": 2.455, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.069014549255371, |
|
"eval_runtime": 747.7568, |
|
"eval_samples_per_second": 19.91, |
|
"eval_steps_per_second": 2.489, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.5928293466567993, |
|
"learning_rate": 0.000165, |
|
"loss": 2.1936, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.0676755905151367, |
|
"eval_runtime": 798.6576, |
|
"eval_samples_per_second": 18.641, |
|
"eval_steps_per_second": 2.33, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.0652225017547607, |
|
"eval_runtime": 806.0571, |
|
"eval_samples_per_second": 18.47, |
|
"eval_steps_per_second": 2.309, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"grad_norm": 0.7964838147163391, |
|
"learning_rate": 0.000164, |
|
"loss": 2.1177, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.0634725093841553, |
|
"eval_runtime": 805.9924, |
|
"eval_samples_per_second": 18.472, |
|
"eval_steps_per_second": 2.309, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.08, |
|
"eval_loss": 2.063730478286743, |
|
"eval_runtime": 805.9942, |
|
"eval_samples_per_second": 18.472, |
|
"eval_steps_per_second": 2.309, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.5622439980506897, |
|
"learning_rate": 0.000163, |
|
"loss": 2.1331, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.06245756149292, |
|
"eval_runtime": 806.0779, |
|
"eval_samples_per_second": 18.47, |
|
"eval_steps_per_second": 2.309, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.0626060962677, |
|
"eval_runtime": 806.2506, |
|
"eval_samples_per_second": 18.466, |
|
"eval_steps_per_second": 2.308, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.7700399160385132, |
|
"learning_rate": 0.000162, |
|
"loss": 2.3266, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.0606532096862793, |
|
"eval_runtime": 806.0725, |
|
"eval_samples_per_second": 18.47, |
|
"eval_steps_per_second": 2.309, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.060309410095215, |
|
"eval_runtime": 806.3975, |
|
"eval_samples_per_second": 18.462, |
|
"eval_steps_per_second": 2.308, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.7712746858596802, |
|
"learning_rate": 0.000161, |
|
"loss": 2.0627, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.0599002838134766, |
|
"eval_runtime": 806.2192, |
|
"eval_samples_per_second": 18.466, |
|
"eval_steps_per_second": 2.308, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.0586650371551514, |
|
"eval_runtime": 806.299, |
|
"eval_samples_per_second": 18.465, |
|
"eval_steps_per_second": 2.308, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.5348338484764099, |
|
"learning_rate": 0.00016, |
|
"loss": 2.3373, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.056227445602417, |
|
"eval_runtime": 806.1063, |
|
"eval_samples_per_second": 18.469, |
|
"eval_steps_per_second": 2.309, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.054941177368164, |
|
"eval_runtime": 806.3058, |
|
"eval_samples_per_second": 18.464, |
|
"eval_steps_per_second": 2.308, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"grad_norm": 0.577366828918457, |
|
"learning_rate": 0.00015900000000000002, |
|
"loss": 2.0422, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.09, |
|
"eval_loss": 2.0555078983306885, |
|
"eval_runtime": 806.2774, |
|
"eval_samples_per_second": 18.465, |
|
"eval_steps_per_second": 2.308, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.054901361465454, |
|
"eval_runtime": 806.2591, |
|
"eval_samples_per_second": 18.466, |
|
"eval_steps_per_second": 2.308, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.7189317345619202, |
|
"learning_rate": 0.00015800000000000002, |
|
"loss": 2.1189, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.0532333850860596, |
|
"eval_runtime": 806.381, |
|
"eval_samples_per_second": 18.463, |
|
"eval_steps_per_second": 2.308, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.0509543418884277, |
|
"eval_runtime": 806.2323, |
|
"eval_samples_per_second": 18.466, |
|
"eval_steps_per_second": 2.308, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.548724353313446, |
|
"learning_rate": 0.00015700000000000002, |
|
"loss": 2.0847, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.0487470626831055, |
|
"eval_runtime": 806.372, |
|
"eval_samples_per_second": 18.463, |
|
"eval_steps_per_second": 2.308, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.0478439331054688, |
|
"eval_runtime": 806.31, |
|
"eval_samples_per_second": 18.464, |
|
"eval_steps_per_second": 2.308, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.5101466774940491, |
|
"learning_rate": 0.00015600000000000002, |
|
"loss": 2.0727, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.0454509258270264, |
|
"eval_runtime": 806.399, |
|
"eval_samples_per_second": 18.462, |
|
"eval_steps_per_second": 2.308, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.043710708618164, |
|
"eval_runtime": 806.2814, |
|
"eval_samples_per_second": 18.465, |
|
"eval_steps_per_second": 2.308, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"grad_norm": 0.7228645086288452, |
|
"learning_rate": 0.000155, |
|
"loss": 2.0467, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.042335033416748, |
|
"eval_runtime": 806.3613, |
|
"eval_samples_per_second": 18.463, |
|
"eval_steps_per_second": 2.308, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.1, |
|
"eval_loss": 2.0407748222351074, |
|
"eval_runtime": 806.5248, |
|
"eval_samples_per_second": 18.459, |
|
"eval_steps_per_second": 2.307, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.7004246115684509, |
|
"learning_rate": 0.000154, |
|
"loss": 2.2736, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.0401110649108887, |
|
"eval_runtime": 807.0877, |
|
"eval_samples_per_second": 18.447, |
|
"eval_steps_per_second": 2.306, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.039562702178955, |
|
"eval_runtime": 808.9767, |
|
"eval_samples_per_second": 18.403, |
|
"eval_steps_per_second": 2.3, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.5844886898994446, |
|
"learning_rate": 0.000153, |
|
"loss": 2.0475, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.039823055267334, |
|
"eval_runtime": 768.1785, |
|
"eval_samples_per_second": 19.381, |
|
"eval_steps_per_second": 2.423, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.0399796962738037, |
|
"eval_runtime": 745.2193, |
|
"eval_samples_per_second": 19.978, |
|
"eval_steps_per_second": 2.497, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.9306110143661499, |
|
"learning_rate": 0.000152, |
|
"loss": 2.1269, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.0386431217193604, |
|
"eval_runtime": 744.6033, |
|
"eval_samples_per_second": 19.995, |
|
"eval_steps_per_second": 2.499, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.038109302520752, |
|
"eval_runtime": 744.4904, |
|
"eval_samples_per_second": 19.998, |
|
"eval_steps_per_second": 2.5, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"grad_norm": 0.7265554666519165, |
|
"learning_rate": 0.000151, |
|
"loss": 2.1965, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.039163589477539, |
|
"eval_runtime": 745.1648, |
|
"eval_samples_per_second": 19.979, |
|
"eval_steps_per_second": 2.497, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.11, |
|
"eval_loss": 2.039876699447632, |
|
"eval_runtime": 745.0115, |
|
"eval_samples_per_second": 19.984, |
|
"eval_steps_per_second": 2.498, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.7146623134613037, |
|
"learning_rate": 0.00015000000000000001, |
|
"loss": 1.9578, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.0375914573669434, |
|
"eval_runtime": 744.9842, |
|
"eval_samples_per_second": 19.984, |
|
"eval_steps_per_second": 2.498, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.0346121788024902, |
|
"eval_runtime": 746.8332, |
|
"eval_samples_per_second": 19.935, |
|
"eval_steps_per_second": 2.492, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 1.0184926986694336, |
|
"learning_rate": 0.00014900000000000002, |
|
"loss": 2.1056, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.032033920288086, |
|
"eval_runtime": 755.7533, |
|
"eval_samples_per_second": 19.7, |
|
"eval_steps_per_second": 2.462, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.0300207138061523, |
|
"eval_runtime": 752.8834, |
|
"eval_samples_per_second": 19.775, |
|
"eval_steps_per_second": 2.472, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.6594335436820984, |
|
"learning_rate": 0.000148, |
|
"loss": 2.3502, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.028749465942383, |
|
"eval_runtime": 782.3884, |
|
"eval_samples_per_second": 19.029, |
|
"eval_steps_per_second": 2.379, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.0278258323669434, |
|
"eval_runtime": 785.0029, |
|
"eval_samples_per_second": 18.966, |
|
"eval_steps_per_second": 2.371, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.8392504453659058, |
|
"learning_rate": 0.000147, |
|
"loss": 2.203, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.02721905708313, |
|
"eval_runtime": 744.3301, |
|
"eval_samples_per_second": 20.002, |
|
"eval_steps_per_second": 2.5, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.0271682739257812, |
|
"eval_runtime": 744.0199, |
|
"eval_samples_per_second": 20.01, |
|
"eval_steps_per_second": 2.501, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"grad_norm": 0.7347747683525085, |
|
"learning_rate": 0.000146, |
|
"loss": 2.2609, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.12, |
|
"eval_loss": 2.0258853435516357, |
|
"eval_runtime": 744.263, |
|
"eval_samples_per_second": 20.004, |
|
"eval_steps_per_second": 2.5, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.026393413543701, |
|
"eval_runtime": 744.209, |
|
"eval_samples_per_second": 20.005, |
|
"eval_steps_per_second": 2.501, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.4300149083137512, |
|
"learning_rate": 0.000145, |
|
"loss": 1.9083, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0268478393554688, |
|
"eval_runtime": 744.0279, |
|
"eval_samples_per_second": 20.01, |
|
"eval_steps_per_second": 2.501, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0266873836517334, |
|
"eval_runtime": 744.0752, |
|
"eval_samples_per_second": 20.009, |
|
"eval_steps_per_second": 2.501, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.46820569038391113, |
|
"learning_rate": 0.000144, |
|
"loss": 2.1976, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0266172885894775, |
|
"eval_runtime": 744.0609, |
|
"eval_samples_per_second": 20.009, |
|
"eval_steps_per_second": 2.501, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0261571407318115, |
|
"eval_runtime": 744.0034, |
|
"eval_samples_per_second": 20.011, |
|
"eval_steps_per_second": 2.501, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.7849207520484924, |
|
"learning_rate": 0.000143, |
|
"loss": 1.8726, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0230391025543213, |
|
"eval_runtime": 744.0221, |
|
"eval_samples_per_second": 20.01, |
|
"eval_steps_per_second": 2.501, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0199837684631348, |
|
"eval_runtime": 744.0989, |
|
"eval_samples_per_second": 20.008, |
|
"eval_steps_per_second": 2.501, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"grad_norm": 0.6005426645278931, |
|
"learning_rate": 0.000142, |
|
"loss": 2.0767, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0186009407043457, |
|
"eval_runtime": 744.1002, |
|
"eval_samples_per_second": 20.008, |
|
"eval_steps_per_second": 2.501, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.13, |
|
"eval_loss": 2.0188121795654297, |
|
"eval_runtime": 768.6396, |
|
"eval_samples_per_second": 19.369, |
|
"eval_steps_per_second": 2.421, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.4590970575809479, |
|
"learning_rate": 0.000141, |
|
"loss": 2.0767, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.01908278465271, |
|
"eval_runtime": 767.2243, |
|
"eval_samples_per_second": 19.405, |
|
"eval_steps_per_second": 2.426, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.0185952186584473, |
|
"eval_runtime": 744.1236, |
|
"eval_samples_per_second": 20.007, |
|
"eval_steps_per_second": 2.501, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.8970805406570435, |
|
"learning_rate": 0.00014, |
|
"loss": 2.1128, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.0173678398132324, |
|
"eval_runtime": 744.3957, |
|
"eval_samples_per_second": 20.0, |
|
"eval_steps_per_second": 2.5, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.015655279159546, |
|
"eval_runtime": 744.2471, |
|
"eval_samples_per_second": 20.004, |
|
"eval_steps_per_second": 2.501, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.6531006693840027, |
|
"learning_rate": 0.000139, |
|
"loss": 2.0945, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.0149075984954834, |
|
"eval_runtime": 744.0697, |
|
"eval_samples_per_second": 20.009, |
|
"eval_steps_per_second": 2.501, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.0139400959014893, |
|
"eval_runtime": 744.1056, |
|
"eval_samples_per_second": 20.008, |
|
"eval_steps_per_second": 2.501, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"grad_norm": 0.6212542653083801, |
|
"learning_rate": 0.000138, |
|
"loss": 2.0035, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.0124478340148926, |
|
"eval_runtime": 747.8028, |
|
"eval_samples_per_second": 19.909, |
|
"eval_steps_per_second": 2.489, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.14, |
|
"eval_loss": 2.0124478340148926, |
|
"eval_runtime": 763.4128, |
|
"eval_samples_per_second": 19.502, |
|
"eval_steps_per_second": 2.438, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.6082640290260315, |
|
"learning_rate": 0.00013700000000000002, |
|
"loss": 2.1103, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.0123772621154785, |
|
"eval_runtime": 745.1902, |
|
"eval_samples_per_second": 19.979, |
|
"eval_steps_per_second": 2.497, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.011105537414551, |
|
"eval_runtime": 799.942, |
|
"eval_samples_per_second": 18.611, |
|
"eval_steps_per_second": 2.326, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.7771211862564087, |
|
"learning_rate": 0.00013600000000000003, |
|
"loss": 2.1028, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.009859323501587, |
|
"eval_runtime": 796.318, |
|
"eval_samples_per_second": 18.696, |
|
"eval_steps_per_second": 2.337, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.009398937225342, |
|
"eval_runtime": 745.0813, |
|
"eval_samples_per_second": 19.982, |
|
"eval_steps_per_second": 2.498, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.8797178268432617, |
|
"learning_rate": 0.00013500000000000003, |
|
"loss": 2.2711, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.0092194080352783, |
|
"eval_runtime": 743.6761, |
|
"eval_samples_per_second": 20.019, |
|
"eval_steps_per_second": 2.502, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.00980806350708, |
|
"eval_runtime": 742.9948, |
|
"eval_samples_per_second": 20.038, |
|
"eval_steps_per_second": 2.505, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 1.1657720804214478, |
|
"learning_rate": 0.000134, |
|
"loss": 2.243, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.0087344646453857, |
|
"eval_runtime": 742.7774, |
|
"eval_samples_per_second": 20.044, |
|
"eval_steps_per_second": 2.505, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.0070958137512207, |
|
"eval_runtime": 748.9587, |
|
"eval_samples_per_second": 19.878, |
|
"eval_steps_per_second": 2.485, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"grad_norm": 0.6517012715339661, |
|
"learning_rate": 0.000133, |
|
"loss": 1.9059, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.15, |
|
"eval_loss": 2.0066118240356445, |
|
"eval_runtime": 746.8276, |
|
"eval_samples_per_second": 19.935, |
|
"eval_steps_per_second": 2.492, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.006216287612915, |
|
"eval_runtime": 746.1771, |
|
"eval_samples_per_second": 19.952, |
|
"eval_steps_per_second": 2.494, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.6520020365715027, |
|
"learning_rate": 0.000132, |
|
"loss": 1.9953, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.0057804584503174, |
|
"eval_runtime": 800.2909, |
|
"eval_samples_per_second": 18.603, |
|
"eval_steps_per_second": 2.325, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.0061757564544678, |
|
"eval_runtime": 743.2104, |
|
"eval_samples_per_second": 20.032, |
|
"eval_steps_per_second": 2.504, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.7605012059211731, |
|
"learning_rate": 0.000131, |
|
"loss": 2.0684, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.0061566829681396, |
|
"eval_runtime": 743.1127, |
|
"eval_samples_per_second": 20.035, |
|
"eval_steps_per_second": 2.504, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.005591869354248, |
|
"eval_runtime": 743.2532, |
|
"eval_samples_per_second": 20.031, |
|
"eval_steps_per_second": 2.504, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.8598791360855103, |
|
"learning_rate": 0.00013000000000000002, |
|
"loss": 1.9624, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.00502347946167, |
|
"eval_runtime": 743.0596, |
|
"eval_samples_per_second": 20.036, |
|
"eval_steps_per_second": 2.505, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.0032622814178467, |
|
"eval_runtime": 743.2083, |
|
"eval_samples_per_second": 20.032, |
|
"eval_steps_per_second": 2.504, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"grad_norm": 0.5449765920639038, |
|
"learning_rate": 0.00012900000000000002, |
|
"loss": 2.0671, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.0018348693847656, |
|
"eval_runtime": 742.4305, |
|
"eval_samples_per_second": 20.053, |
|
"eval_steps_per_second": 2.507, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.16, |
|
"eval_loss": 2.001499652862549, |
|
"eval_runtime": 742.3408, |
|
"eval_samples_per_second": 20.055, |
|
"eval_steps_per_second": 2.507, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.7031337022781372, |
|
"learning_rate": 0.00012800000000000002, |
|
"loss": 2.1012, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 2.0012526512145996, |
|
"eval_runtime": 742.3685, |
|
"eval_samples_per_second": 20.055, |
|
"eval_steps_per_second": 2.507, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 2.0008952617645264, |
|
"eval_runtime": 742.5083, |
|
"eval_samples_per_second": 20.051, |
|
"eval_steps_per_second": 2.506, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.9199779033660889, |
|
"learning_rate": 0.000127, |
|
"loss": 2.1093, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 2.0010104179382324, |
|
"eval_runtime": 742.3134, |
|
"eval_samples_per_second": 20.056, |
|
"eval_steps_per_second": 2.507, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 2.000081777572632, |
|
"eval_runtime": 742.5728, |
|
"eval_samples_per_second": 20.049, |
|
"eval_steps_per_second": 2.506, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.5495649576187134, |
|
"learning_rate": 0.000126, |
|
"loss": 2.0745, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 1.9993902444839478, |
|
"eval_runtime": 742.2628, |
|
"eval_samples_per_second": 20.058, |
|
"eval_steps_per_second": 2.507, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 1.9984277486801147, |
|
"eval_runtime": 742.413, |
|
"eval_samples_per_second": 20.054, |
|
"eval_steps_per_second": 2.507, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"grad_norm": 0.6500754952430725, |
|
"learning_rate": 0.000125, |
|
"loss": 2.1672, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 1.9985421895980835, |
|
"eval_runtime": 742.5432, |
|
"eval_samples_per_second": 20.05, |
|
"eval_steps_per_second": 2.506, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.17, |
|
"eval_loss": 1.9995174407958984, |
|
"eval_runtime": 742.7051, |
|
"eval_samples_per_second": 20.046, |
|
"eval_steps_per_second": 2.506, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.6539570689201355, |
|
"learning_rate": 0.000124, |
|
"loss": 2.0329, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.999580979347229, |
|
"eval_runtime": 742.8251, |
|
"eval_samples_per_second": 20.042, |
|
"eval_steps_per_second": 2.505, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.9998208284378052, |
|
"eval_runtime": 742.7139, |
|
"eval_samples_per_second": 20.045, |
|
"eval_steps_per_second": 2.506, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.7342298626899719, |
|
"learning_rate": 0.000123, |
|
"loss": 2.1127, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.998628854751587, |
|
"eval_runtime": 742.6862, |
|
"eval_samples_per_second": 20.046, |
|
"eval_steps_per_second": 2.506, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.9968189001083374, |
|
"eval_runtime": 742.6582, |
|
"eval_samples_per_second": 20.047, |
|
"eval_steps_per_second": 2.506, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.5857991576194763, |
|
"learning_rate": 0.000122, |
|
"loss": 1.83, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.9953583478927612, |
|
"eval_runtime": 742.6099, |
|
"eval_samples_per_second": 20.048, |
|
"eval_steps_per_second": 2.506, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.9942070245742798, |
|
"eval_runtime": 742.6392, |
|
"eval_samples_per_second": 20.047, |
|
"eval_steps_per_second": 2.506, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.7684934735298157, |
|
"learning_rate": 0.000121, |
|
"loss": 2.1953, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.9930860996246338, |
|
"eval_runtime": 742.5366, |
|
"eval_samples_per_second": 20.05, |
|
"eval_steps_per_second": 2.506, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.9929825067520142, |
|
"eval_runtime": 742.4635, |
|
"eval_samples_per_second": 20.052, |
|
"eval_steps_per_second": 2.507, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"grad_norm": 0.7895638346672058, |
|
"learning_rate": 0.00012, |
|
"loss": 2.0523, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.18, |
|
"eval_loss": 1.993416428565979, |
|
"eval_runtime": 742.3749, |
|
"eval_samples_per_second": 20.055, |
|
"eval_steps_per_second": 2.507, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.9942388534545898, |
|
"eval_runtime": 742.6099, |
|
"eval_samples_per_second": 20.048, |
|
"eval_steps_per_second": 2.506, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.7561328411102295, |
|
"learning_rate": 0.000119, |
|
"loss": 1.9476, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.9943628311157227, |
|
"eval_runtime": 742.6406, |
|
"eval_samples_per_second": 20.047, |
|
"eval_steps_per_second": 2.506, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.993873953819275, |
|
"eval_runtime": 742.6764, |
|
"eval_samples_per_second": 20.046, |
|
"eval_steps_per_second": 2.506, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.6752389073371887, |
|
"learning_rate": 0.000118, |
|
"loss": 2.2179, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.993316411972046, |
|
"eval_runtime": 742.8432, |
|
"eval_samples_per_second": 20.042, |
|
"eval_steps_per_second": 2.505, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.9925029277801514, |
|
"eval_runtime": 742.8122, |
|
"eval_samples_per_second": 20.043, |
|
"eval_steps_per_second": 2.505, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8891268372535706, |
|
"learning_rate": 0.000117, |
|
"loss": 2.0997, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.9914040565490723, |
|
"eval_runtime": 742.5596, |
|
"eval_samples_per_second": 20.05, |
|
"eval_steps_per_second": 2.506, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.989838719367981, |
|
"eval_runtime": 742.7205, |
|
"eval_samples_per_second": 20.045, |
|
"eval_steps_per_second": 2.506, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"grad_norm": 0.8138695359230042, |
|
"learning_rate": 0.000116, |
|
"loss": 2.0146, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.9878920316696167, |
|
"eval_runtime": 746.0661, |
|
"eval_samples_per_second": 19.955, |
|
"eval_steps_per_second": 2.494, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.19, |
|
"eval_loss": 1.986674427986145, |
|
"eval_runtime": 753.8888, |
|
"eval_samples_per_second": 19.748, |
|
"eval_steps_per_second": 2.469, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8682060241699219, |
|
"learning_rate": 0.00011499999999999999, |
|
"loss": 2.1124, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9865528345108032, |
|
"eval_runtime": 754.5751, |
|
"eval_samples_per_second": 19.73, |
|
"eval_steps_per_second": 2.466, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9874132871627808, |
|
"eval_runtime": 758.8328, |
|
"eval_samples_per_second": 19.62, |
|
"eval_steps_per_second": 2.452, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8115680813789368, |
|
"learning_rate": 0.00011399999999999999, |
|
"loss": 2.3409, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9879497289657593, |
|
"eval_runtime": 752.4654, |
|
"eval_samples_per_second": 19.786, |
|
"eval_steps_per_second": 2.473, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9877504110336304, |
|
"eval_runtime": 777.0301, |
|
"eval_samples_per_second": 19.16, |
|
"eval_steps_per_second": 2.395, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.8121163845062256, |
|
"learning_rate": 0.000113, |
|
"loss": 2.1718, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9862942695617676, |
|
"eval_runtime": 789.4357, |
|
"eval_samples_per_second": 18.859, |
|
"eval_steps_per_second": 2.357, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9849509000778198, |
|
"eval_runtime": 766.4313, |
|
"eval_samples_per_second": 19.425, |
|
"eval_steps_per_second": 2.428, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.6633480787277222, |
|
"learning_rate": 0.00011200000000000001, |
|
"loss": 2.0154, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9856353998184204, |
|
"eval_runtime": 797.5238, |
|
"eval_samples_per_second": 18.668, |
|
"eval_steps_per_second": 2.333, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9872477054595947, |
|
"eval_runtime": 796.5052, |
|
"eval_samples_per_second": 18.692, |
|
"eval_steps_per_second": 2.336, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"grad_norm": 0.6087526679039001, |
|
"learning_rate": 0.00011100000000000001, |
|
"loss": 2.03, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.2, |
|
"eval_loss": 1.9866758584976196, |
|
"eval_runtime": 776.7422, |
|
"eval_samples_per_second": 19.167, |
|
"eval_steps_per_second": 2.396, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.985320806503296, |
|
"eval_runtime": 801.3372, |
|
"eval_samples_per_second": 18.579, |
|
"eval_steps_per_second": 2.322, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8433945178985596, |
|
"learning_rate": 0.00011000000000000002, |
|
"loss": 2.0573, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.9836574792861938, |
|
"eval_runtime": 767.1796, |
|
"eval_samples_per_second": 19.406, |
|
"eval_steps_per_second": 2.426, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.9830906391143799, |
|
"eval_runtime": 755.8391, |
|
"eval_samples_per_second": 19.697, |
|
"eval_steps_per_second": 2.462, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.9100953340530396, |
|
"learning_rate": 0.000109, |
|
"loss": 1.9915, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.9824570417404175, |
|
"eval_runtime": 766.9525, |
|
"eval_samples_per_second": 19.412, |
|
"eval_steps_per_second": 2.426, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.983730435371399, |
|
"eval_runtime": 761.0826, |
|
"eval_samples_per_second": 19.562, |
|
"eval_steps_per_second": 2.445, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.5524155497550964, |
|
"learning_rate": 0.00010800000000000001, |
|
"loss": 2.4334, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.9855434894561768, |
|
"eval_runtime": 806.2529, |
|
"eval_samples_per_second": 18.466, |
|
"eval_steps_per_second": 2.308, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.9850364923477173, |
|
"eval_runtime": 759.49, |
|
"eval_samples_per_second": 19.603, |
|
"eval_steps_per_second": 2.45, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"grad_norm": 0.8319659233093262, |
|
"learning_rate": 0.00010700000000000001, |
|
"loss": 2.0895, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.21, |
|
"eval_loss": 1.9837418794631958, |
|
"eval_runtime": 742.155, |
|
"eval_samples_per_second": 20.06, |
|
"eval_steps_per_second": 2.508, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9824495315551758, |
|
"eval_runtime": 742.1038, |
|
"eval_samples_per_second": 20.062, |
|
"eval_steps_per_second": 2.508, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.5936623811721802, |
|
"learning_rate": 0.00010600000000000002, |
|
"loss": 2.018, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9815988540649414, |
|
"eval_runtime": 742.0378, |
|
"eval_samples_per_second": 20.064, |
|
"eval_steps_per_second": 2.508, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9809609651565552, |
|
"eval_runtime": 742.3078, |
|
"eval_samples_per_second": 20.056, |
|
"eval_steps_per_second": 2.507, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.6863654255867004, |
|
"learning_rate": 0.000105, |
|
"loss": 2.1028, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9804768562316895, |
|
"eval_runtime": 742.0981, |
|
"eval_samples_per_second": 20.062, |
|
"eval_steps_per_second": 2.508, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.979957103729248, |
|
"eval_runtime": 742.167, |
|
"eval_samples_per_second": 20.06, |
|
"eval_steps_per_second": 2.508, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 0.8558996915817261, |
|
"learning_rate": 0.00010400000000000001, |
|
"loss": 2.0259, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9790544509887695, |
|
"eval_runtime": 742.0942, |
|
"eval_samples_per_second": 20.062, |
|
"eval_steps_per_second": 2.508, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9782884120941162, |
|
"eval_runtime": 742.2666, |
|
"eval_samples_per_second": 20.057, |
|
"eval_steps_per_second": 2.507, |
|
"step": 965 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"grad_norm": 1.0001534223556519, |
|
"learning_rate": 0.00010300000000000001, |
|
"loss": 1.9648, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9775859117507935, |
|
"eval_runtime": 742.263, |
|
"eval_samples_per_second": 20.058, |
|
"eval_steps_per_second": 2.507, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.22, |
|
"eval_loss": 1.9767318964004517, |
|
"eval_runtime": 742.3891, |
|
"eval_samples_per_second": 20.054, |
|
"eval_steps_per_second": 2.507, |
|
"step": 975 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.8744415640830994, |
|
"learning_rate": 0.00010200000000000001, |
|
"loss": 2.0649, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.976646065711975, |
|
"eval_runtime": 742.1965, |
|
"eval_samples_per_second": 20.059, |
|
"eval_steps_per_second": 2.507, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.9765182733535767, |
|
"eval_runtime": 742.453, |
|
"eval_samples_per_second": 20.052, |
|
"eval_steps_per_second": 2.507, |
|
"step": 985 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.7335282564163208, |
|
"learning_rate": 0.000101, |
|
"loss": 2.0636, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.9757829904556274, |
|
"eval_runtime": 742.3078, |
|
"eval_samples_per_second": 20.056, |
|
"eval_steps_per_second": 2.507, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.9756759405136108, |
|
"eval_runtime": 742.3788, |
|
"eval_samples_per_second": 20.054, |
|
"eval_steps_per_second": 2.507, |
|
"step": 995 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.9827843904495239, |
|
"learning_rate": 0.0001, |
|
"loss": 1.9848, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.9751605987548828, |
|
"eval_runtime": 742.3183, |
|
"eval_samples_per_second": 20.056, |
|
"eval_steps_per_second": 2.507, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.974611759185791, |
|
"eval_runtime": 741.7661, |
|
"eval_samples_per_second": 20.071, |
|
"eval_steps_per_second": 2.509, |
|
"step": 1005 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.8337591290473938, |
|
"learning_rate": 9.900000000000001e-05, |
|
"loss": 1.8473, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.974379301071167, |
|
"eval_runtime": 741.9424, |
|
"eval_samples_per_second": 20.066, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.973697304725647, |
|
"eval_runtime": 741.9395, |
|
"eval_samples_per_second": 20.066, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1015 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"grad_norm": 0.5600916147232056, |
|
"learning_rate": 9.8e-05, |
|
"loss": 1.9928, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.23, |
|
"eval_loss": 1.9728336334228516, |
|
"eval_runtime": 742.058, |
|
"eval_samples_per_second": 20.063, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9720138311386108, |
|
"eval_runtime": 741.909, |
|
"eval_samples_per_second": 20.067, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1025 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.7983824014663696, |
|
"learning_rate": 9.7e-05, |
|
"loss": 1.9903, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9715585708618164, |
|
"eval_runtime": 741.9989, |
|
"eval_samples_per_second": 20.065, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.971234917640686, |
|
"eval_runtime": 742.1566, |
|
"eval_samples_per_second": 20.06, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1035 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 1.0410112142562866, |
|
"learning_rate": 9.6e-05, |
|
"loss": 2.2528, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9711899757385254, |
|
"eval_runtime": 742.4115, |
|
"eval_samples_per_second": 20.054, |
|
"eval_steps_per_second": 2.507, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9705572128295898, |
|
"eval_runtime": 742.1151, |
|
"eval_samples_per_second": 20.062, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1045 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.9996401071548462, |
|
"learning_rate": 9.5e-05, |
|
"loss": 1.8971, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9699151515960693, |
|
"eval_runtime": 742.02, |
|
"eval_samples_per_second": 20.064, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9693446159362793, |
|
"eval_runtime": 742.2312, |
|
"eval_samples_per_second": 20.058, |
|
"eval_steps_per_second": 2.507, |
|
"step": 1055 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"grad_norm": 0.6173760294914246, |
|
"learning_rate": 9.4e-05, |
|
"loss": 1.9011, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.24, |
|
"eval_loss": 1.9691505432128906, |
|
"eval_runtime": 742.0593, |
|
"eval_samples_per_second": 20.063, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9691288471221924, |
|
"eval_runtime": 741.9923, |
|
"eval_samples_per_second": 20.065, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1065 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.7394955158233643, |
|
"learning_rate": 9.300000000000001e-05, |
|
"loss": 2.0518, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9698320627212524, |
|
"eval_runtime": 742.0516, |
|
"eval_samples_per_second": 20.063, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9706062078475952, |
|
"eval_runtime": 746.4611, |
|
"eval_samples_per_second": 19.945, |
|
"eval_steps_per_second": 2.493, |
|
"step": 1075 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.6462586522102356, |
|
"learning_rate": 9.200000000000001e-05, |
|
"loss": 2.04, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9709925651550293, |
|
"eval_runtime": 793.086, |
|
"eval_samples_per_second": 18.772, |
|
"eval_steps_per_second": 2.347, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9699233770370483, |
|
"eval_runtime": 772.4813, |
|
"eval_samples_per_second": 19.273, |
|
"eval_steps_per_second": 2.409, |
|
"step": 1085 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 0.8134130835533142, |
|
"learning_rate": 9.1e-05, |
|
"loss": 1.9992, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9690241813659668, |
|
"eval_runtime": 742.7771, |
|
"eval_samples_per_second": 20.044, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9685667753219604, |
|
"eval_runtime": 742.8374, |
|
"eval_samples_per_second": 20.042, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1095 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"grad_norm": 1.065739631652832, |
|
"learning_rate": 9e-05, |
|
"loss": 2.2447, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.9679232835769653, |
|
"eval_runtime": 743.058, |
|
"eval_samples_per_second": 20.036, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.25, |
|
"eval_loss": 1.967824101448059, |
|
"eval_runtime": 742.6615, |
|
"eval_samples_per_second": 20.047, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1105 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.8471609950065613, |
|
"learning_rate": 8.900000000000001e-05, |
|
"loss": 2.0276, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.9677886962890625, |
|
"eval_runtime": 742.686, |
|
"eval_samples_per_second": 20.046, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.9673609733581543, |
|
"eval_runtime": 742.7476, |
|
"eval_samples_per_second": 20.044, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1115 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.0944584608078003, |
|
"learning_rate": 8.800000000000001e-05, |
|
"loss": 1.9626, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.9674118757247925, |
|
"eval_runtime": 742.8656, |
|
"eval_samples_per_second": 20.041, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.966983675956726, |
|
"eval_runtime": 742.5557, |
|
"eval_samples_per_second": 20.05, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1125 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 1.0942668914794922, |
|
"learning_rate": 8.7e-05, |
|
"loss": 2.1137, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.966120719909668, |
|
"eval_runtime": 743.2494, |
|
"eval_samples_per_second": 20.031, |
|
"eval_steps_per_second": 2.504, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.9657062292099, |
|
"eval_runtime": 742.7716, |
|
"eval_samples_per_second": 20.044, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1135 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.9553364515304565, |
|
"learning_rate": 8.6e-05, |
|
"loss": 1.9172, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.9652293920516968, |
|
"eval_runtime": 742.8386, |
|
"eval_samples_per_second": 20.042, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.965266466140747, |
|
"eval_runtime": 742.7224, |
|
"eval_samples_per_second": 20.045, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1145 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"grad_norm": 0.7700403332710266, |
|
"learning_rate": 8.5e-05, |
|
"loss": 1.969, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.26, |
|
"eval_loss": 1.9645556211471558, |
|
"eval_runtime": 742.7935, |
|
"eval_samples_per_second": 20.043, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.9638077020645142, |
|
"eval_runtime": 742.9212, |
|
"eval_samples_per_second": 20.04, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1155 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.6622421741485596, |
|
"learning_rate": 8.4e-05, |
|
"loss": 2.0767, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.9633898735046387, |
|
"eval_runtime": 743.0434, |
|
"eval_samples_per_second": 20.037, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.9631668329238892, |
|
"eval_runtime": 743.0892, |
|
"eval_samples_per_second": 20.035, |
|
"eval_steps_per_second": 2.504, |
|
"step": 1165 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.902953565120697, |
|
"learning_rate": 8.3e-05, |
|
"loss": 1.9907, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.9637709856033325, |
|
"eval_runtime": 742.9531, |
|
"eval_samples_per_second": 20.039, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.964076042175293, |
|
"eval_runtime": 752.1865, |
|
"eval_samples_per_second": 19.793, |
|
"eval_steps_per_second": 2.474, |
|
"step": 1175 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 0.6186763644218445, |
|
"learning_rate": 8.2e-05, |
|
"loss": 2.0733, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.9636982679367065, |
|
"eval_runtime": 806.7383, |
|
"eval_samples_per_second": 18.455, |
|
"eval_steps_per_second": 2.307, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.9626468420028687, |
|
"eval_runtime": 794.7659, |
|
"eval_samples_per_second": 18.733, |
|
"eval_steps_per_second": 2.342, |
|
"step": 1185 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"grad_norm": 2.3641421794891357, |
|
"learning_rate": 8.1e-05, |
|
"loss": 2.2387, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.27, |
|
"eval_loss": 1.962171196937561, |
|
"eval_runtime": 742.946, |
|
"eval_samples_per_second": 20.039, |
|
"eval_steps_per_second": 2.505, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9623751640319824, |
|
"eval_runtime": 748.6093, |
|
"eval_samples_per_second": 19.888, |
|
"eval_steps_per_second": 2.486, |
|
"step": 1195 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.8681169748306274, |
|
"learning_rate": 8e-05, |
|
"loss": 1.7697, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9623966217041016, |
|
"eval_runtime": 747.3889, |
|
"eval_samples_per_second": 19.92, |
|
"eval_steps_per_second": 2.49, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9618523120880127, |
|
"eval_runtime": 747.8306, |
|
"eval_samples_per_second": 19.908, |
|
"eval_steps_per_second": 2.489, |
|
"step": 1205 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.8061439990997314, |
|
"learning_rate": 7.900000000000001e-05, |
|
"loss": 1.8372, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9610604047775269, |
|
"eval_runtime": 754.4243, |
|
"eval_samples_per_second": 19.734, |
|
"eval_steps_per_second": 2.467, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9600740671157837, |
|
"eval_runtime": 773.0575, |
|
"eval_samples_per_second": 19.259, |
|
"eval_steps_per_second": 2.407, |
|
"step": 1215 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.6680261492729187, |
|
"learning_rate": 7.800000000000001e-05, |
|
"loss": 1.8579, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9600106477737427, |
|
"eval_runtime": 758.3783, |
|
"eval_samples_per_second": 19.631, |
|
"eval_steps_per_second": 2.454, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.960121512413025, |
|
"eval_runtime": 742.6905, |
|
"eval_samples_per_second": 20.046, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1225 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"grad_norm": 0.7453675866127014, |
|
"learning_rate": 7.7e-05, |
|
"loss": 2.0669, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9599344730377197, |
|
"eval_runtime": 742.6985, |
|
"eval_samples_per_second": 20.046, |
|
"eval_steps_per_second": 2.506, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.28, |
|
"eval_loss": 1.9595812559127808, |
|
"eval_runtime": 743.459, |
|
"eval_samples_per_second": 20.025, |
|
"eval_steps_per_second": 2.503, |
|
"step": 1235 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.8137094378471375, |
|
"learning_rate": 7.6e-05, |
|
"loss": 1.8052, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.959162950515747, |
|
"eval_runtime": 762.4324, |
|
"eval_samples_per_second": 19.527, |
|
"eval_steps_per_second": 2.441, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.958164930343628, |
|
"eval_runtime": 760.8758, |
|
"eval_samples_per_second": 19.567, |
|
"eval_steps_per_second": 2.446, |
|
"step": 1245 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.9641257524490356, |
|
"learning_rate": 7.500000000000001e-05, |
|
"loss": 2.012, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.9580010175704956, |
|
"eval_runtime": 758.6208, |
|
"eval_samples_per_second": 19.625, |
|
"eval_steps_per_second": 2.453, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.9579390287399292, |
|
"eval_runtime": 756.8034, |
|
"eval_samples_per_second": 19.672, |
|
"eval_steps_per_second": 2.459, |
|
"step": 1255 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.9015198349952698, |
|
"learning_rate": 7.4e-05, |
|
"loss": 1.7978, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.957844614982605, |
|
"eval_runtime": 762.8272, |
|
"eval_samples_per_second": 19.517, |
|
"eval_steps_per_second": 2.44, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.9587702751159668, |
|
"eval_runtime": 757.6095, |
|
"eval_samples_per_second": 19.651, |
|
"eval_steps_per_second": 2.456, |
|
"step": 1265 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 0.6779741644859314, |
|
"learning_rate": 7.3e-05, |
|
"loss": 2.2184, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.9595640897750854, |
|
"eval_runtime": 752.2491, |
|
"eval_samples_per_second": 19.791, |
|
"eval_steps_per_second": 2.474, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.9594327211380005, |
|
"eval_runtime": 753.8257, |
|
"eval_samples_per_second": 19.75, |
|
"eval_steps_per_second": 2.469, |
|
"step": 1275 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"grad_norm": 1.2027363777160645, |
|
"learning_rate": 7.2e-05, |
|
"loss": 2.1702, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.29, |
|
"eval_loss": 1.9583239555358887, |
|
"eval_runtime": 762.8807, |
|
"eval_samples_per_second": 19.516, |
|
"eval_steps_per_second": 2.439, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9568368196487427, |
|
"eval_runtime": 760.0025, |
|
"eval_samples_per_second": 19.589, |
|
"eval_steps_per_second": 2.449, |
|
"step": 1285 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.690773069858551, |
|
"learning_rate": 7.1e-05, |
|
"loss": 2.1691, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9561883211135864, |
|
"eval_runtime": 763.7974, |
|
"eval_samples_per_second": 19.492, |
|
"eval_steps_per_second": 2.437, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9560072422027588, |
|
"eval_runtime": 750.1629, |
|
"eval_samples_per_second": 19.846, |
|
"eval_steps_per_second": 2.481, |
|
"step": 1295 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.5568638443946838, |
|
"learning_rate": 7e-05, |
|
"loss": 2.0352, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9558581113815308, |
|
"eval_runtime": 771.8238, |
|
"eval_samples_per_second": 19.289, |
|
"eval_steps_per_second": 2.411, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9561656713485718, |
|
"eval_runtime": 761.3654, |
|
"eval_samples_per_second": 19.554, |
|
"eval_steps_per_second": 2.444, |
|
"step": 1305 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.9243978261947632, |
|
"learning_rate": 6.9e-05, |
|
"loss": 2.1049, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9566450119018555, |
|
"eval_runtime": 774.2138, |
|
"eval_samples_per_second": 19.23, |
|
"eval_steps_per_second": 2.404, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.956655502319336, |
|
"eval_runtime": 757.6655, |
|
"eval_samples_per_second": 19.65, |
|
"eval_steps_per_second": 2.456, |
|
"step": 1315 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"grad_norm": 0.799391508102417, |
|
"learning_rate": 6.800000000000001e-05, |
|
"loss": 1.9632, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.3, |
|
"eval_loss": 1.9563167095184326, |
|
"eval_runtime": 775.3068, |
|
"eval_samples_per_second": 19.203, |
|
"eval_steps_per_second": 2.4, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.955706000328064, |
|
"eval_runtime": 775.6046, |
|
"eval_samples_per_second": 19.195, |
|
"eval_steps_per_second": 2.399, |
|
"step": 1325 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.5652315020561218, |
|
"learning_rate": 6.7e-05, |
|
"loss": 2.1231, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9552385807037354, |
|
"eval_runtime": 774.2845, |
|
"eval_samples_per_second": 19.228, |
|
"eval_steps_per_second": 2.404, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9548553228378296, |
|
"eval_runtime": 804.564, |
|
"eval_samples_per_second": 18.504, |
|
"eval_steps_per_second": 2.313, |
|
"step": 1335 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.838487446308136, |
|
"learning_rate": 6.6e-05, |
|
"loss": 2.1477, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9545986652374268, |
|
"eval_runtime": 804.7596, |
|
"eval_samples_per_second": 18.5, |
|
"eval_steps_per_second": 2.312, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.954168438911438, |
|
"eval_runtime": 804.8099, |
|
"eval_samples_per_second": 18.499, |
|
"eval_steps_per_second": 2.312, |
|
"step": 1345 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.6252363920211792, |
|
"learning_rate": 6.500000000000001e-05, |
|
"loss": 1.9284, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9540753364562988, |
|
"eval_runtime": 756.5934, |
|
"eval_samples_per_second": 19.678, |
|
"eval_steps_per_second": 2.46, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9538133144378662, |
|
"eval_runtime": 753.786, |
|
"eval_samples_per_second": 19.751, |
|
"eval_steps_per_second": 2.469, |
|
"step": 1355 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"grad_norm": 0.7139986753463745, |
|
"learning_rate": 6.400000000000001e-05, |
|
"loss": 1.9112, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9534461498260498, |
|
"eval_runtime": 752.6982, |
|
"eval_samples_per_second": 19.78, |
|
"eval_steps_per_second": 2.472, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.31, |
|
"eval_loss": 1.9531402587890625, |
|
"eval_runtime": 753.832, |
|
"eval_samples_per_second": 19.75, |
|
"eval_steps_per_second": 2.469, |
|
"step": 1365 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.7954660058021545, |
|
"learning_rate": 6.3e-05, |
|
"loss": 1.879, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9530184268951416, |
|
"eval_runtime": 749.4503, |
|
"eval_samples_per_second": 19.865, |
|
"eval_steps_per_second": 2.483, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9531172513961792, |
|
"eval_runtime": 751.563, |
|
"eval_samples_per_second": 19.809, |
|
"eval_steps_per_second": 2.476, |
|
"step": 1375 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.6721971035003662, |
|
"learning_rate": 6.2e-05, |
|
"loss": 1.793, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9533251523971558, |
|
"eval_runtime": 760.6235, |
|
"eval_samples_per_second": 19.573, |
|
"eval_steps_per_second": 2.447, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9533405303955078, |
|
"eval_runtime": 762.9017, |
|
"eval_samples_per_second": 19.515, |
|
"eval_steps_per_second": 2.439, |
|
"step": 1385 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.9299044609069824, |
|
"learning_rate": 6.1e-05, |
|
"loss": 2.0359, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9533050060272217, |
|
"eval_runtime": 770.8252, |
|
"eval_samples_per_second": 19.314, |
|
"eval_steps_per_second": 2.414, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9530737400054932, |
|
"eval_runtime": 774.7678, |
|
"eval_samples_per_second": 19.216, |
|
"eval_steps_per_second": 2.402, |
|
"step": 1395 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8456938862800598, |
|
"learning_rate": 6e-05, |
|
"loss": 2.1071, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9526972770690918, |
|
"eval_runtime": 785.1544, |
|
"eval_samples_per_second": 18.962, |
|
"eval_steps_per_second": 2.37, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.9520927667617798, |
|
"eval_runtime": 774.7517, |
|
"eval_samples_per_second": 19.216, |
|
"eval_steps_per_second": 2.402, |
|
"step": 1405 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"grad_norm": 0.8644620180130005, |
|
"learning_rate": 5.9e-05, |
|
"loss": 1.922, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.32, |
|
"eval_loss": 1.951416015625, |
|
"eval_runtime": 770.2102, |
|
"eval_samples_per_second": 19.33, |
|
"eval_steps_per_second": 2.416, |
|
"step": 1410 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.951120138168335, |
|
"eval_runtime": 776.0343, |
|
"eval_samples_per_second": 19.185, |
|
"eval_steps_per_second": 2.398, |
|
"step": 1415 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.7249743342399597, |
|
"learning_rate": 5.8e-05, |
|
"loss": 1.8578, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.9509921073913574, |
|
"eval_runtime": 786.6, |
|
"eval_samples_per_second": 18.927, |
|
"eval_steps_per_second": 2.366, |
|
"step": 1420 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.9510488510131836, |
|
"eval_runtime": 776.8377, |
|
"eval_samples_per_second": 19.165, |
|
"eval_steps_per_second": 2.396, |
|
"step": 1425 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.6929222941398621, |
|
"learning_rate": 5.6999999999999996e-05, |
|
"loss": 2.2087, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.9513132572174072, |
|
"eval_runtime": 759.5413, |
|
"eval_samples_per_second": 19.601, |
|
"eval_steps_per_second": 2.45, |
|
"step": 1430 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.9508575201034546, |
|
"eval_runtime": 762.6392, |
|
"eval_samples_per_second": 19.522, |
|
"eval_steps_per_second": 2.44, |
|
"step": 1435 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.7323941588401794, |
|
"learning_rate": 5.6000000000000006e-05, |
|
"loss": 2.1452, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.950052261352539, |
|
"eval_runtime": 744.7027, |
|
"eval_samples_per_second": 19.992, |
|
"eval_steps_per_second": 2.499, |
|
"step": 1440 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.9493366479873657, |
|
"eval_runtime": 740.8795, |
|
"eval_samples_per_second": 20.095, |
|
"eval_steps_per_second": 2.512, |
|
"step": 1445 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"grad_norm": 0.9700380563735962, |
|
"learning_rate": 5.500000000000001e-05, |
|
"loss": 1.945, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.33, |
|
"eval_loss": 1.9489691257476807, |
|
"eval_runtime": 740.6194, |
|
"eval_samples_per_second": 20.102, |
|
"eval_steps_per_second": 2.513, |
|
"step": 1450 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9490820169448853, |
|
"eval_runtime": 740.5722, |
|
"eval_samples_per_second": 20.103, |
|
"eval_steps_per_second": 2.513, |
|
"step": 1455 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.704833984375, |
|
"learning_rate": 5.4000000000000005e-05, |
|
"loss": 2.158, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.949270486831665, |
|
"eval_runtime": 743.446, |
|
"eval_samples_per_second": 20.026, |
|
"eval_steps_per_second": 2.503, |
|
"step": 1460 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9491441249847412, |
|
"eval_runtime": 768.2882, |
|
"eval_samples_per_second": 19.378, |
|
"eval_steps_per_second": 2.422, |
|
"step": 1465 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.7452517747879028, |
|
"learning_rate": 5.300000000000001e-05, |
|
"loss": 2.1391, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.949285864830017, |
|
"eval_runtime": 784.4371, |
|
"eval_samples_per_second": 18.979, |
|
"eval_steps_per_second": 2.372, |
|
"step": 1470 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9491676092147827, |
|
"eval_runtime": 779.509, |
|
"eval_samples_per_second": 19.099, |
|
"eval_steps_per_second": 2.387, |
|
"step": 1475 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 1.1882747411727905, |
|
"learning_rate": 5.2000000000000004e-05, |
|
"loss": 2.0945, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9490594863891602, |
|
"eval_runtime": 773.5132, |
|
"eval_samples_per_second": 19.247, |
|
"eval_steps_per_second": 2.406, |
|
"step": 1480 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9491394758224487, |
|
"eval_runtime": 769.1156, |
|
"eval_samples_per_second": 19.357, |
|
"eval_steps_per_second": 2.42, |
|
"step": 1485 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"grad_norm": 0.636154055595398, |
|
"learning_rate": 5.1000000000000006e-05, |
|
"loss": 1.812, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9490303993225098, |
|
"eval_runtime": 800.4965, |
|
"eval_samples_per_second": 18.598, |
|
"eval_steps_per_second": 2.325, |
|
"step": 1490 |
|
}, |
|
{ |
|
"epoch": 0.34, |
|
"eval_loss": 1.9486448764801025, |
|
"eval_runtime": 800.2384, |
|
"eval_samples_per_second": 18.604, |
|
"eval_steps_per_second": 2.326, |
|
"step": 1495 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.7470951676368713, |
|
"learning_rate": 5e-05, |
|
"loss": 2.0111, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9481943845748901, |
|
"eval_runtime": 799.7192, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9475739002227783, |
|
"eval_runtime": 799.7239, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1505 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.1213610172271729, |
|
"learning_rate": 4.9e-05, |
|
"loss": 2.2712, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9474812746047974, |
|
"eval_runtime": 799.7984, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1510 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.947679877281189, |
|
"eval_runtime": 799.9014, |
|
"eval_samples_per_second": 18.612, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1515 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.1590468883514404, |
|
"learning_rate": 4.8e-05, |
|
"loss": 2.0296, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.947396159172058, |
|
"eval_runtime": 799.8052, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1520 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9468810558319092, |
|
"eval_runtime": 799.9551, |
|
"eval_samples_per_second": 18.611, |
|
"eval_steps_per_second": 2.326, |
|
"step": 1525 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 0.7615826725959778, |
|
"learning_rate": 4.7e-05, |
|
"loss": 1.8507, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9464668035507202, |
|
"eval_runtime": 799.7366, |
|
"eval_samples_per_second": 18.616, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1530 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9460117816925049, |
|
"eval_runtime": 799.5384, |
|
"eval_samples_per_second": 18.621, |
|
"eval_steps_per_second": 2.328, |
|
"step": 1535 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"grad_norm": 1.255834698677063, |
|
"learning_rate": 4.600000000000001e-05, |
|
"loss": 1.9103, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.35, |
|
"eval_loss": 1.9458439350128174, |
|
"eval_runtime": 799.7058, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1540 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.9456100463867188, |
|
"eval_runtime": 799.6359, |
|
"eval_samples_per_second": 18.618, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1545 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.8211413621902466, |
|
"learning_rate": 4.5e-05, |
|
"loss": 1.9487, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.94538152217865, |
|
"eval_runtime": 799.8615, |
|
"eval_samples_per_second": 18.613, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1550 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.9452861547470093, |
|
"eval_runtime": 799.6185, |
|
"eval_samples_per_second": 18.619, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1555 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 1.0679750442504883, |
|
"learning_rate": 4.4000000000000006e-05, |
|
"loss": 2.0504, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.9451079368591309, |
|
"eval_runtime": 799.6427, |
|
"eval_samples_per_second": 18.618, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1560 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.945074439048767, |
|
"eval_runtime": 799.6799, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1565 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.6025732159614563, |
|
"learning_rate": 4.3e-05, |
|
"loss": 2.1497, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.9450632333755493, |
|
"eval_runtime": 799.6958, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1570 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.9450970888137817, |
|
"eval_runtime": 799.7035, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1575 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"grad_norm": 0.9095989465713501, |
|
"learning_rate": 4.2e-05, |
|
"loss": 1.9836, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.36, |
|
"eval_loss": 1.9450123310089111, |
|
"eval_runtime": 799.6334, |
|
"eval_samples_per_second": 18.619, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1580 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9449422359466553, |
|
"eval_runtime": 799.7912, |
|
"eval_samples_per_second": 18.615, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1585 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.714409351348877, |
|
"learning_rate": 4.1e-05, |
|
"loss": 1.8096, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.944868803024292, |
|
"eval_runtime": 799.8495, |
|
"eval_samples_per_second": 18.614, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1590 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9445998668670654, |
|
"eval_runtime": 799.4538, |
|
"eval_samples_per_second": 18.623, |
|
"eval_steps_per_second": 2.328, |
|
"step": 1595 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.5383341908454895, |
|
"learning_rate": 4e-05, |
|
"loss": 2.0759, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9442588090896606, |
|
"eval_runtime": 777.9784, |
|
"eval_samples_per_second": 19.137, |
|
"eval_steps_per_second": 2.392, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9438750743865967, |
|
"eval_runtime": 738.7615, |
|
"eval_samples_per_second": 20.153, |
|
"eval_steps_per_second": 2.519, |
|
"step": 1605 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.8301254510879517, |
|
"learning_rate": 3.9000000000000006e-05, |
|
"loss": 2.0214, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9435011148452759, |
|
"eval_runtime": 739.0062, |
|
"eval_samples_per_second": 20.146, |
|
"eval_steps_per_second": 2.518, |
|
"step": 1610 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9431562423706055, |
|
"eval_runtime": 738.2469, |
|
"eval_samples_per_second": 20.167, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1615 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"grad_norm": 0.9721934795379639, |
|
"learning_rate": 3.8e-05, |
|
"loss": 1.9382, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9430102109909058, |
|
"eval_runtime": 738.2353, |
|
"eval_samples_per_second": 20.167, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1620 |
|
}, |
|
{ |
|
"epoch": 0.37, |
|
"eval_loss": 1.9429086446762085, |
|
"eval_runtime": 738.1793, |
|
"eval_samples_per_second": 20.169, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1625 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9744455218315125, |
|
"learning_rate": 3.7e-05, |
|
"loss": 2.0086, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.942738652229309, |
|
"eval_runtime": 738.1789, |
|
"eval_samples_per_second": 20.169, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1630 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9425232410430908, |
|
"eval_runtime": 738.3004, |
|
"eval_samples_per_second": 20.165, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1635 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9008609652519226, |
|
"learning_rate": 3.6e-05, |
|
"loss": 2.1409, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9424079656600952, |
|
"eval_runtime": 738.2426, |
|
"eval_samples_per_second": 20.167, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1640 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9424738883972168, |
|
"eval_runtime": 738.2907, |
|
"eval_samples_per_second": 20.165, |
|
"eval_steps_per_second": 2.521, |
|
"step": 1645 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.7677657604217529, |
|
"learning_rate": 3.5e-05, |
|
"loss": 2.067, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9425007104873657, |
|
"eval_runtime": 742.0127, |
|
"eval_samples_per_second": 20.064, |
|
"eval_steps_per_second": 2.508, |
|
"step": 1650 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9425512552261353, |
|
"eval_runtime": 750.5258, |
|
"eval_samples_per_second": 19.837, |
|
"eval_steps_per_second": 2.48, |
|
"step": 1655 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.6732711791992188, |
|
"learning_rate": 3.4000000000000007e-05, |
|
"loss": 1.7463, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9426753520965576, |
|
"eval_runtime": 748.9114, |
|
"eval_samples_per_second": 19.88, |
|
"eval_steps_per_second": 2.485, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.942758560180664, |
|
"eval_runtime": 746.9293, |
|
"eval_samples_per_second": 19.932, |
|
"eval_steps_per_second": 2.492, |
|
"step": 1665 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"grad_norm": 0.9339891076087952, |
|
"learning_rate": 3.3e-05, |
|
"loss": 2.001, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.38, |
|
"eval_loss": 1.9427272081375122, |
|
"eval_runtime": 797.2689, |
|
"eval_samples_per_second": 18.674, |
|
"eval_steps_per_second": 2.334, |
|
"step": 1670 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9428359270095825, |
|
"eval_runtime": 800.5836, |
|
"eval_samples_per_second": 18.596, |
|
"eval_steps_per_second": 2.325, |
|
"step": 1675 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8810248374938965, |
|
"learning_rate": 3.2000000000000005e-05, |
|
"loss": 1.8671, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9427249431610107, |
|
"eval_runtime": 799.8884, |
|
"eval_samples_per_second": 18.613, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1680 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9424289464950562, |
|
"eval_runtime": 799.7098, |
|
"eval_samples_per_second": 18.617, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1685 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.9490606784820557, |
|
"learning_rate": 3.1e-05, |
|
"loss": 1.9094, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9420220851898193, |
|
"eval_runtime": 752.2407, |
|
"eval_samples_per_second": 19.792, |
|
"eval_steps_per_second": 2.474, |
|
"step": 1690 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9419386386871338, |
|
"eval_runtime": 738.4421, |
|
"eval_samples_per_second": 20.161, |
|
"eval_steps_per_second": 2.52, |
|
"step": 1695 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.7382946610450745, |
|
"learning_rate": 3e-05, |
|
"loss": 2.048, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9419728517532349, |
|
"eval_runtime": 738.4148, |
|
"eval_samples_per_second": 20.162, |
|
"eval_steps_per_second": 2.52, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9419692754745483, |
|
"eval_runtime": 744.3255, |
|
"eval_samples_per_second": 20.002, |
|
"eval_steps_per_second": 2.5, |
|
"step": 1705 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"grad_norm": 0.8462529182434082, |
|
"learning_rate": 2.9e-05, |
|
"loss": 1.9912, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9418373107910156, |
|
"eval_runtime": 745.3964, |
|
"eval_samples_per_second": 19.973, |
|
"eval_steps_per_second": 2.497, |
|
"step": 1710 |
|
}, |
|
{ |
|
"epoch": 0.39, |
|
"eval_loss": 1.9416935443878174, |
|
"eval_runtime": 745.4722, |
|
"eval_samples_per_second": 19.971, |
|
"eval_steps_per_second": 2.496, |
|
"step": 1715 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 1.0404905080795288, |
|
"learning_rate": 2.8000000000000003e-05, |
|
"loss": 2.1166, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.9415353536605835, |
|
"eval_runtime": 744.4599, |
|
"eval_samples_per_second": 19.998, |
|
"eval_steps_per_second": 2.5, |
|
"step": 1720 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.941475510597229, |
|
"eval_runtime": 747.4114, |
|
"eval_samples_per_second": 19.919, |
|
"eval_steps_per_second": 2.49, |
|
"step": 1725 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.9904809594154358, |
|
"learning_rate": 2.7000000000000002e-05, |
|
"loss": 1.8904, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.9413694143295288, |
|
"eval_runtime": 745.7476, |
|
"eval_samples_per_second": 19.964, |
|
"eval_steps_per_second": 2.495, |
|
"step": 1730 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.9412492513656616, |
|
"eval_runtime": 746.5069, |
|
"eval_samples_per_second": 19.944, |
|
"eval_steps_per_second": 2.493, |
|
"step": 1735 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8735015988349915, |
|
"learning_rate": 2.6000000000000002e-05, |
|
"loss": 1.926, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.9410345554351807, |
|
"eval_runtime": 745.8571, |
|
"eval_samples_per_second": 19.961, |
|
"eval_steps_per_second": 2.495, |
|
"step": 1740 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.940790057182312, |
|
"eval_runtime": 745.3703, |
|
"eval_samples_per_second": 19.974, |
|
"eval_steps_per_second": 2.497, |
|
"step": 1745 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"grad_norm": 0.8582627177238464, |
|
"learning_rate": 2.5e-05, |
|
"loss": 2.0692, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.9406797885894775, |
|
"eval_runtime": 759.2984, |
|
"eval_samples_per_second": 19.608, |
|
"eval_steps_per_second": 2.451, |
|
"step": 1750 |
|
}, |
|
{ |
|
"epoch": 0.4, |
|
"eval_loss": 1.9406994581222534, |
|
"eval_runtime": 799.6272, |
|
"eval_samples_per_second": 18.619, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1755 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.8529316782951355, |
|
"learning_rate": 2.4e-05, |
|
"loss": 2.0355, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9407473802566528, |
|
"eval_runtime": 799.5818, |
|
"eval_samples_per_second": 18.62, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1760 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9407432079315186, |
|
"eval_runtime": 799.2768, |
|
"eval_samples_per_second": 18.627, |
|
"eval_steps_per_second": 2.328, |
|
"step": 1765 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.9113526940345764, |
|
"learning_rate": 2.3000000000000003e-05, |
|
"loss": 1.9542, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9405388832092285, |
|
"eval_runtime": 799.4163, |
|
"eval_samples_per_second": 18.624, |
|
"eval_steps_per_second": 2.328, |
|
"step": 1770 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9403679370880127, |
|
"eval_runtime": 799.618, |
|
"eval_samples_per_second": 18.619, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1775 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.6997771859169006, |
|
"learning_rate": 2.2000000000000003e-05, |
|
"loss": 1.9599, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9402695894241333, |
|
"eval_runtime": 799.4944, |
|
"eval_samples_per_second": 18.622, |
|
"eval_steps_per_second": 2.328, |
|
"step": 1780 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9402552843093872, |
|
"eval_runtime": 799.5449, |
|
"eval_samples_per_second": 18.621, |
|
"eval_steps_per_second": 2.328, |
|
"step": 1785 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 0.6518684029579163, |
|
"learning_rate": 2.1e-05, |
|
"loss": 1.8072, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9402391910552979, |
|
"eval_runtime": 799.5838, |
|
"eval_samples_per_second": 18.62, |
|
"eval_steps_per_second": 2.327, |
|
"step": 1790 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9402625560760498, |
|
"eval_runtime": 764.0189, |
|
"eval_samples_per_second": 19.486, |
|
"eval_steps_per_second": 2.436, |
|
"step": 1795 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"grad_norm": 1.0698702335357666, |
|
"learning_rate": 2e-05, |
|
"loss": 2.0761, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.41, |
|
"eval_loss": 1.9402350187301636, |
|
"eval_runtime": 766.71, |
|
"eval_samples_per_second": 19.418, |
|
"eval_steps_per_second": 2.427, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.94016695022583, |
|
"eval_runtime": 767.2932, |
|
"eval_samples_per_second": 19.403, |
|
"eval_steps_per_second": 2.425, |
|
"step": 1805 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.5021382570266724, |
|
"learning_rate": 1.9e-05, |
|
"loss": 2.1248, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.9400701522827148, |
|
"eval_runtime": 751.0521, |
|
"eval_samples_per_second": 19.823, |
|
"eval_steps_per_second": 2.478, |
|
"step": 1810 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.9399473667144775, |
|
"eval_runtime": 781.0617, |
|
"eval_samples_per_second": 19.061, |
|
"eval_steps_per_second": 2.383, |
|
"step": 1815 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.5781351327896118, |
|
"learning_rate": 1.8e-05, |
|
"loss": 1.8231, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.9398096799850464, |
|
"eval_runtime": 786.8736, |
|
"eval_samples_per_second": 18.92, |
|
"eval_steps_per_second": 2.365, |
|
"step": 1820 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.93966805934906, |
|
"eval_runtime": 768.2598, |
|
"eval_samples_per_second": 19.379, |
|
"eval_steps_per_second": 2.422, |
|
"step": 1825 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 0.6509262919425964, |
|
"learning_rate": 1.7000000000000003e-05, |
|
"loss": 2.0489, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.9395979642868042, |
|
"eval_runtime": 751.1152, |
|
"eval_samples_per_second": 19.821, |
|
"eval_steps_per_second": 2.478, |
|
"step": 1830 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.9395313262939453, |
|
"eval_runtime": 759.4106, |
|
"eval_samples_per_second": 19.605, |
|
"eval_steps_per_second": 2.451, |
|
"step": 1835 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"grad_norm": 1.2815340757369995, |
|
"learning_rate": 1.6000000000000003e-05, |
|
"loss": 2.0146, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.939477562904358, |
|
"eval_runtime": 749.6182, |
|
"eval_samples_per_second": 19.861, |
|
"eval_steps_per_second": 2.483, |
|
"step": 1840 |
|
}, |
|
{ |
|
"epoch": 0.42, |
|
"eval_loss": 1.9394245147705078, |
|
"eval_runtime": 752.8584, |
|
"eval_samples_per_second": 19.775, |
|
"eval_steps_per_second": 2.472, |
|
"step": 1845 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.43942534923553467, |
|
"learning_rate": 1.5e-05, |
|
"loss": 1.942, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.9393796920776367, |
|
"eval_runtime": 759.4255, |
|
"eval_samples_per_second": 19.604, |
|
"eval_steps_per_second": 2.451, |
|
"step": 1850 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.9393460750579834, |
|
"eval_runtime": 804.7704, |
|
"eval_samples_per_second": 18.5, |
|
"eval_steps_per_second": 2.312, |
|
"step": 1855 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.7411811947822571, |
|
"learning_rate": 1.4000000000000001e-05, |
|
"loss": 2.1642, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.9392962455749512, |
|
"eval_runtime": 804.5633, |
|
"eval_samples_per_second": 18.504, |
|
"eval_steps_per_second": 2.313, |
|
"step": 1860 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.939192533493042, |
|
"eval_runtime": 795.592, |
|
"eval_samples_per_second": 18.713, |
|
"eval_steps_per_second": 2.339, |
|
"step": 1865 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.844436764717102, |
|
"learning_rate": 1.3000000000000001e-05, |
|
"loss": 1.8236, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.939056634902954, |
|
"eval_runtime": 752.4134, |
|
"eval_samples_per_second": 19.787, |
|
"eval_steps_per_second": 2.473, |
|
"step": 1870 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.9389359951019287, |
|
"eval_runtime": 748.7084, |
|
"eval_samples_per_second": 19.885, |
|
"eval_steps_per_second": 2.486, |
|
"step": 1875 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"grad_norm": 0.8722310662269592, |
|
"learning_rate": 1.2e-05, |
|
"loss": 2.134, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.9388750791549683, |
|
"eval_runtime": 751.6317, |
|
"eval_samples_per_second": 19.808, |
|
"eval_steps_per_second": 2.476, |
|
"step": 1880 |
|
}, |
|
{ |
|
"epoch": 0.43, |
|
"eval_loss": 1.9388281106948853, |
|
"eval_runtime": 754.3186, |
|
"eval_samples_per_second": 19.737, |
|
"eval_steps_per_second": 2.467, |
|
"step": 1885 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8985447287559509, |
|
"learning_rate": 1.1000000000000001e-05, |
|
"loss": 1.9291, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.938772439956665, |
|
"eval_runtime": 769.6876, |
|
"eval_samples_per_second": 19.343, |
|
"eval_steps_per_second": 2.418, |
|
"step": 1890 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.9387296438217163, |
|
"eval_runtime": 804.3454, |
|
"eval_samples_per_second": 18.509, |
|
"eval_steps_per_second": 2.314, |
|
"step": 1895 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 1.002416968345642, |
|
"learning_rate": 1e-05, |
|
"loss": 1.9172, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.938694953918457, |
|
"eval_runtime": 804.4582, |
|
"eval_samples_per_second": 18.507, |
|
"eval_steps_per_second": 2.313, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.9386274814605713, |
|
"eval_runtime": 804.296, |
|
"eval_samples_per_second": 18.511, |
|
"eval_steps_per_second": 2.314, |
|
"step": 1905 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.7875113487243652, |
|
"learning_rate": 9e-06, |
|
"loss": 2.0179, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.93858003616333, |
|
"eval_runtime": 767.3044, |
|
"eval_samples_per_second": 19.403, |
|
"eval_steps_per_second": 2.425, |
|
"step": 1910 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.9385550022125244, |
|
"eval_runtime": 753.0534, |
|
"eval_samples_per_second": 19.77, |
|
"eval_steps_per_second": 2.471, |
|
"step": 1915 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.8589398264884949, |
|
"learning_rate": 8.000000000000001e-06, |
|
"loss": 2.1198, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.9385381937026978, |
|
"eval_runtime": 752.7415, |
|
"eval_samples_per_second": 19.778, |
|
"eval_steps_per_second": 2.472, |
|
"step": 1920 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.9385038614273071, |
|
"eval_runtime": 751.344, |
|
"eval_samples_per_second": 19.815, |
|
"eval_steps_per_second": 2.477, |
|
"step": 1925 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"grad_norm": 0.7445068359375, |
|
"learning_rate": 7.000000000000001e-06, |
|
"loss": 1.7865, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.44, |
|
"eval_loss": 1.938494086265564, |
|
"eval_runtime": 754.4106, |
|
"eval_samples_per_second": 19.735, |
|
"eval_steps_per_second": 2.467, |
|
"step": 1930 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9384897947311401, |
|
"eval_runtime": 751.5617, |
|
"eval_samples_per_second": 19.809, |
|
"eval_steps_per_second": 2.476, |
|
"step": 1935 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.9787230491638184, |
|
"learning_rate": 6e-06, |
|
"loss": 1.8503, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9384899139404297, |
|
"eval_runtime": 762.4984, |
|
"eval_samples_per_second": 19.525, |
|
"eval_steps_per_second": 2.441, |
|
"step": 1940 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9384738206863403, |
|
"eval_runtime": 759.6919, |
|
"eval_samples_per_second": 19.597, |
|
"eval_steps_per_second": 2.45, |
|
"step": 1945 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.5382779240608215, |
|
"learning_rate": 5e-06, |
|
"loss": 1.9748, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9384503364562988, |
|
"eval_runtime": 765.0189, |
|
"eval_samples_per_second": 19.461, |
|
"eval_steps_per_second": 2.433, |
|
"step": 1950 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9384208917617798, |
|
"eval_runtime": 759.892, |
|
"eval_samples_per_second": 19.592, |
|
"eval_steps_per_second": 2.449, |
|
"step": 1955 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.9979357123374939, |
|
"learning_rate": 4.000000000000001e-06, |
|
"loss": 1.8969, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.938381314277649, |
|
"eval_runtime": 779.852, |
|
"eval_samples_per_second": 19.091, |
|
"eval_steps_per_second": 2.386, |
|
"step": 1960 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9383498430252075, |
|
"eval_runtime": 790.6207, |
|
"eval_samples_per_second": 18.831, |
|
"eval_steps_per_second": 2.354, |
|
"step": 1965 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"grad_norm": 0.6065915822982788, |
|
"learning_rate": 3e-06, |
|
"loss": 1.9022, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9383246898651123, |
|
"eval_runtime": 768.8896, |
|
"eval_samples_per_second": 19.363, |
|
"eval_steps_per_second": 2.42, |
|
"step": 1970 |
|
}, |
|
{ |
|
"epoch": 0.45, |
|
"eval_loss": 1.9383031129837036, |
|
"eval_runtime": 777.816, |
|
"eval_samples_per_second": 19.141, |
|
"eval_steps_per_second": 2.393, |
|
"step": 1975 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.9139683246612549, |
|
"learning_rate": 2.0000000000000003e-06, |
|
"loss": 1.9378, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"eval_loss": 1.938288927078247, |
|
"eval_runtime": 800.6894, |
|
"eval_samples_per_second": 18.594, |
|
"eval_steps_per_second": 2.324, |
|
"step": 1980 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"eval_loss": 1.9382776021957397, |
|
"eval_runtime": 796.1175, |
|
"eval_samples_per_second": 18.701, |
|
"eval_steps_per_second": 2.338, |
|
"step": 1985 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 1.0108898878097534, |
|
"learning_rate": 1.0000000000000002e-06, |
|
"loss": 2.0676, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"eval_loss": 1.938266396522522, |
|
"eval_runtime": 764.489, |
|
"eval_samples_per_second": 19.474, |
|
"eval_steps_per_second": 2.434, |
|
"step": 1990 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"eval_loss": 1.9382603168487549, |
|
"eval_runtime": 774.5865, |
|
"eval_samples_per_second": 19.221, |
|
"eval_steps_per_second": 2.403, |
|
"step": 1995 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"grad_norm": 0.8565602898597717, |
|
"learning_rate": 0.0, |
|
"loss": 1.9374, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.46, |
|
"eval_loss": 1.9382596015930176, |
|
"eval_runtime": 749.4454, |
|
"eval_samples_per_second": 19.865, |
|
"eval_steps_per_second": 2.483, |
|
"step": 2000 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 2000, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 500, |
|
"total_flos": 1244221462020096.0, |
|
"train_batch_size": 2, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|