|
{ |
|
"best_metric": 1.20897376537323, |
|
"best_model_checkpoint": "model_training/reprover_err/checkpoints-random-09-08-18-00/checkpoint-250", |
|
"epoch": 7.302231237322515, |
|
"eval_steps": 25, |
|
"global_step": 450, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.08113590263691683, |
|
"grad_norm": 23.601030349731445, |
|
"learning_rate": 5.102040816326531e-05, |
|
"loss": 12.5644, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.16227180527383367, |
|
"grad_norm": 10.2500638961792, |
|
"learning_rate": 0.00010204081632653062, |
|
"loss": 7.6824, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.2434077079107505, |
|
"grad_norm": 0.5588569641113281, |
|
"learning_rate": 0.00015306122448979594, |
|
"loss": 4.0204, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.32454361054766734, |
|
"grad_norm": 0.23660239577293396, |
|
"learning_rate": 0.00020408163265306123, |
|
"loss": 3.6085, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.4056795131845842, |
|
"grad_norm": 0.249436616897583, |
|
"learning_rate": 0.00025510204081632655, |
|
"loss": 3.5044, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.4056795131845842, |
|
"eval_loss": 3.385061025619507, |
|
"eval_runtime": 98.6159, |
|
"eval_samples_per_second": 10.14, |
|
"eval_steps_per_second": 0.639, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.486815415821501, |
|
"grad_norm": 1.356109857559204, |
|
"learning_rate": 0.0003061224489795919, |
|
"loss": 3.4346, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.5679513184584178, |
|
"grad_norm": 1.5906275510787964, |
|
"learning_rate": 0.00035714285714285714, |
|
"loss": 2.5708, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.6490872210953347, |
|
"grad_norm": 0.7469751834869385, |
|
"learning_rate": 0.00040816326530612246, |
|
"loss": 1.9541, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.7302231237322515, |
|
"grad_norm": 0.46847933530807495, |
|
"learning_rate": 0.0004591836734693878, |
|
"loss": 1.7316, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.8113590263691683, |
|
"grad_norm": 0.7208541035652161, |
|
"learning_rate": 0.0004999935985425297, |
|
"loss": 1.6138, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8113590263691683, |
|
"eval_loss": 1.387947678565979, |
|
"eval_runtime": 98.6024, |
|
"eval_samples_per_second": 10.142, |
|
"eval_steps_per_second": 0.639, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.8924949290060852, |
|
"grad_norm": 0.31425440311431885, |
|
"learning_rate": 0.0004997695819512612, |
|
"loss": 1.5905, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.973630831643002, |
|
"grad_norm": 0.4940365254878998, |
|
"learning_rate": 0.0004992258202402822, |
|
"loss": 1.5013, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 1.054766734279919, |
|
"grad_norm": 0.3743211328983307, |
|
"learning_rate": 0.0004983630095117843, |
|
"loss": 1.4282, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 1.1359026369168357, |
|
"grad_norm": 0.46512871980667114, |
|
"learning_rate": 0.0004971822543018662, |
|
"loss": 1.386, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 1.2170385395537526, |
|
"grad_norm": 0.2764676511287689, |
|
"learning_rate": 0.0004956850661665511, |
|
"loss": 1.4375, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.2170385395537526, |
|
"eval_loss": 1.2822275161743164, |
|
"eval_runtime": 98.6288, |
|
"eval_samples_per_second": 10.139, |
|
"eval_steps_per_second": 0.639, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 1.2981744421906694, |
|
"grad_norm": 0.2621470093727112, |
|
"learning_rate": 0.0004938733617467517, |
|
"loss": 1.4209, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 1.3793103448275863, |
|
"grad_norm": 0.14090226590633392, |
|
"learning_rate": 0.0004917494603146632, |
|
"loss": 1.3766, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 1.460446247464503, |
|
"grad_norm": 0.11969256401062012, |
|
"learning_rate": 0.0004893160808047222, |
|
"loss": 1.3755, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 1.5415821501014197, |
|
"grad_norm": 0.18346261978149414, |
|
"learning_rate": 0.00048657633833293557, |
|
"loss": 1.4168, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 1.6227180527383367, |
|
"grad_norm": 0.3624500036239624, |
|
"learning_rate": 0.0004835337402090316, |
|
"loss": 1.3587, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.6227180527383367, |
|
"eval_loss": 1.2857593297958374, |
|
"eval_runtime": 98.6659, |
|
"eval_samples_per_second": 10.135, |
|
"eval_steps_per_second": 0.639, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 1.7038539553752536, |
|
"grad_norm": 0.5834723114967346, |
|
"learning_rate": 0.0004801921814465414, |
|
"loss": 1.3857, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 1.7849898580121704, |
|
"grad_norm": 0.37480711936950684, |
|
"learning_rate": 0.00047655593977655674, |
|
"loss": 1.3745, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 1.866125760649087, |
|
"grad_norm": 0.31830307841300964, |
|
"learning_rate": 0.0004726296701715489, |
|
"loss": 1.3517, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 1.947261663286004, |
|
"grad_norm": 0.14451760053634644, |
|
"learning_rate": 0.00046841839888625623, |
|
"loss": 1.3346, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 2.028397565922921, |
|
"grad_norm": 0.22967207431793213, |
|
"learning_rate": 0.0004639275170232734, |
|
"loss": 1.3735, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.028397565922921, |
|
"eval_loss": 1.2710317373275757, |
|
"eval_runtime": 98.685, |
|
"eval_samples_per_second": 10.133, |
|
"eval_steps_per_second": 0.638, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 2.109533468559838, |
|
"grad_norm": 0.16856279969215393, |
|
"learning_rate": 0.0004591627736315743, |
|
"loss": 1.3417, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 2.1906693711967544, |
|
"grad_norm": 0.2670596241950989, |
|
"learning_rate": 0.0004541302683468084, |
|
"loss": 1.3135, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 2.2718052738336714, |
|
"grad_norm": 0.8827645778656006, |
|
"learning_rate": 0.0004488364435827881, |
|
"loss": 1.3399, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 2.3529411764705883, |
|
"grad_norm": 0.3872511088848114, |
|
"learning_rate": 0.00044328807628416644, |
|
"loss": 1.3888, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 2.4340770791075053, |
|
"grad_norm": 0.5543010234832764, |
|
"learning_rate": 0.0004374922692508611, |
|
"loss": 1.3678, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.4340770791075053, |
|
"eval_loss": 1.283129096031189, |
|
"eval_runtime": 98.6825, |
|
"eval_samples_per_second": 10.134, |
|
"eval_steps_per_second": 0.638, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 2.5152129817444218, |
|
"grad_norm": 0.3389835059642792, |
|
"learning_rate": 0.0004314564420453311, |
|
"loss": 1.3614, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 2.5963488843813387, |
|
"grad_norm": 0.38079357147216797, |
|
"learning_rate": 0.0004251883214943475, |
|
"loss": 1.342, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 2.6774847870182557, |
|
"grad_norm": 0.23523402214050293, |
|
"learning_rate": 0.0004186959317974155, |
|
"loss": 1.3378, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 2.7586206896551726, |
|
"grad_norm": 0.18377311527729034, |
|
"learning_rate": 0.00041198758425451266, |
|
"loss": 1.3272, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 2.839756592292089, |
|
"grad_norm": 0.4048319458961487, |
|
"learning_rate": 0.00040507186662629185, |
|
"loss": 1.3431, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.839756592292089, |
|
"eval_loss": 1.252098798751831, |
|
"eval_runtime": 98.7407, |
|
"eval_samples_per_second": 10.128, |
|
"eval_steps_per_second": 0.638, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 2.920892494929006, |
|
"grad_norm": 1.1706334352493286, |
|
"learning_rate": 0.0003979576321403705, |
|
"loss": 1.3111, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 3.002028397565923, |
|
"grad_norm": 0.3772040903568268, |
|
"learning_rate": 0.0003906539881577793, |
|
"loss": 1.3515, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 3.08316430020284, |
|
"grad_norm": 0.4847799241542816, |
|
"learning_rate": 0.0003831702845140801, |
|
"loss": 1.3442, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 3.1643002028397564, |
|
"grad_norm": 0.35955315828323364, |
|
"learning_rate": 0.00037551610155007613, |
|
"loss": 1.338, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 3.2454361054766734, |
|
"grad_norm": 0.30124932527542114, |
|
"learning_rate": 0.00036770123784744027, |
|
"loss": 1.3001, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.2454361054766734, |
|
"eval_loss": 1.2483381032943726, |
|
"eval_runtime": 98.609, |
|
"eval_samples_per_second": 10.141, |
|
"eval_steps_per_second": 0.639, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 3.3265720081135903, |
|
"grad_norm": 0.4706196188926697, |
|
"learning_rate": 0.00035973569768495855, |
|
"loss": 1.3099, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 3.4077079107505073, |
|
"grad_norm": 0.4052978754043579, |
|
"learning_rate": 0.0003516296782314491, |
|
"loss": 1.2863, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 3.4888438133874238, |
|
"grad_norm": 0.4775325655937195, |
|
"learning_rate": 0.00034339355649175095, |
|
"loss": 1.3306, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 3.5699797160243407, |
|
"grad_norm": 0.2688348591327667, |
|
"learning_rate": 0.00033503787602249364, |
|
"loss": 1.2971, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 3.6511156186612577, |
|
"grad_norm": 0.16574963927268982, |
|
"learning_rate": 0.00032657333343465356, |
|
"loss": 1.2761, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.6511156186612577, |
|
"eval_loss": 1.2276986837387085, |
|
"eval_runtime": 98.6417, |
|
"eval_samples_per_second": 10.138, |
|
"eval_steps_per_second": 0.639, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 3.732251521298174, |
|
"grad_norm": 0.16287527978420258, |
|
"learning_rate": 0.0003180107647001769, |
|
"loss": 1.2611, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 3.813387423935091, |
|
"grad_norm": 0.12902116775512695, |
|
"learning_rate": 0.0003093611312801979, |
|
"loss": 1.2933, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 3.894523326572008, |
|
"grad_norm": 0.19913575053215027, |
|
"learning_rate": 0.00030063550609261025, |
|
"loss": 1.2926, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 3.975659229208925, |
|
"grad_norm": 0.39822715520858765, |
|
"learning_rate": 0.000291845059336957, |
|
"loss": 1.3097, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 4.056795131845842, |
|
"grad_norm": 0.5150278806686401, |
|
"learning_rate": 0.0002830010441947834, |
|
"loss": 1.2907, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.056795131845842, |
|
"eval_loss": 1.20897376537323, |
|
"eval_runtime": 98.62, |
|
"eval_samples_per_second": 10.14, |
|
"eval_steps_per_second": 0.639, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 4.137931034482759, |
|
"grad_norm": 0.3091956675052643, |
|
"learning_rate": 0.00027411478242376017, |
|
"loss": 1.2904, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 4.219066937119676, |
|
"grad_norm": 0.2656278908252716, |
|
"learning_rate": 0.00026519764986401774, |
|
"loss": 1.2999, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 4.300202839756592, |
|
"grad_norm": 0.1790953278541565, |
|
"learning_rate": 0.000256261061875247, |
|
"loss": 1.2549, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 4.381338742393509, |
|
"grad_norm": 0.15690313279628754, |
|
"learning_rate": 0.0002473164587232079, |
|
"loss": 1.3, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 4.462474645030426, |
|
"grad_norm": 0.12892165780067444, |
|
"learning_rate": 0.0002383752909343547, |
|
"loss": 1.2946, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.462474645030426, |
|
"eval_loss": 1.2234361171722412, |
|
"eval_runtime": 98.5999, |
|
"eval_samples_per_second": 10.142, |
|
"eval_steps_per_second": 0.639, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 4.543610547667343, |
|
"grad_norm": 0.136368989944458, |
|
"learning_rate": 0.0002294490046373259, |
|
"loss": 1.2971, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 4.62474645030426, |
|
"grad_norm": 0.07033926248550415, |
|
"learning_rate": 0.00022054902691006405, |
|
"loss": 1.2991, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 4.705882352941177, |
|
"grad_norm": 0.08667398989200592, |
|
"learning_rate": 0.00021168675115132315, |
|
"loss": 1.291, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 4.787018255578094, |
|
"grad_norm": 0.08972382545471191, |
|
"learning_rate": 0.00020287352249529153, |
|
"loss": 1.3007, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 4.8681541582150105, |
|
"grad_norm": 0.1722905933856964, |
|
"learning_rate": 0.00019412062328800044, |
|
"loss": 1.2729, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.8681541582150105, |
|
"eval_loss": 1.2213460206985474, |
|
"eval_runtime": 98.5784, |
|
"eval_samples_per_second": 10.144, |
|
"eval_steps_per_second": 0.639, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 4.9492900608519275, |
|
"grad_norm": 0.10548505187034607, |
|
"learning_rate": 0.000185439258644112, |
|
"loss": 1.2854, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 5.0304259634888435, |
|
"grad_norm": 0.08486346155405045, |
|
"learning_rate": 0.00017684054210257517, |
|
"loss": 1.236, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 5.1115618661257605, |
|
"grad_norm": 0.10991238802671432, |
|
"learning_rate": 0.00016833548139951395, |
|
"loss": 1.2712, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 5.192697768762677, |
|
"grad_norm": 0.06149250268936157, |
|
"learning_rate": 0.0001599349643765599, |
|
"loss": 1.2807, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 5.273833671399594, |
|
"grad_norm": 0.17640946805477142, |
|
"learning_rate": 0.0001516497450426686, |
|
"loss": 1.3158, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.273833671399594, |
|
"eval_loss": 1.2180273532867432, |
|
"eval_runtime": 98.5761, |
|
"eval_samples_per_second": 10.144, |
|
"eval_steps_per_second": 0.639, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 5.354969574036511, |
|
"grad_norm": 0.46290215849876404, |
|
"learning_rate": 0.00014349042980726362, |
|
"loss": 1.2482, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 5.436105476673428, |
|
"grad_norm": 0.16729195415973663, |
|
"learning_rate": 0.0001354674639023318, |
|
"loss": 1.2903, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 5.517241379310345, |
|
"grad_norm": 0.24489770829677582, |
|
"learning_rate": 0.00012759111801085066, |
|
"loss": 1.2757, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 5.598377281947261, |
|
"grad_norm": 0.23532100021839142, |
|
"learning_rate": 0.00011987147511866788, |
|
"loss": 1.2928, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 5.679513184584178, |
|
"grad_norm": 0.07568836212158203, |
|
"learning_rate": 0.00011231841760666186, |
|
"loss": 1.2722, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.679513184584178, |
|
"eval_loss": 1.2159314155578613, |
|
"eval_runtime": 98.5804, |
|
"eval_samples_per_second": 10.144, |
|
"eval_steps_per_second": 0.639, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 5.760649087221095, |
|
"grad_norm": 0.07196550071239471, |
|
"learning_rate": 0.0001049416145997094, |
|
"loss": 1.2969, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 5.841784989858012, |
|
"grad_norm": 0.0853937566280365, |
|
"learning_rate": 9.775050958865584e-05, |
|
"loss": 1.306, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 5.922920892494929, |
|
"grad_norm": 0.05001223087310791, |
|
"learning_rate": 9.075430834113152e-05, |
|
"loss": 1.2304, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 6.004056795131846, |
|
"grad_norm": 0.058833617717027664, |
|
"learning_rate": 8.396196711669335e-05, |
|
"loss": 1.2746, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 6.085192697768763, |
|
"grad_norm": 0.07061880826950073, |
|
"learning_rate": 7.738218120137671e-05, |
|
"loss": 1.2522, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.085192697768763, |
|
"eval_loss": 1.2192745208740234, |
|
"eval_runtime": 98.5743, |
|
"eval_samples_per_second": 10.145, |
|
"eval_steps_per_second": 0.639, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 6.16632860040568, |
|
"grad_norm": 0.06949684768915176, |
|
"learning_rate": 7.102337377633394e-05, |
|
"loss": 1.2678, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 6.247464503042596, |
|
"grad_norm": 0.08404209464788437, |
|
"learning_rate": 6.489368513481228e-05, |
|
"loss": 1.2792, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 6.328600405679513, |
|
"grad_norm": 0.08185689151287079, |
|
"learning_rate": 5.9000962261273136e-05, |
|
"loss": 1.2779, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 6.40973630831643, |
|
"grad_norm": 0.04312283918261528, |
|
"learning_rate": 5.3352748785993164e-05, |
|
"loss": 1.3018, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 6.490872210953347, |
|
"grad_norm": 0.07406298071146011, |
|
"learning_rate": 4.795627532800806e-05, |
|
"loss": 1.2636, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.490872210953347, |
|
"eval_loss": 1.2143380641937256, |
|
"eval_runtime": 98.6065, |
|
"eval_samples_per_second": 10.141, |
|
"eval_steps_per_second": 0.639, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 6.572008113590264, |
|
"grad_norm": 0.043406181037425995, |
|
"learning_rate": 4.281845023876074e-05, |
|
"loss": 1.2578, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 6.653144016227181, |
|
"grad_norm": 0.07357024401426315, |
|
"learning_rate": 3.794585075830329e-05, |
|
"loss": 1.2472, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 6.734279918864098, |
|
"grad_norm": 0.05774565041065216, |
|
"learning_rate": 3.334471459537497e-05, |
|
"loss": 1.3082, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 6.8154158215010145, |
|
"grad_norm": 0.054679855704307556, |
|
"learning_rate": 2.902093194213526e-05, |
|
"loss": 1.2782, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 6.896551724137931, |
|
"grad_norm": 0.050430841743946075, |
|
"learning_rate": 2.4980037933772488e-05, |
|
"loss": 1.3043, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 6.896551724137931, |
|
"eval_loss": 1.2151918411254883, |
|
"eval_runtime": 98.562, |
|
"eval_samples_per_second": 10.146, |
|
"eval_steps_per_second": 0.639, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 6.9776876267748476, |
|
"grad_norm": 0.041136015206575394, |
|
"learning_rate": 2.122720556264357e-05, |
|
"loss": 1.2673, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 7.0588235294117645, |
|
"grad_norm": 0.03708193823695183, |
|
"learning_rate": 1.776723905601438e-05, |
|
"loss": 1.2496, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 7.139959432048681, |
|
"grad_norm": 0.055347807705402374, |
|
"learning_rate": 1.4604567725877926e-05, |
|
"loss": 1.2835, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 7.221095334685598, |
|
"grad_norm": 0.040714431554079056, |
|
"learning_rate": 1.1743240298725116e-05, |
|
"loss": 1.3163, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 7.302231237322515, |
|
"grad_norm": 0.03776278346776962, |
|
"learning_rate": 9.18691973252539e-06, |
|
"loss": 1.2647, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 7.302231237322515, |
|
"eval_loss": 1.2177557945251465, |
|
"eval_runtime": 98.5774, |
|
"eval_samples_per_second": 10.144, |
|
"eval_steps_per_second": 0.639, |
|
"step": 450 |
|
} |
|
], |
|
"logging_steps": 5, |
|
"max_steps": 488, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 8, |
|
"save_steps": 25, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 3.7034254328345395e+17, |
|
"train_batch_size": 16, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|