|
{ |
|
"best_metric": 1.2448393106460571, |
|
"best_model_checkpoint": "./lora_bn/checkpoint-1400", |
|
"epoch": 0.9003215434083601, |
|
"eval_steps": 200, |
|
"global_step": 1400, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.006430868167202572, |
|
"grad_norm": 1.3939015865325928, |
|
"learning_rate": 2.9999999999999997e-05, |
|
"loss": 3.6079, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.012861736334405145, |
|
"grad_norm": 1.4347032308578491, |
|
"learning_rate": 5.9999999999999995e-05, |
|
"loss": 3.3531, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.01929260450160772, |
|
"grad_norm": 1.10429847240448, |
|
"learning_rate": 8.999999999999999e-05, |
|
"loss": 2.7805, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.02572347266881029, |
|
"grad_norm": 0.6873103976249695, |
|
"learning_rate": 0.00011999999999999999, |
|
"loss": 2.3846, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.03215434083601286, |
|
"grad_norm": 0.749026894569397, |
|
"learning_rate": 0.00015, |
|
"loss": 2.2196, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.03858520900321544, |
|
"grad_norm": 0.6276043057441711, |
|
"learning_rate": 0.00017999999999999998, |
|
"loss": 2.0483, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.04501607717041801, |
|
"grad_norm": 0.7210396528244019, |
|
"learning_rate": 0.00020999999999999998, |
|
"loss": 1.8821, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.05144694533762058, |
|
"grad_norm": 0.6538994312286377, |
|
"learning_rate": 0.00023999999999999998, |
|
"loss": 1.7903, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.05787781350482315, |
|
"grad_norm": 0.7155820727348328, |
|
"learning_rate": 0.00027, |
|
"loss": 1.7298, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.06430868167202572, |
|
"grad_norm": 0.6406599283218384, |
|
"learning_rate": 0.0003, |
|
"loss": 1.6995, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.0707395498392283, |
|
"grad_norm": 0.5722670555114746, |
|
"learning_rate": 0.00029793814432989686, |
|
"loss": 1.6895, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.07717041800643087, |
|
"grad_norm": 0.624355137348175, |
|
"learning_rate": 0.0002958762886597938, |
|
"loss": 1.6073, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.08360128617363344, |
|
"grad_norm": 0.5818140506744385, |
|
"learning_rate": 0.0002938144329896907, |
|
"loss": 1.6228, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.09003215434083602, |
|
"grad_norm": 0.5287107229232788, |
|
"learning_rate": 0.0002917525773195876, |
|
"loss": 1.6014, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.09646302250803858, |
|
"grad_norm": 0.5458894968032837, |
|
"learning_rate": 0.0002896907216494845, |
|
"loss": 1.5688, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.10289389067524116, |
|
"grad_norm": 0.5477120876312256, |
|
"learning_rate": 0.0002876288659793814, |
|
"loss": 1.5029, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.10932475884244373, |
|
"grad_norm": 0.5350768566131592, |
|
"learning_rate": 0.0002855670103092783, |
|
"loss": 1.559, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.1157556270096463, |
|
"grad_norm": 0.5387317538261414, |
|
"learning_rate": 0.00028350515463917525, |
|
"loss": 1.5013, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.12218649517684887, |
|
"grad_norm": 0.5648557543754578, |
|
"learning_rate": 0.00028144329896907214, |
|
"loss": 1.5107, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.12861736334405144, |
|
"grad_norm": 0.5028502345085144, |
|
"learning_rate": 0.0002793814432989691, |
|
"loss": 1.525, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.12861736334405144, |
|
"eval_loss": 1.4951640367507935, |
|
"eval_runtime": 340.2657, |
|
"eval_samples_per_second": 5.878, |
|
"eval_steps_per_second": 0.735, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.13504823151125403, |
|
"grad_norm": 0.520252525806427, |
|
"learning_rate": 0.00027731958762886597, |
|
"loss": 1.5075, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.1414790996784566, |
|
"grad_norm": 0.5032276511192322, |
|
"learning_rate": 0.00027525773195876286, |
|
"loss": 1.4617, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.14790996784565916, |
|
"grad_norm": 0.47061678767204285, |
|
"learning_rate": 0.00027319587628865975, |
|
"loss": 1.4809, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.15434083601286175, |
|
"grad_norm": 0.4570547640323639, |
|
"learning_rate": 0.0002711340206185567, |
|
"loss": 1.4195, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.1607717041800643, |
|
"grad_norm": 0.49685975909233093, |
|
"learning_rate": 0.0002690721649484536, |
|
"loss": 1.4835, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.16720257234726688, |
|
"grad_norm": 0.5451533794403076, |
|
"learning_rate": 0.00026701030927835047, |
|
"loss": 1.4514, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.17363344051446947, |
|
"grad_norm": 0.472751647233963, |
|
"learning_rate": 0.0002649484536082474, |
|
"loss": 1.4538, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.18006430868167203, |
|
"grad_norm": 0.4936389923095703, |
|
"learning_rate": 0.0002628865979381443, |
|
"loss": 1.4562, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.1864951768488746, |
|
"grad_norm": 0.5179173350334167, |
|
"learning_rate": 0.00026082474226804124, |
|
"loss": 1.45, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.19292604501607716, |
|
"grad_norm": 0.48392704129219055, |
|
"learning_rate": 0.00025876288659793813, |
|
"loss": 1.4381, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.19935691318327975, |
|
"grad_norm": 0.4685576856136322, |
|
"learning_rate": 0.000256701030927835, |
|
"loss": 1.439, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.2057877813504823, |
|
"grad_norm": 0.5039055347442627, |
|
"learning_rate": 0.0002546391752577319, |
|
"loss": 1.3921, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.21221864951768488, |
|
"grad_norm": 0.49153512716293335, |
|
"learning_rate": 0.00025257731958762885, |
|
"loss": 1.4272, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.21864951768488747, |
|
"grad_norm": 0.48751088976860046, |
|
"learning_rate": 0.00025051546391752574, |
|
"loss": 1.3853, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.22508038585209003, |
|
"grad_norm": 0.4660916328430176, |
|
"learning_rate": 0.0002484536082474227, |
|
"loss": 1.3963, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.2315112540192926, |
|
"grad_norm": 0.44358980655670166, |
|
"learning_rate": 0.0002463917525773196, |
|
"loss": 1.4165, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.2379421221864952, |
|
"grad_norm": 0.49413594603538513, |
|
"learning_rate": 0.00024432989690721646, |
|
"loss": 1.3768, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.24437299035369775, |
|
"grad_norm": 0.44838714599609375, |
|
"learning_rate": 0.00024226804123711338, |
|
"loss": 1.3877, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.2508038585209003, |
|
"grad_norm": 0.46237707138061523, |
|
"learning_rate": 0.0002402061855670103, |
|
"loss": 1.3781, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.2572347266881029, |
|
"grad_norm": 0.49087774753570557, |
|
"learning_rate": 0.00023814432989690718, |
|
"loss": 1.4006, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.2572347266881029, |
|
"eval_loss": 1.3819265365600586, |
|
"eval_runtime": 340.2482, |
|
"eval_samples_per_second": 5.878, |
|
"eval_steps_per_second": 0.735, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.26366559485530544, |
|
"grad_norm": 0.47525086998939514, |
|
"learning_rate": 0.0002360824742268041, |
|
"loss": 1.3835, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.27009646302250806, |
|
"grad_norm": 0.4709095358848572, |
|
"learning_rate": 0.00023402061855670102, |
|
"loss": 1.4037, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.2765273311897106, |
|
"grad_norm": 0.4556897282600403, |
|
"learning_rate": 0.00023195876288659793, |
|
"loss": 1.4335, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.2829581993569132, |
|
"grad_norm": 0.5150389671325684, |
|
"learning_rate": 0.00022989690721649485, |
|
"loss": 1.3855, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.28938906752411575, |
|
"grad_norm": 0.4785289764404297, |
|
"learning_rate": 0.0002278350515463917, |
|
"loss": 1.3572, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.2958199356913183, |
|
"grad_norm": 0.4818146526813507, |
|
"learning_rate": 0.00022577319587628863, |
|
"loss": 1.4095, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.3022508038585209, |
|
"grad_norm": 0.44462355971336365, |
|
"learning_rate": 0.00022371134020618554, |
|
"loss": 1.3607, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.3086816720257235, |
|
"grad_norm": 0.4708723723888397, |
|
"learning_rate": 0.00022164948453608246, |
|
"loss": 1.3985, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 0.31511254019292606, |
|
"grad_norm": 0.46771880984306335, |
|
"learning_rate": 0.00021958762886597935, |
|
"loss": 1.364, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 0.3215434083601286, |
|
"grad_norm": 0.4421800971031189, |
|
"learning_rate": 0.00021752577319587626, |
|
"loss": 1.3938, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.3279742765273312, |
|
"grad_norm": 0.4229006767272949, |
|
"learning_rate": 0.00021546391752577318, |
|
"loss": 1.3547, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 0.33440514469453375, |
|
"grad_norm": 0.4854479730129242, |
|
"learning_rate": 0.0002134020618556701, |
|
"loss": 1.3448, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 0.3408360128617363, |
|
"grad_norm": 0.4800192713737488, |
|
"learning_rate": 0.000211340206185567, |
|
"loss": 1.3716, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 0.34726688102893893, |
|
"grad_norm": 0.4886428415775299, |
|
"learning_rate": 0.0002092783505154639, |
|
"loss": 1.381, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 0.3536977491961415, |
|
"grad_norm": 0.4030856788158417, |
|
"learning_rate": 0.0002072164948453608, |
|
"loss": 1.3242, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 0.36012861736334406, |
|
"grad_norm": 0.5465673208236694, |
|
"learning_rate": 0.0002051546391752577, |
|
"loss": 1.3309, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 0.3665594855305466, |
|
"grad_norm": 0.47386476397514343, |
|
"learning_rate": 0.00020309278350515462, |
|
"loss": 1.399, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 0.3729903536977492, |
|
"grad_norm": 0.4967169761657715, |
|
"learning_rate": 0.0002010309278350515, |
|
"loss": 1.3441, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 0.37942122186495175, |
|
"grad_norm": 0.44418269395828247, |
|
"learning_rate": 0.00019896907216494843, |
|
"loss": 1.3417, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 0.3858520900321543, |
|
"grad_norm": 0.447294682264328, |
|
"learning_rate": 0.00019690721649484534, |
|
"loss": 1.3011, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.3858520900321543, |
|
"eval_loss": 1.331955909729004, |
|
"eval_runtime": 340.7655, |
|
"eval_samples_per_second": 5.869, |
|
"eval_steps_per_second": 0.734, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.39228295819935693, |
|
"grad_norm": 0.46866482496261597, |
|
"learning_rate": 0.00019484536082474226, |
|
"loss": 1.3152, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 0.3987138263665595, |
|
"grad_norm": 0.4665449857711792, |
|
"learning_rate": 0.00019278350515463918, |
|
"loss": 1.3253, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 0.40514469453376206, |
|
"grad_norm": 0.4765641391277313, |
|
"learning_rate": 0.00019072164948453606, |
|
"loss": 1.3493, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 0.4115755627009646, |
|
"grad_norm": 0.49187275767326355, |
|
"learning_rate": 0.00018865979381443298, |
|
"loss": 1.3454, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 0.4180064308681672, |
|
"grad_norm": 0.4786391258239746, |
|
"learning_rate": 0.0001865979381443299, |
|
"loss": 1.3464, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 0.42443729903536975, |
|
"grad_norm": 0.5260841846466064, |
|
"learning_rate": 0.00018453608247422679, |
|
"loss": 1.3165, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 0.43086816720257237, |
|
"grad_norm": 0.47295427322387695, |
|
"learning_rate": 0.00018247422680412367, |
|
"loss": 1.3064, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 0.43729903536977494, |
|
"grad_norm": 0.49421828985214233, |
|
"learning_rate": 0.0001804123711340206, |
|
"loss": 1.3315, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 0.4437299035369775, |
|
"grad_norm": 0.4635757505893707, |
|
"learning_rate": 0.0001783505154639175, |
|
"loss": 1.3549, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 0.45016077170418006, |
|
"grad_norm": 0.47637075185775757, |
|
"learning_rate": 0.00017628865979381442, |
|
"loss": 1.3124, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.4565916398713826, |
|
"grad_norm": 0.442953884601593, |
|
"learning_rate": 0.00017422680412371134, |
|
"loss": 1.3702, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 0.4630225080385852, |
|
"grad_norm": 0.49394121766090393, |
|
"learning_rate": 0.00017216494845360823, |
|
"loss": 1.3104, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 0.4694533762057878, |
|
"grad_norm": 0.4604041576385498, |
|
"learning_rate": 0.00017010309278350514, |
|
"loss": 1.2666, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 0.4758842443729904, |
|
"grad_norm": 0.45597556233406067, |
|
"learning_rate": 0.00016804123711340206, |
|
"loss": 1.3049, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 0.48231511254019294, |
|
"grad_norm": 0.4338426887989044, |
|
"learning_rate": 0.00016597938144329898, |
|
"loss": 1.3581, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 0.4887459807073955, |
|
"grad_norm": 0.4623814821243286, |
|
"learning_rate": 0.00016391752577319584, |
|
"loss": 1.2777, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 0.49517684887459806, |
|
"grad_norm": 0.4995267689228058, |
|
"learning_rate": 0.00016185567010309275, |
|
"loss": 1.3087, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 0.5016077170418006, |
|
"grad_norm": 0.4469720721244812, |
|
"learning_rate": 0.00015979381443298967, |
|
"loss": 1.2893, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 0.5080385852090032, |
|
"grad_norm": 0.4728144407272339, |
|
"learning_rate": 0.00015773195876288659, |
|
"loss": 1.2496, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 0.5144694533762058, |
|
"grad_norm": 0.4181075096130371, |
|
"learning_rate": 0.0001556701030927835, |
|
"loss": 1.2912, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.5144694533762058, |
|
"eval_loss": 1.2991915941238403, |
|
"eval_runtime": 340.7607, |
|
"eval_samples_per_second": 5.869, |
|
"eval_steps_per_second": 0.734, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.5209003215434084, |
|
"grad_norm": 0.46053722500801086, |
|
"learning_rate": 0.0001536082474226804, |
|
"loss": 1.2695, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 0.5273311897106109, |
|
"grad_norm": 0.5585050582885742, |
|
"learning_rate": 0.0001515463917525773, |
|
"loss": 1.2968, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 0.5337620578778135, |
|
"grad_norm": 0.4734201431274414, |
|
"learning_rate": 0.00014948453608247422, |
|
"loss": 1.2478, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 0.5401929260450161, |
|
"grad_norm": 0.45543360710144043, |
|
"learning_rate": 0.0001474226804123711, |
|
"loss": 1.3217, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 0.5466237942122186, |
|
"grad_norm": 0.5416027903556824, |
|
"learning_rate": 0.00014536082474226803, |
|
"loss": 1.2522, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 0.5530546623794212, |
|
"grad_norm": 0.42681366205215454, |
|
"learning_rate": 0.00014329896907216494, |
|
"loss": 1.3112, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 0.5594855305466238, |
|
"grad_norm": 0.45384252071380615, |
|
"learning_rate": 0.00014123711340206183, |
|
"loss": 1.2793, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 0.5659163987138264, |
|
"grad_norm": 0.450888454914093, |
|
"learning_rate": 0.00013917525773195875, |
|
"loss": 1.3204, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 0.572347266881029, |
|
"grad_norm": 0.4524657428264618, |
|
"learning_rate": 0.00013711340206185566, |
|
"loss": 1.272, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 0.5787781350482315, |
|
"grad_norm": 0.46384352445602417, |
|
"learning_rate": 0.00013505154639175258, |
|
"loss": 1.2948, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.5852090032154341, |
|
"grad_norm": 0.4582608938217163, |
|
"learning_rate": 0.00013298969072164947, |
|
"loss": 1.3027, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 0.5916398713826366, |
|
"grad_norm": 0.4848160147666931, |
|
"learning_rate": 0.00013092783505154639, |
|
"loss": 1.3172, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 0.5980707395498392, |
|
"grad_norm": 0.484418660402298, |
|
"learning_rate": 0.00012886597938144327, |
|
"loss": 1.3117, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 0.6045016077170418, |
|
"grad_norm": 0.39582470059394836, |
|
"learning_rate": 0.0001268041237113402, |
|
"loss": 1.2813, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 0.6109324758842444, |
|
"grad_norm": 0.4482058882713318, |
|
"learning_rate": 0.0001247422680412371, |
|
"loss": 1.2639, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 0.617363344051447, |
|
"grad_norm": 0.49503323435783386, |
|
"learning_rate": 0.000122680412371134, |
|
"loss": 1.302, |
|
"step": 960 |
|
}, |
|
{ |
|
"epoch": 0.6237942122186495, |
|
"grad_norm": 0.42037469148635864, |
|
"learning_rate": 0.00012061855670103093, |
|
"loss": 1.2943, |
|
"step": 970 |
|
}, |
|
{ |
|
"epoch": 0.6302250803858521, |
|
"grad_norm": 0.449434757232666, |
|
"learning_rate": 0.00011855670103092781, |
|
"loss": 1.2941, |
|
"step": 980 |
|
}, |
|
{ |
|
"epoch": 0.6366559485530546, |
|
"grad_norm": 0.44711926579475403, |
|
"learning_rate": 0.00011649484536082473, |
|
"loss": 1.2906, |
|
"step": 990 |
|
}, |
|
{ |
|
"epoch": 0.6430868167202572, |
|
"grad_norm": 0.43769001960754395, |
|
"learning_rate": 0.00011443298969072163, |
|
"loss": 1.3039, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.6430868167202572, |
|
"eval_loss": 1.2755881547927856, |
|
"eval_runtime": 341.9539, |
|
"eval_samples_per_second": 5.849, |
|
"eval_steps_per_second": 0.731, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.6495176848874598, |
|
"grad_norm": 0.4710148870944977, |
|
"learning_rate": 0.00011237113402061855, |
|
"loss": 1.2645, |
|
"step": 1010 |
|
}, |
|
{ |
|
"epoch": 0.6559485530546624, |
|
"grad_norm": 0.44989457726478577, |
|
"learning_rate": 0.00011030927835051547, |
|
"loss": 1.282, |
|
"step": 1020 |
|
}, |
|
{ |
|
"epoch": 0.662379421221865, |
|
"grad_norm": 0.4453730285167694, |
|
"learning_rate": 0.00010824742268041235, |
|
"loss": 1.256, |
|
"step": 1030 |
|
}, |
|
{ |
|
"epoch": 0.6688102893890675, |
|
"grad_norm": 0.5148506164550781, |
|
"learning_rate": 0.00010618556701030927, |
|
"loss": 1.2541, |
|
"step": 1040 |
|
}, |
|
{ |
|
"epoch": 0.6752411575562701, |
|
"grad_norm": 0.43415772914886475, |
|
"learning_rate": 0.00010412371134020617, |
|
"loss": 1.2902, |
|
"step": 1050 |
|
}, |
|
{ |
|
"epoch": 0.6816720257234726, |
|
"grad_norm": 0.4594573974609375, |
|
"learning_rate": 0.00010206185567010309, |
|
"loss": 1.2425, |
|
"step": 1060 |
|
}, |
|
{ |
|
"epoch": 0.6881028938906752, |
|
"grad_norm": 0.4503769278526306, |
|
"learning_rate": 9.999999999999999e-05, |
|
"loss": 1.2896, |
|
"step": 1070 |
|
}, |
|
{ |
|
"epoch": 0.6945337620578779, |
|
"grad_norm": 0.4645858407020569, |
|
"learning_rate": 9.79381443298969e-05, |
|
"loss": 1.2735, |
|
"step": 1080 |
|
}, |
|
{ |
|
"epoch": 0.7009646302250804, |
|
"grad_norm": 0.48965829610824585, |
|
"learning_rate": 9.58762886597938e-05, |
|
"loss": 1.3047, |
|
"step": 1090 |
|
}, |
|
{ |
|
"epoch": 0.707395498392283, |
|
"grad_norm": 0.46829336881637573, |
|
"learning_rate": 9.381443298969071e-05, |
|
"loss": 1.2751, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.7138263665594855, |
|
"grad_norm": 0.4561706483364105, |
|
"learning_rate": 9.175257731958763e-05, |
|
"loss": 1.2858, |
|
"step": 1110 |
|
}, |
|
{ |
|
"epoch": 0.7202572347266881, |
|
"grad_norm": 0.4418432414531708, |
|
"learning_rate": 8.969072164948453e-05, |
|
"loss": 1.2254, |
|
"step": 1120 |
|
}, |
|
{ |
|
"epoch": 0.7266881028938906, |
|
"grad_norm": 0.46051761507987976, |
|
"learning_rate": 8.762886597938145e-05, |
|
"loss": 1.2198, |
|
"step": 1130 |
|
}, |
|
{ |
|
"epoch": 0.7331189710610932, |
|
"grad_norm": 0.48182615637779236, |
|
"learning_rate": 8.556701030927834e-05, |
|
"loss": 1.2631, |
|
"step": 1140 |
|
}, |
|
{ |
|
"epoch": 0.7395498392282959, |
|
"grad_norm": 0.42658165097236633, |
|
"learning_rate": 8.350515463917525e-05, |
|
"loss": 1.2667, |
|
"step": 1150 |
|
}, |
|
{ |
|
"epoch": 0.7459807073954984, |
|
"grad_norm": 0.48048514127731323, |
|
"learning_rate": 8.144329896907215e-05, |
|
"loss": 1.27, |
|
"step": 1160 |
|
}, |
|
{ |
|
"epoch": 0.752411575562701, |
|
"grad_norm": 0.4658808410167694, |
|
"learning_rate": 7.938144329896907e-05, |
|
"loss": 1.2778, |
|
"step": 1170 |
|
}, |
|
{ |
|
"epoch": 0.7588424437299035, |
|
"grad_norm": 0.47830724716186523, |
|
"learning_rate": 7.731958762886596e-05, |
|
"loss": 1.2327, |
|
"step": 1180 |
|
}, |
|
{ |
|
"epoch": 0.7652733118971061, |
|
"grad_norm": 0.4181123971939087, |
|
"learning_rate": 7.525773195876288e-05, |
|
"loss": 1.2559, |
|
"step": 1190 |
|
}, |
|
{ |
|
"epoch": 0.7717041800643086, |
|
"grad_norm": 0.46117720007896423, |
|
"learning_rate": 7.319587628865979e-05, |
|
"loss": 1.2572, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.7717041800643086, |
|
"eval_loss": 1.2570642232894897, |
|
"eval_runtime": 340.7404, |
|
"eval_samples_per_second": 5.87, |
|
"eval_steps_per_second": 0.734, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.7781350482315113, |
|
"grad_norm": 0.4613102078437805, |
|
"learning_rate": 7.11340206185567e-05, |
|
"loss": 1.2659, |
|
"step": 1210 |
|
}, |
|
{ |
|
"epoch": 0.7845659163987139, |
|
"grad_norm": 0.4278354346752167, |
|
"learning_rate": 6.90721649484536e-05, |
|
"loss": 1.2858, |
|
"step": 1220 |
|
}, |
|
{ |
|
"epoch": 0.7909967845659164, |
|
"grad_norm": 0.4664517343044281, |
|
"learning_rate": 6.701030927835051e-05, |
|
"loss": 1.293, |
|
"step": 1230 |
|
}, |
|
{ |
|
"epoch": 0.797427652733119, |
|
"grad_norm": 0.44390469789505005, |
|
"learning_rate": 6.494845360824742e-05, |
|
"loss": 1.2234, |
|
"step": 1240 |
|
}, |
|
{ |
|
"epoch": 0.8038585209003215, |
|
"grad_norm": 0.4815686345100403, |
|
"learning_rate": 6.288659793814433e-05, |
|
"loss": 1.2891, |
|
"step": 1250 |
|
}, |
|
{ |
|
"epoch": 0.8102893890675241, |
|
"grad_norm": 0.43138524889945984, |
|
"learning_rate": 6.0824742268041234e-05, |
|
"loss": 1.2344, |
|
"step": 1260 |
|
}, |
|
{ |
|
"epoch": 0.8167202572347267, |
|
"grad_norm": 0.49563485383987427, |
|
"learning_rate": 5.8762886597938136e-05, |
|
"loss": 1.2214, |
|
"step": 1270 |
|
}, |
|
{ |
|
"epoch": 0.8231511254019293, |
|
"grad_norm": 0.4346306324005127, |
|
"learning_rate": 5.6701030927835046e-05, |
|
"loss": 1.2804, |
|
"step": 1280 |
|
}, |
|
{ |
|
"epoch": 0.8295819935691319, |
|
"grad_norm": 0.4574083685874939, |
|
"learning_rate": 5.463917525773195e-05, |
|
"loss": 1.237, |
|
"step": 1290 |
|
}, |
|
{ |
|
"epoch": 0.8360128617363344, |
|
"grad_norm": 0.43636971712112427, |
|
"learning_rate": 5.257731958762886e-05, |
|
"loss": 1.218, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.842443729903537, |
|
"grad_norm": 0.411658376455307, |
|
"learning_rate": 5.051546391752577e-05, |
|
"loss": 1.2383, |
|
"step": 1310 |
|
}, |
|
{ |
|
"epoch": 0.8488745980707395, |
|
"grad_norm": 0.4198724627494812, |
|
"learning_rate": 4.8453608247422676e-05, |
|
"loss": 1.2599, |
|
"step": 1320 |
|
}, |
|
{ |
|
"epoch": 0.8553054662379421, |
|
"grad_norm": 0.4522201418876648, |
|
"learning_rate": 4.6391752577319585e-05, |
|
"loss": 1.2277, |
|
"step": 1330 |
|
}, |
|
{ |
|
"epoch": 0.8617363344051447, |
|
"grad_norm": 0.4966380298137665, |
|
"learning_rate": 4.4329896907216494e-05, |
|
"loss": 1.2268, |
|
"step": 1340 |
|
}, |
|
{ |
|
"epoch": 0.8681672025723473, |
|
"grad_norm": 0.4492814838886261, |
|
"learning_rate": 4.22680412371134e-05, |
|
"loss": 1.2589, |
|
"step": 1350 |
|
}, |
|
{ |
|
"epoch": 0.8745980707395499, |
|
"grad_norm": 0.4740350544452667, |
|
"learning_rate": 4.0206185567010306e-05, |
|
"loss": 1.2196, |
|
"step": 1360 |
|
}, |
|
{ |
|
"epoch": 0.8810289389067524, |
|
"grad_norm": 0.46403658390045166, |
|
"learning_rate": 3.814432989690721e-05, |
|
"loss": 1.2811, |
|
"step": 1370 |
|
}, |
|
{ |
|
"epoch": 0.887459807073955, |
|
"grad_norm": 0.46236127614974976, |
|
"learning_rate": 3.608247422680412e-05, |
|
"loss": 1.2403, |
|
"step": 1380 |
|
}, |
|
{ |
|
"epoch": 0.8938906752411575, |
|
"grad_norm": 0.44528549909591675, |
|
"learning_rate": 3.402061855670103e-05, |
|
"loss": 1.2374, |
|
"step": 1390 |
|
}, |
|
{ |
|
"epoch": 0.9003215434083601, |
|
"grad_norm": 0.4524483382701874, |
|
"learning_rate": 3.1958762886597937e-05, |
|
"loss": 1.2833, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.9003215434083601, |
|
"eval_loss": 1.2448393106460571, |
|
"eval_runtime": 340.581, |
|
"eval_samples_per_second": 5.872, |
|
"eval_steps_per_second": 0.734, |
|
"step": 1400 |
|
} |
|
], |
|
"logging_steps": 10, |
|
"max_steps": 1555, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 200, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": false |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 1.1406337433445335e+18, |
|
"train_batch_size": 8, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|