reward_modeling / checkpoint-2000 /trainer_state.json
hchang's picture
Upload folder using huggingface_hub
e937342 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.6657789613848203,
"eval_steps": 500,
"global_step": 2000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.003328894806924101,
"grad_norm": 37.75,
"learning_rate": 1.9933422103861518e-05,
"loss": 1.4084,
"step": 10
},
{
"epoch": 0.006657789613848202,
"grad_norm": 12.125,
"learning_rate": 1.9866844207723038e-05,
"loss": 1.0474,
"step": 20
},
{
"epoch": 0.009986684420772303,
"grad_norm": 45.25,
"learning_rate": 1.9800266311584554e-05,
"loss": 1.2455,
"step": 30
},
{
"epoch": 0.013315579227696404,
"grad_norm": 13.9375,
"learning_rate": 1.9733688415446073e-05,
"loss": 1.1969,
"step": 40
},
{
"epoch": 0.016644474034620507,
"grad_norm": 32.0,
"learning_rate": 1.966711051930759e-05,
"loss": 1.1659,
"step": 50
},
{
"epoch": 0.019973368841544607,
"grad_norm": 19.0,
"learning_rate": 1.960053262316911e-05,
"loss": 1.1159,
"step": 60
},
{
"epoch": 0.02330226364846871,
"grad_norm": 27.5,
"learning_rate": 1.953395472703063e-05,
"loss": 1.1861,
"step": 70
},
{
"epoch": 0.02663115845539281,
"grad_norm": 17.5,
"learning_rate": 1.9467376830892145e-05,
"loss": 1.1073,
"step": 80
},
{
"epoch": 0.02996005326231691,
"grad_norm": 33.75,
"learning_rate": 1.9400798934753665e-05,
"loss": 1.1506,
"step": 90
},
{
"epoch": 0.033288948069241014,
"grad_norm": 9.5,
"learning_rate": 1.933422103861518e-05,
"loss": 1.246,
"step": 100
},
{
"epoch": 0.03661784287616511,
"grad_norm": 11.3125,
"learning_rate": 1.92676431424767e-05,
"loss": 1.1708,
"step": 110
},
{
"epoch": 0.03994673768308921,
"grad_norm": 14.375,
"learning_rate": 1.9201065246338217e-05,
"loss": 1.1302,
"step": 120
},
{
"epoch": 0.043275632490013316,
"grad_norm": 20.625,
"learning_rate": 1.9134487350199737e-05,
"loss": 1.0537,
"step": 130
},
{
"epoch": 0.04660452729693742,
"grad_norm": 21.75,
"learning_rate": 1.9067909454061253e-05,
"loss": 1.0669,
"step": 140
},
{
"epoch": 0.049933422103861515,
"grad_norm": 14.125,
"learning_rate": 1.900133155792277e-05,
"loss": 1.0482,
"step": 150
},
{
"epoch": 0.05326231691078562,
"grad_norm": 27.625,
"learning_rate": 1.893475366178429e-05,
"loss": 1.1468,
"step": 160
},
{
"epoch": 0.05659121171770972,
"grad_norm": 4.53125,
"learning_rate": 1.8868175765645805e-05,
"loss": 1.092,
"step": 170
},
{
"epoch": 0.05992010652463382,
"grad_norm": 32.5,
"learning_rate": 1.8801597869507325e-05,
"loss": 1.2208,
"step": 180
},
{
"epoch": 0.06324900133155792,
"grad_norm": 56.0,
"learning_rate": 1.873501997336884e-05,
"loss": 1.2359,
"step": 190
},
{
"epoch": 0.06657789613848203,
"grad_norm": 8.375,
"learning_rate": 1.866844207723036e-05,
"loss": 1.0712,
"step": 200
},
{
"epoch": 0.06990679094540612,
"grad_norm": 26.375,
"learning_rate": 1.860186418109188e-05,
"loss": 1.0156,
"step": 210
},
{
"epoch": 0.07323568575233022,
"grad_norm": 12.0,
"learning_rate": 1.8535286284953397e-05,
"loss": 1.0407,
"step": 220
},
{
"epoch": 0.07656458055925433,
"grad_norm": 13.75,
"learning_rate": 1.8468708388814916e-05,
"loss": 1.0161,
"step": 230
},
{
"epoch": 0.07989347536617843,
"grad_norm": 27.125,
"learning_rate": 1.8402130492676432e-05,
"loss": 1.1466,
"step": 240
},
{
"epoch": 0.08322237017310254,
"grad_norm": 26.125,
"learning_rate": 1.8335552596537952e-05,
"loss": 1.3192,
"step": 250
},
{
"epoch": 0.08655126498002663,
"grad_norm": 25.75,
"learning_rate": 1.826897470039947e-05,
"loss": 1.0628,
"step": 260
},
{
"epoch": 0.08988015978695073,
"grad_norm": 29.25,
"learning_rate": 1.8202396804260988e-05,
"loss": 0.967,
"step": 270
},
{
"epoch": 0.09320905459387484,
"grad_norm": 29.75,
"learning_rate": 1.8135818908122504e-05,
"loss": 1.0356,
"step": 280
},
{
"epoch": 0.09653794940079893,
"grad_norm": 11.6875,
"learning_rate": 1.806924101198402e-05,
"loss": 0.9853,
"step": 290
},
{
"epoch": 0.09986684420772303,
"grad_norm": 27.5,
"learning_rate": 1.800266311584554e-05,
"loss": 1.0279,
"step": 300
},
{
"epoch": 0.10319573901464714,
"grad_norm": 5.65625,
"learning_rate": 1.7936085219707056e-05,
"loss": 1.0115,
"step": 310
},
{
"epoch": 0.10652463382157124,
"grad_norm": 8.1875,
"learning_rate": 1.7869507323568576e-05,
"loss": 0.942,
"step": 320
},
{
"epoch": 0.10985352862849534,
"grad_norm": 65.0,
"learning_rate": 1.7802929427430096e-05,
"loss": 1.0208,
"step": 330
},
{
"epoch": 0.11318242343541944,
"grad_norm": 9.625,
"learning_rate": 1.7736351531291612e-05,
"loss": 1.1729,
"step": 340
},
{
"epoch": 0.11651131824234354,
"grad_norm": 17.5,
"learning_rate": 1.766977363515313e-05,
"loss": 0.9322,
"step": 350
},
{
"epoch": 0.11984021304926765,
"grad_norm": 28.125,
"learning_rate": 1.7603195739014648e-05,
"loss": 1.0417,
"step": 360
},
{
"epoch": 0.12316910785619174,
"grad_norm": 18.25,
"learning_rate": 1.7536617842876168e-05,
"loss": 1.2491,
"step": 370
},
{
"epoch": 0.12649800266311584,
"grad_norm": 27.375,
"learning_rate": 1.7470039946737684e-05,
"loss": 0.9388,
"step": 380
},
{
"epoch": 0.12982689747003995,
"grad_norm": 22.125,
"learning_rate": 1.7403462050599203e-05,
"loss": 1.0488,
"step": 390
},
{
"epoch": 0.13315579227696406,
"grad_norm": 21.25,
"learning_rate": 1.733688415446072e-05,
"loss": 0.9951,
"step": 400
},
{
"epoch": 0.13648468708388814,
"grad_norm": 11.75,
"learning_rate": 1.727030625832224e-05,
"loss": 0.9212,
"step": 410
},
{
"epoch": 0.13981358189081225,
"grad_norm": 16.25,
"learning_rate": 1.7203728362183756e-05,
"loss": 1.0548,
"step": 420
},
{
"epoch": 0.14314247669773636,
"grad_norm": 19.875,
"learning_rate": 1.7137150466045275e-05,
"loss": 1.0238,
"step": 430
},
{
"epoch": 0.14647137150466044,
"grad_norm": 18.875,
"learning_rate": 1.707057256990679e-05,
"loss": 1.0327,
"step": 440
},
{
"epoch": 0.14980026631158455,
"grad_norm": 7.84375,
"learning_rate": 1.7003994673768308e-05,
"loss": 1.1155,
"step": 450
},
{
"epoch": 0.15312916111850866,
"grad_norm": 14.125,
"learning_rate": 1.693741677762983e-05,
"loss": 0.9627,
"step": 460
},
{
"epoch": 0.15645805592543274,
"grad_norm": 7.4375,
"learning_rate": 1.6870838881491347e-05,
"loss": 1.0216,
"step": 470
},
{
"epoch": 0.15978695073235685,
"grad_norm": 9.375,
"learning_rate": 1.6804260985352863e-05,
"loss": 1.0489,
"step": 480
},
{
"epoch": 0.16311584553928096,
"grad_norm": 19.5,
"learning_rate": 1.6737683089214383e-05,
"loss": 1.1254,
"step": 490
},
{
"epoch": 0.16644474034620507,
"grad_norm": 26.125,
"learning_rate": 1.66711051930759e-05,
"loss": 1.0134,
"step": 500
},
{
"epoch": 0.16644474034620507,
"eval_accuracy": 0.5015762402521985,
"eval_loss": 0.9942083358764648,
"eval_runtime": 211.8123,
"eval_samples_per_second": 113.818,
"eval_steps_per_second": 28.454,
"step": 500
},
{
"epoch": 0.16977363515312915,
"grad_norm": 12.3125,
"learning_rate": 1.660452729693742e-05,
"loss": 1.0164,
"step": 510
},
{
"epoch": 0.17310252996005326,
"grad_norm": 38.75,
"learning_rate": 1.6537949400798935e-05,
"loss": 1.0042,
"step": 520
},
{
"epoch": 0.17643142476697737,
"grad_norm": 20.125,
"learning_rate": 1.6471371504660455e-05,
"loss": 0.9786,
"step": 530
},
{
"epoch": 0.17976031957390146,
"grad_norm": 11.4375,
"learning_rate": 1.640479360852197e-05,
"loss": 0.9678,
"step": 540
},
{
"epoch": 0.18308921438082557,
"grad_norm": 19.375,
"learning_rate": 1.633821571238349e-05,
"loss": 1.0477,
"step": 550
},
{
"epoch": 0.18641810918774968,
"grad_norm": 16.75,
"learning_rate": 1.6271637816245007e-05,
"loss": 1.02,
"step": 560
},
{
"epoch": 0.18974700399467376,
"grad_norm": 15.5,
"learning_rate": 1.6205059920106527e-05,
"loss": 0.9864,
"step": 570
},
{
"epoch": 0.19307589880159787,
"grad_norm": 9.75,
"learning_rate": 1.6138482023968043e-05,
"loss": 1.0019,
"step": 580
},
{
"epoch": 0.19640479360852198,
"grad_norm": 12.125,
"learning_rate": 1.6071904127829563e-05,
"loss": 0.9483,
"step": 590
},
{
"epoch": 0.19973368841544606,
"grad_norm": 30.875,
"learning_rate": 1.6005326231691082e-05,
"loss": 1.0692,
"step": 600
},
{
"epoch": 0.20306258322237017,
"grad_norm": 10.125,
"learning_rate": 1.59387483355526e-05,
"loss": 0.9279,
"step": 610
},
{
"epoch": 0.20639147802929428,
"grad_norm": 11.6875,
"learning_rate": 1.5872170439414115e-05,
"loss": 0.941,
"step": 620
},
{
"epoch": 0.2097203728362184,
"grad_norm": 3.03125,
"learning_rate": 1.5805592543275634e-05,
"loss": 1.0259,
"step": 630
},
{
"epoch": 0.21304926764314247,
"grad_norm": 10.4375,
"learning_rate": 1.573901464713715e-05,
"loss": 0.9687,
"step": 640
},
{
"epoch": 0.21637816245006658,
"grad_norm": 12.8125,
"learning_rate": 1.567243675099867e-05,
"loss": 0.9512,
"step": 650
},
{
"epoch": 0.2197070572569907,
"grad_norm": 9.75,
"learning_rate": 1.5605858854860187e-05,
"loss": 0.9569,
"step": 660
},
{
"epoch": 0.22303595206391477,
"grad_norm": 17.375,
"learning_rate": 1.5539280958721706e-05,
"loss": 0.9694,
"step": 670
},
{
"epoch": 0.22636484687083888,
"grad_norm": 9.625,
"learning_rate": 1.5472703062583222e-05,
"loss": 0.97,
"step": 680
},
{
"epoch": 0.229693741677763,
"grad_norm": 21.0,
"learning_rate": 1.5406125166444742e-05,
"loss": 1.0506,
"step": 690
},
{
"epoch": 0.23302263648468707,
"grad_norm": 10.3125,
"learning_rate": 1.533954727030626e-05,
"loss": 0.9184,
"step": 700
},
{
"epoch": 0.23635153129161118,
"grad_norm": 21.5,
"learning_rate": 1.5272969374167778e-05,
"loss": 1.005,
"step": 710
},
{
"epoch": 0.2396804260985353,
"grad_norm": 4.40625,
"learning_rate": 1.5206391478029296e-05,
"loss": 0.8624,
"step": 720
},
{
"epoch": 0.24300932090545938,
"grad_norm": 13.9375,
"learning_rate": 1.5139813581890814e-05,
"loss": 1.0051,
"step": 730
},
{
"epoch": 0.24633821571238348,
"grad_norm": 16.625,
"learning_rate": 1.5073235685752332e-05,
"loss": 1.0378,
"step": 740
},
{
"epoch": 0.2496671105193076,
"grad_norm": 6.4375,
"learning_rate": 1.500665778961385e-05,
"loss": 0.9191,
"step": 750
},
{
"epoch": 0.2529960053262317,
"grad_norm": 5.5625,
"learning_rate": 1.4940079893475368e-05,
"loss": 0.9722,
"step": 760
},
{
"epoch": 0.2563249001331558,
"grad_norm": 22.625,
"learning_rate": 1.4873501997336886e-05,
"loss": 1.014,
"step": 770
},
{
"epoch": 0.2596537949400799,
"grad_norm": 21.0,
"learning_rate": 1.4806924101198404e-05,
"loss": 1.0986,
"step": 780
},
{
"epoch": 0.262982689747004,
"grad_norm": 10.4375,
"learning_rate": 1.4740346205059922e-05,
"loss": 0.9672,
"step": 790
},
{
"epoch": 0.2663115845539281,
"grad_norm": 17.0,
"learning_rate": 1.467376830892144e-05,
"loss": 0.9605,
"step": 800
},
{
"epoch": 0.26964047936085217,
"grad_norm": 31.75,
"learning_rate": 1.4607190412782957e-05,
"loss": 1.0308,
"step": 810
},
{
"epoch": 0.2729693741677763,
"grad_norm": 12.4375,
"learning_rate": 1.4540612516644474e-05,
"loss": 0.9268,
"step": 820
},
{
"epoch": 0.2762982689747004,
"grad_norm": 17.125,
"learning_rate": 1.4474034620505992e-05,
"loss": 0.9656,
"step": 830
},
{
"epoch": 0.2796271637816245,
"grad_norm": 4.53125,
"learning_rate": 1.440745672436751e-05,
"loss": 0.9865,
"step": 840
},
{
"epoch": 0.2829560585885486,
"grad_norm": 22.625,
"learning_rate": 1.434087882822903e-05,
"loss": 1.1118,
"step": 850
},
{
"epoch": 0.2862849533954727,
"grad_norm": 13.0,
"learning_rate": 1.4274300932090547e-05,
"loss": 1.0043,
"step": 860
},
{
"epoch": 0.28961384820239683,
"grad_norm": 9.0625,
"learning_rate": 1.4207723035952065e-05,
"loss": 0.9768,
"step": 870
},
{
"epoch": 0.2929427430093209,
"grad_norm": 6.75,
"learning_rate": 1.4141145139813583e-05,
"loss": 1.0237,
"step": 880
},
{
"epoch": 0.296271637816245,
"grad_norm": 10.875,
"learning_rate": 1.4074567243675101e-05,
"loss": 1.1597,
"step": 890
},
{
"epoch": 0.2996005326231691,
"grad_norm": 5.75,
"learning_rate": 1.4007989347536619e-05,
"loss": 0.7993,
"step": 900
},
{
"epoch": 0.3029294274300932,
"grad_norm": 31.5,
"learning_rate": 1.3941411451398137e-05,
"loss": 0.9802,
"step": 910
},
{
"epoch": 0.3062583222370173,
"grad_norm": 12.25,
"learning_rate": 1.3874833555259655e-05,
"loss": 0.9565,
"step": 920
},
{
"epoch": 0.30958721704394143,
"grad_norm": 11.8125,
"learning_rate": 1.3808255659121173e-05,
"loss": 1.0024,
"step": 930
},
{
"epoch": 0.3129161118508655,
"grad_norm": 16.625,
"learning_rate": 1.3741677762982691e-05,
"loss": 1.0209,
"step": 940
},
{
"epoch": 0.3162450066577896,
"grad_norm": 13.375,
"learning_rate": 1.3675099866844209e-05,
"loss": 0.8767,
"step": 950
},
{
"epoch": 0.3195739014647137,
"grad_norm": 20.875,
"learning_rate": 1.3608521970705725e-05,
"loss": 0.9642,
"step": 960
},
{
"epoch": 0.3229027962716378,
"grad_norm": 17.25,
"learning_rate": 1.3541944074567243e-05,
"loss": 0.9508,
"step": 970
},
{
"epoch": 0.3262316910785619,
"grad_norm": 22.25,
"learning_rate": 1.3475366178428764e-05,
"loss": 1.1324,
"step": 980
},
{
"epoch": 0.32956058588548603,
"grad_norm": 11.5,
"learning_rate": 1.3408788282290282e-05,
"loss": 1.0655,
"step": 990
},
{
"epoch": 0.33288948069241014,
"grad_norm": 15.625,
"learning_rate": 1.3342210386151799e-05,
"loss": 0.9608,
"step": 1000
},
{
"epoch": 0.33288948069241014,
"eval_accuracy": 0.507300481168077,
"eval_loss": 0.9467151165008545,
"eval_runtime": 211.5004,
"eval_samples_per_second": 113.986,
"eval_steps_per_second": 28.496,
"step": 1000
},
{
"epoch": 0.3362183754993342,
"grad_norm": 20.0,
"learning_rate": 1.3275632490013317e-05,
"loss": 1.0432,
"step": 1010
},
{
"epoch": 0.3395472703062583,
"grad_norm": 22.375,
"learning_rate": 1.3209054593874834e-05,
"loss": 1.0393,
"step": 1020
},
{
"epoch": 0.3428761651131824,
"grad_norm": 18.375,
"learning_rate": 1.3142476697736352e-05,
"loss": 1.0398,
"step": 1030
},
{
"epoch": 0.34620505992010653,
"grad_norm": 17.25,
"learning_rate": 1.307589880159787e-05,
"loss": 0.9225,
"step": 1040
},
{
"epoch": 0.34953395472703064,
"grad_norm": 11.75,
"learning_rate": 1.3009320905459388e-05,
"loss": 0.9262,
"step": 1050
},
{
"epoch": 0.35286284953395475,
"grad_norm": 10.25,
"learning_rate": 1.2942743009320906e-05,
"loss": 0.7248,
"step": 1060
},
{
"epoch": 0.3561917443408788,
"grad_norm": 12.875,
"learning_rate": 1.2876165113182424e-05,
"loss": 0.8358,
"step": 1070
},
{
"epoch": 0.3595206391478029,
"grad_norm": 4.65625,
"learning_rate": 1.2809587217043942e-05,
"loss": 0.8156,
"step": 1080
},
{
"epoch": 0.362849533954727,
"grad_norm": 23.5,
"learning_rate": 1.274300932090546e-05,
"loss": 1.0211,
"step": 1090
},
{
"epoch": 0.36617842876165113,
"grad_norm": 17.5,
"learning_rate": 1.2676431424766978e-05,
"loss": 0.9804,
"step": 1100
},
{
"epoch": 0.36950732356857524,
"grad_norm": 6.375,
"learning_rate": 1.2609853528628498e-05,
"loss": 0.9636,
"step": 1110
},
{
"epoch": 0.37283621837549935,
"grad_norm": 20.25,
"learning_rate": 1.2543275632490016e-05,
"loss": 0.9873,
"step": 1120
},
{
"epoch": 0.37616511318242346,
"grad_norm": 17.0,
"learning_rate": 1.2476697736351534e-05,
"loss": 0.9674,
"step": 1130
},
{
"epoch": 0.3794940079893475,
"grad_norm": 20.5,
"learning_rate": 1.241011984021305e-05,
"loss": 1.0073,
"step": 1140
},
{
"epoch": 0.3828229027962716,
"grad_norm": 4.125,
"learning_rate": 1.2343541944074568e-05,
"loss": 1.0919,
"step": 1150
},
{
"epoch": 0.38615179760319573,
"grad_norm": 15.3125,
"learning_rate": 1.2276964047936086e-05,
"loss": 1.0103,
"step": 1160
},
{
"epoch": 0.38948069241011984,
"grad_norm": 22.125,
"learning_rate": 1.2210386151797604e-05,
"loss": 0.9629,
"step": 1170
},
{
"epoch": 0.39280958721704395,
"grad_norm": 2.953125,
"learning_rate": 1.2143808255659122e-05,
"loss": 0.9545,
"step": 1180
},
{
"epoch": 0.39613848202396806,
"grad_norm": 7.59375,
"learning_rate": 1.207723035952064e-05,
"loss": 0.8836,
"step": 1190
},
{
"epoch": 0.3994673768308921,
"grad_norm": 28.125,
"learning_rate": 1.2010652463382158e-05,
"loss": 0.9632,
"step": 1200
},
{
"epoch": 0.40279627163781623,
"grad_norm": 15.125,
"learning_rate": 1.1944074567243676e-05,
"loss": 1.0765,
"step": 1210
},
{
"epoch": 0.40612516644474034,
"grad_norm": 11.625,
"learning_rate": 1.1877496671105194e-05,
"loss": 0.7911,
"step": 1220
},
{
"epoch": 0.40945406125166445,
"grad_norm": 15.5625,
"learning_rate": 1.1810918774966711e-05,
"loss": 1.0908,
"step": 1230
},
{
"epoch": 0.41278295605858856,
"grad_norm": 6.3125,
"learning_rate": 1.1744340878828231e-05,
"loss": 1.0418,
"step": 1240
},
{
"epoch": 0.41611185086551267,
"grad_norm": 2.671875,
"learning_rate": 1.1677762982689749e-05,
"loss": 0.9041,
"step": 1250
},
{
"epoch": 0.4194407456724368,
"grad_norm": 19.125,
"learning_rate": 1.1611185086551267e-05,
"loss": 1.0392,
"step": 1260
},
{
"epoch": 0.42276964047936083,
"grad_norm": 6.8125,
"learning_rate": 1.1544607190412785e-05,
"loss": 0.8439,
"step": 1270
},
{
"epoch": 0.42609853528628494,
"grad_norm": 7.75,
"learning_rate": 1.1478029294274303e-05,
"loss": 1.0188,
"step": 1280
},
{
"epoch": 0.42942743009320905,
"grad_norm": 17.25,
"learning_rate": 1.141145139813582e-05,
"loss": 1.1353,
"step": 1290
},
{
"epoch": 0.43275632490013316,
"grad_norm": 22.125,
"learning_rate": 1.1344873501997337e-05,
"loss": 0.987,
"step": 1300
},
{
"epoch": 0.43608521970705727,
"grad_norm": 16.0,
"learning_rate": 1.1278295605858855e-05,
"loss": 0.8446,
"step": 1310
},
{
"epoch": 0.4394141145139814,
"grad_norm": 48.25,
"learning_rate": 1.1211717709720373e-05,
"loss": 1.0588,
"step": 1320
},
{
"epoch": 0.44274300932090543,
"grad_norm": 17.625,
"learning_rate": 1.1145139813581891e-05,
"loss": 0.8579,
"step": 1330
},
{
"epoch": 0.44607190412782954,
"grad_norm": 20.25,
"learning_rate": 1.1078561917443409e-05,
"loss": 1.0973,
"step": 1340
},
{
"epoch": 0.44940079893475365,
"grad_norm": 14.25,
"learning_rate": 1.1011984021304927e-05,
"loss": 0.975,
"step": 1350
},
{
"epoch": 0.45272969374167776,
"grad_norm": 22.875,
"learning_rate": 1.0945406125166447e-05,
"loss": 0.9035,
"step": 1360
},
{
"epoch": 0.4560585885486019,
"grad_norm": 13.0,
"learning_rate": 1.0878828229027965e-05,
"loss": 1.0748,
"step": 1370
},
{
"epoch": 0.459387483355526,
"grad_norm": 13.125,
"learning_rate": 1.0812250332889482e-05,
"loss": 1.0373,
"step": 1380
},
{
"epoch": 0.4627163781624501,
"grad_norm": 15.0,
"learning_rate": 1.0745672436751e-05,
"loss": 0.9623,
"step": 1390
},
{
"epoch": 0.46604527296937415,
"grad_norm": 26.25,
"learning_rate": 1.0679094540612518e-05,
"loss": 0.9579,
"step": 1400
},
{
"epoch": 0.46937416777629826,
"grad_norm": 12.125,
"learning_rate": 1.0612516644474036e-05,
"loss": 0.918,
"step": 1410
},
{
"epoch": 0.47270306258322237,
"grad_norm": 30.75,
"learning_rate": 1.0545938748335554e-05,
"loss": 0.9823,
"step": 1420
},
{
"epoch": 0.4760319573901465,
"grad_norm": 7.0,
"learning_rate": 1.047936085219707e-05,
"loss": 0.9885,
"step": 1430
},
{
"epoch": 0.4793608521970706,
"grad_norm": 19.5,
"learning_rate": 1.0412782956058588e-05,
"loss": 0.9654,
"step": 1440
},
{
"epoch": 0.4826897470039947,
"grad_norm": 13.5,
"learning_rate": 1.0346205059920106e-05,
"loss": 0.8212,
"step": 1450
},
{
"epoch": 0.48601864181091875,
"grad_norm": 17.0,
"learning_rate": 1.0279627163781624e-05,
"loss": 0.9406,
"step": 1460
},
{
"epoch": 0.48934753661784286,
"grad_norm": 28.0,
"learning_rate": 1.0213049267643142e-05,
"loss": 0.9908,
"step": 1470
},
{
"epoch": 0.49267643142476697,
"grad_norm": 14.25,
"learning_rate": 1.014647137150466e-05,
"loss": 0.9418,
"step": 1480
},
{
"epoch": 0.4960053262316911,
"grad_norm": 12.25,
"learning_rate": 1.007989347536618e-05,
"loss": 1.075,
"step": 1490
},
{
"epoch": 0.4993342210386152,
"grad_norm": 10.625,
"learning_rate": 1.0013315579227698e-05,
"loss": 1.0111,
"step": 1500
},
{
"epoch": 0.4993342210386152,
"eval_accuracy": 0.508752281400365,
"eval_loss": 0.9421009421348572,
"eval_runtime": 209.8514,
"eval_samples_per_second": 114.881,
"eval_steps_per_second": 28.72,
"step": 1500
},
{
"epoch": 0.5026631158455392,
"grad_norm": 31.5,
"learning_rate": 9.946737683089214e-06,
"loss": 0.9077,
"step": 1510
},
{
"epoch": 0.5059920106524634,
"grad_norm": 19.375,
"learning_rate": 9.880159786950732e-06,
"loss": 1.1292,
"step": 1520
},
{
"epoch": 0.5093209054593875,
"grad_norm": 8.4375,
"learning_rate": 9.813581890812252e-06,
"loss": 1.0658,
"step": 1530
},
{
"epoch": 0.5126498002663116,
"grad_norm": 17.375,
"learning_rate": 9.74700399467377e-06,
"loss": 0.9691,
"step": 1540
},
{
"epoch": 0.5159786950732357,
"grad_norm": 8.125,
"learning_rate": 9.680426098535288e-06,
"loss": 1.0146,
"step": 1550
},
{
"epoch": 0.5193075898801598,
"grad_norm": 17.5,
"learning_rate": 9.613848202396806e-06,
"loss": 1.0196,
"step": 1560
},
{
"epoch": 0.5226364846870839,
"grad_norm": 15.1875,
"learning_rate": 9.547270306258324e-06,
"loss": 0.8996,
"step": 1570
},
{
"epoch": 0.525965379494008,
"grad_norm": 55.5,
"learning_rate": 9.48069241011984e-06,
"loss": 1.0211,
"step": 1580
},
{
"epoch": 0.5292942743009321,
"grad_norm": 4.21875,
"learning_rate": 9.41411451398136e-06,
"loss": 0.9728,
"step": 1590
},
{
"epoch": 0.5326231691078562,
"grad_norm": 6.71875,
"learning_rate": 9.347536617842877e-06,
"loss": 1.0168,
"step": 1600
},
{
"epoch": 0.5359520639147803,
"grad_norm": 24.75,
"learning_rate": 9.280958721704395e-06,
"loss": 0.9252,
"step": 1610
},
{
"epoch": 0.5392809587217043,
"grad_norm": 9.0625,
"learning_rate": 9.214380825565913e-06,
"loss": 1.0479,
"step": 1620
},
{
"epoch": 0.5426098535286284,
"grad_norm": 15.375,
"learning_rate": 9.147802929427431e-06,
"loss": 1.0699,
"step": 1630
},
{
"epoch": 0.5459387483355526,
"grad_norm": 17.5,
"learning_rate": 9.08122503328895e-06,
"loss": 0.9407,
"step": 1640
},
{
"epoch": 0.5492676431424767,
"grad_norm": 9.625,
"learning_rate": 9.014647137150465e-06,
"loss": 1.0124,
"step": 1650
},
{
"epoch": 0.5525965379494008,
"grad_norm": 10.375,
"learning_rate": 8.948069241011985e-06,
"loss": 0.8612,
"step": 1660
},
{
"epoch": 0.5559254327563249,
"grad_norm": 21.375,
"learning_rate": 8.881491344873503e-06,
"loss": 0.9613,
"step": 1670
},
{
"epoch": 0.559254327563249,
"grad_norm": 9.3125,
"learning_rate": 8.814913448735021e-06,
"loss": 0.9752,
"step": 1680
},
{
"epoch": 0.5625832223701731,
"grad_norm": 8.375,
"learning_rate": 8.748335552596539e-06,
"loss": 0.8926,
"step": 1690
},
{
"epoch": 0.5659121171770972,
"grad_norm": 17.625,
"learning_rate": 8.681757656458057e-06,
"loss": 1.0054,
"step": 1700
},
{
"epoch": 0.5692410119840213,
"grad_norm": 20.125,
"learning_rate": 8.615179760319575e-06,
"loss": 0.9203,
"step": 1710
},
{
"epoch": 0.5725699067909454,
"grad_norm": 28.75,
"learning_rate": 8.548601864181093e-06,
"loss": 0.9583,
"step": 1720
},
{
"epoch": 0.5758988015978695,
"grad_norm": 17.0,
"learning_rate": 8.48202396804261e-06,
"loss": 0.9582,
"step": 1730
},
{
"epoch": 0.5792276964047937,
"grad_norm": 13.4375,
"learning_rate": 8.415446071904129e-06,
"loss": 1.0034,
"step": 1740
},
{
"epoch": 0.5825565912117177,
"grad_norm": 27.75,
"learning_rate": 8.348868175765647e-06,
"loss": 1.0647,
"step": 1750
},
{
"epoch": 0.5858854860186418,
"grad_norm": 33.75,
"learning_rate": 8.282290279627165e-06,
"loss": 1.0575,
"step": 1760
},
{
"epoch": 0.5892143808255659,
"grad_norm": 14.375,
"learning_rate": 8.215712383488683e-06,
"loss": 1.0699,
"step": 1770
},
{
"epoch": 0.59254327563249,
"grad_norm": 6.09375,
"learning_rate": 8.1491344873502e-06,
"loss": 0.7706,
"step": 1780
},
{
"epoch": 0.5958721704394141,
"grad_norm": 8.5625,
"learning_rate": 8.082556591211719e-06,
"loss": 0.8626,
"step": 1790
},
{
"epoch": 0.5992010652463382,
"grad_norm": 7.46875,
"learning_rate": 8.015978695073236e-06,
"loss": 0.9427,
"step": 1800
},
{
"epoch": 0.6025299600532623,
"grad_norm": 23.375,
"learning_rate": 7.949400798934754e-06,
"loss": 0.962,
"step": 1810
},
{
"epoch": 0.6058588548601864,
"grad_norm": 7.65625,
"learning_rate": 7.882822902796272e-06,
"loss": 1.1216,
"step": 1820
},
{
"epoch": 0.6091877496671105,
"grad_norm": 10.0625,
"learning_rate": 7.81624500665779e-06,
"loss": 0.9076,
"step": 1830
},
{
"epoch": 0.6125166444740346,
"grad_norm": 6.78125,
"learning_rate": 7.749667110519308e-06,
"loss": 0.8158,
"step": 1840
},
{
"epoch": 0.6158455392809588,
"grad_norm": 29.5,
"learning_rate": 7.683089214380826e-06,
"loss": 0.9828,
"step": 1850
},
{
"epoch": 0.6191744340878829,
"grad_norm": 9.6875,
"learning_rate": 7.616511318242344e-06,
"loss": 0.9181,
"step": 1860
},
{
"epoch": 0.622503328894807,
"grad_norm": 7.6875,
"learning_rate": 7.549933422103862e-06,
"loss": 0.9864,
"step": 1870
},
{
"epoch": 0.625832223701731,
"grad_norm": 33.0,
"learning_rate": 7.48335552596538e-06,
"loss": 0.9898,
"step": 1880
},
{
"epoch": 0.6291611185086551,
"grad_norm": 13.1875,
"learning_rate": 7.416777629826898e-06,
"loss": 0.948,
"step": 1890
},
{
"epoch": 0.6324900133155792,
"grad_norm": 6.59375,
"learning_rate": 7.350199733688416e-06,
"loss": 0.9811,
"step": 1900
},
{
"epoch": 0.6358189081225033,
"grad_norm": 8.5625,
"learning_rate": 7.283621837549935e-06,
"loss": 1.0461,
"step": 1910
},
{
"epoch": 0.6391478029294274,
"grad_norm": 30.25,
"learning_rate": 7.217043941411453e-06,
"loss": 1.0408,
"step": 1920
},
{
"epoch": 0.6424766977363515,
"grad_norm": 25.625,
"learning_rate": 7.15046604527297e-06,
"loss": 1.0327,
"step": 1930
},
{
"epoch": 0.6458055925432756,
"grad_norm": 27.125,
"learning_rate": 7.083888149134488e-06,
"loss": 0.9213,
"step": 1940
},
{
"epoch": 0.6491344873501997,
"grad_norm": 22.125,
"learning_rate": 7.017310252996006e-06,
"loss": 1.1191,
"step": 1950
},
{
"epoch": 0.6524633821571239,
"grad_norm": 10.6875,
"learning_rate": 6.950732356857524e-06,
"loss": 0.9514,
"step": 1960
},
{
"epoch": 0.655792276964048,
"grad_norm": 16.375,
"learning_rate": 6.884154460719042e-06,
"loss": 0.9437,
"step": 1970
},
{
"epoch": 0.6591211717709721,
"grad_norm": 7.5625,
"learning_rate": 6.8175765645805605e-06,
"loss": 1.0645,
"step": 1980
},
{
"epoch": 0.6624500665778962,
"grad_norm": 9.125,
"learning_rate": 6.7509986684420784e-06,
"loss": 0.9085,
"step": 1990
},
{
"epoch": 0.6657789613848203,
"grad_norm": 9.6875,
"learning_rate": 6.6844207723035955e-06,
"loss": 0.9784,
"step": 2000
},
{
"epoch": 0.6657789613848203,
"eval_accuracy": 0.5068027210884354,
"eval_loss": 0.9414482712745667,
"eval_runtime": 213.4336,
"eval_samples_per_second": 112.953,
"eval_steps_per_second": 28.238,
"step": 2000
}
],
"logging_steps": 10,
"max_steps": 3004,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 0.0,
"train_batch_size": 16,
"trial_name": null,
"trial_params": null
}