whisper-small-ar_tsize_0.8 / trainer_state.json
otozz's picture
Upload 9 files
94572f3 verified
raw
history blame contribute delete
No virus
33.2 kB
{
"best_metric": 0.4691739691294871,
"best_model_checkpoint": "./whisper-small-ar_tsize_0.8/checkpoint-5000",
"epoch": 1.6103059581320451,
"eval_steps": 1000,
"global_step": 5000,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.01,
"grad_norm": 74.5753402709961,
"learning_rate": 5.000000000000001e-07,
"loss": 3.3084,
"step": 25
},
{
"epoch": 0.02,
"grad_norm": 23.875898361206055,
"learning_rate": 1.0000000000000002e-06,
"loss": 2.626,
"step": 50
},
{
"epoch": 0.02,
"grad_norm": 22.36936378479004,
"learning_rate": 1.5e-06,
"loss": 1.9536,
"step": 75
},
{
"epoch": 0.03,
"grad_norm": 18.3158016204834,
"learning_rate": 2.0000000000000003e-06,
"loss": 1.315,
"step": 100
},
{
"epoch": 0.04,
"grad_norm": 14.385392189025879,
"learning_rate": 2.5e-06,
"loss": 1.1058,
"step": 125
},
{
"epoch": 0.05,
"grad_norm": 14.316064834594727,
"learning_rate": 3e-06,
"loss": 1.0312,
"step": 150
},
{
"epoch": 0.06,
"grad_norm": 16.03868293762207,
"learning_rate": 3.5e-06,
"loss": 0.9277,
"step": 175
},
{
"epoch": 0.06,
"grad_norm": 16.684885025024414,
"learning_rate": 4.000000000000001e-06,
"loss": 0.867,
"step": 200
},
{
"epoch": 0.07,
"grad_norm": 14.64684772491455,
"learning_rate": 4.5e-06,
"loss": 0.8164,
"step": 225
},
{
"epoch": 0.08,
"grad_norm": 16.835796356201172,
"learning_rate": 5e-06,
"loss": 0.7126,
"step": 250
},
{
"epoch": 0.09,
"grad_norm": 17.927705764770508,
"learning_rate": 5.500000000000001e-06,
"loss": 0.6566,
"step": 275
},
{
"epoch": 0.1,
"grad_norm": 20.162878036499023,
"learning_rate": 6e-06,
"loss": 0.5585,
"step": 300
},
{
"epoch": 0.1,
"grad_norm": 15.939558982849121,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.4888,
"step": 325
},
{
"epoch": 0.11,
"grad_norm": 10.801262855529785,
"learning_rate": 7e-06,
"loss": 0.4609,
"step": 350
},
{
"epoch": 0.12,
"grad_norm": 14.835463523864746,
"learning_rate": 7.500000000000001e-06,
"loss": 0.4432,
"step": 375
},
{
"epoch": 0.13,
"grad_norm": 13.00851821899414,
"learning_rate": 8.000000000000001e-06,
"loss": 0.4361,
"step": 400
},
{
"epoch": 0.14,
"grad_norm": 15.16860294342041,
"learning_rate": 8.5e-06,
"loss": 0.4177,
"step": 425
},
{
"epoch": 0.14,
"grad_norm": 15.027597427368164,
"learning_rate": 9e-06,
"loss": 0.4447,
"step": 450
},
{
"epoch": 0.15,
"grad_norm": 15.810965538024902,
"learning_rate": 9.5e-06,
"loss": 0.4198,
"step": 475
},
{
"epoch": 0.16,
"grad_norm": 16.27136993408203,
"learning_rate": 1e-05,
"loss": 0.4265,
"step": 500
},
{
"epoch": 0.17,
"grad_norm": 12.16751766204834,
"learning_rate": 9.944444444444445e-06,
"loss": 0.4191,
"step": 525
},
{
"epoch": 0.18,
"grad_norm": 10.931772232055664,
"learning_rate": 9.88888888888889e-06,
"loss": 0.3924,
"step": 550
},
{
"epoch": 0.19,
"grad_norm": 12.04465389251709,
"learning_rate": 9.833333333333333e-06,
"loss": 0.4176,
"step": 575
},
{
"epoch": 0.19,
"grad_norm": 12.589746475219727,
"learning_rate": 9.777777777777779e-06,
"loss": 0.4358,
"step": 600
},
{
"epoch": 0.2,
"grad_norm": 14.16213607788086,
"learning_rate": 9.722222222222223e-06,
"loss": 0.3828,
"step": 625
},
{
"epoch": 0.21,
"grad_norm": 15.552674293518066,
"learning_rate": 9.666666666666667e-06,
"loss": 0.4139,
"step": 650
},
{
"epoch": 0.22,
"grad_norm": 12.745460510253906,
"learning_rate": 9.611111111111112e-06,
"loss": 0.3915,
"step": 675
},
{
"epoch": 0.23,
"grad_norm": 10.602766990661621,
"learning_rate": 9.555555555555556e-06,
"loss": 0.365,
"step": 700
},
{
"epoch": 0.23,
"grad_norm": 14.100479125976562,
"learning_rate": 9.5e-06,
"loss": 0.4218,
"step": 725
},
{
"epoch": 0.24,
"grad_norm": 11.610438346862793,
"learning_rate": 9.444444444444445e-06,
"loss": 0.3358,
"step": 750
},
{
"epoch": 0.25,
"grad_norm": 11.454590797424316,
"learning_rate": 9.38888888888889e-06,
"loss": 0.4012,
"step": 775
},
{
"epoch": 0.26,
"grad_norm": 12.808696746826172,
"learning_rate": 9.333333333333334e-06,
"loss": 0.3754,
"step": 800
},
{
"epoch": 0.27,
"grad_norm": 10.447662353515625,
"learning_rate": 9.277777777777778e-06,
"loss": 0.3407,
"step": 825
},
{
"epoch": 0.27,
"grad_norm": 9.822239875793457,
"learning_rate": 9.222222222222224e-06,
"loss": 0.3557,
"step": 850
},
{
"epoch": 0.28,
"grad_norm": 10.579741477966309,
"learning_rate": 9.166666666666666e-06,
"loss": 0.3557,
"step": 875
},
{
"epoch": 0.29,
"grad_norm": 13.126590728759766,
"learning_rate": 9.111111111111112e-06,
"loss": 0.3588,
"step": 900
},
{
"epoch": 0.3,
"grad_norm": 10.920930862426758,
"learning_rate": 9.055555555555556e-06,
"loss": 0.3761,
"step": 925
},
{
"epoch": 0.31,
"grad_norm": 16.436307907104492,
"learning_rate": 9e-06,
"loss": 0.3613,
"step": 950
},
{
"epoch": 0.31,
"grad_norm": 16.763071060180664,
"learning_rate": 8.944444444444446e-06,
"loss": 0.3718,
"step": 975
},
{
"epoch": 0.32,
"grad_norm": 7.7616376876831055,
"learning_rate": 8.888888888888888e-06,
"loss": 0.2768,
"step": 1000
},
{
"epoch": 0.32,
"eval_loss": 0.3424592614173889,
"eval_runtime": 2548.0078,
"eval_samples_per_second": 3.046,
"eval_steps_per_second": 0.381,
"eval_wer": 0.5745963257862194,
"step": 1000
},
{
"epoch": 0.33,
"grad_norm": 9.518533706665039,
"learning_rate": 8.833333333333334e-06,
"loss": 0.3837,
"step": 1025
},
{
"epoch": 0.34,
"grad_norm": 15.829094886779785,
"learning_rate": 8.777777777777778e-06,
"loss": 0.324,
"step": 1050
},
{
"epoch": 0.35,
"grad_norm": 12.936345100402832,
"learning_rate": 8.722222222222224e-06,
"loss": 0.3251,
"step": 1075
},
{
"epoch": 0.35,
"grad_norm": 10.695411682128906,
"learning_rate": 8.666666666666668e-06,
"loss": 0.3596,
"step": 1100
},
{
"epoch": 0.36,
"grad_norm": 10.362096786499023,
"learning_rate": 8.611111111111112e-06,
"loss": 0.3336,
"step": 1125
},
{
"epoch": 0.37,
"grad_norm": 11.786664009094238,
"learning_rate": 8.555555555555556e-06,
"loss": 0.3427,
"step": 1150
},
{
"epoch": 0.38,
"grad_norm": 10.013471603393555,
"learning_rate": 8.5e-06,
"loss": 0.3265,
"step": 1175
},
{
"epoch": 0.39,
"grad_norm": 11.64340877532959,
"learning_rate": 8.444444444444446e-06,
"loss": 0.3561,
"step": 1200
},
{
"epoch": 0.39,
"grad_norm": 11.622845649719238,
"learning_rate": 8.38888888888889e-06,
"loss": 0.3457,
"step": 1225
},
{
"epoch": 0.4,
"grad_norm": 13.175597190856934,
"learning_rate": 8.333333333333334e-06,
"loss": 0.3605,
"step": 1250
},
{
"epoch": 0.41,
"grad_norm": 13.138216018676758,
"learning_rate": 8.277777777777778e-06,
"loss": 0.3493,
"step": 1275
},
{
"epoch": 0.42,
"grad_norm": 13.549818992614746,
"learning_rate": 8.222222222222222e-06,
"loss": 0.2964,
"step": 1300
},
{
"epoch": 0.43,
"grad_norm": 11.058015823364258,
"learning_rate": 8.166666666666668e-06,
"loss": 0.3219,
"step": 1325
},
{
"epoch": 0.43,
"grad_norm": 13.490798950195312,
"learning_rate": 8.111111111111112e-06,
"loss": 0.2941,
"step": 1350
},
{
"epoch": 0.44,
"grad_norm": 9.73831844329834,
"learning_rate": 8.055555555555557e-06,
"loss": 0.329,
"step": 1375
},
{
"epoch": 0.45,
"grad_norm": 6.883594512939453,
"learning_rate": 8.000000000000001e-06,
"loss": 0.3076,
"step": 1400
},
{
"epoch": 0.46,
"grad_norm": 12.517281532287598,
"learning_rate": 7.944444444444445e-06,
"loss": 0.3153,
"step": 1425
},
{
"epoch": 0.47,
"grad_norm": 7.685423851013184,
"learning_rate": 7.88888888888889e-06,
"loss": 0.2991,
"step": 1450
},
{
"epoch": 0.48,
"grad_norm": 9.92092514038086,
"learning_rate": 7.833333333333333e-06,
"loss": 0.3064,
"step": 1475
},
{
"epoch": 0.48,
"grad_norm": 13.166866302490234,
"learning_rate": 7.77777777777778e-06,
"loss": 0.3195,
"step": 1500
},
{
"epoch": 0.49,
"grad_norm": 7.179815292358398,
"learning_rate": 7.722222222222223e-06,
"loss": 0.3214,
"step": 1525
},
{
"epoch": 0.5,
"grad_norm": 12.424599647521973,
"learning_rate": 7.666666666666667e-06,
"loss": 0.3062,
"step": 1550
},
{
"epoch": 0.51,
"grad_norm": 13.103777885437012,
"learning_rate": 7.611111111111111e-06,
"loss": 0.2858,
"step": 1575
},
{
"epoch": 0.52,
"grad_norm": 8.597440719604492,
"learning_rate": 7.555555555555556e-06,
"loss": 0.3194,
"step": 1600
},
{
"epoch": 0.52,
"grad_norm": 10.370293617248535,
"learning_rate": 7.500000000000001e-06,
"loss": 0.3285,
"step": 1625
},
{
"epoch": 0.53,
"grad_norm": 10.53341007232666,
"learning_rate": 7.444444444444445e-06,
"loss": 0.2848,
"step": 1650
},
{
"epoch": 0.54,
"grad_norm": 13.32474422454834,
"learning_rate": 7.38888888888889e-06,
"loss": 0.304,
"step": 1675
},
{
"epoch": 0.55,
"grad_norm": 10.368809700012207,
"learning_rate": 7.333333333333333e-06,
"loss": 0.3135,
"step": 1700
},
{
"epoch": 0.56,
"grad_norm": 13.282475471496582,
"learning_rate": 7.277777777777778e-06,
"loss": 0.2797,
"step": 1725
},
{
"epoch": 0.56,
"grad_norm": 10.41550350189209,
"learning_rate": 7.222222222222223e-06,
"loss": 0.3019,
"step": 1750
},
{
"epoch": 0.57,
"grad_norm": 12.789618492126465,
"learning_rate": 7.166666666666667e-06,
"loss": 0.3085,
"step": 1775
},
{
"epoch": 0.58,
"grad_norm": 6.927020072937012,
"learning_rate": 7.111111111111112e-06,
"loss": 0.271,
"step": 1800
},
{
"epoch": 0.59,
"grad_norm": 14.091415405273438,
"learning_rate": 7.055555555555557e-06,
"loss": 0.2985,
"step": 1825
},
{
"epoch": 0.6,
"grad_norm": 8.496126174926758,
"learning_rate": 7e-06,
"loss": 0.3029,
"step": 1850
},
{
"epoch": 0.6,
"grad_norm": 5.859543323516846,
"learning_rate": 6.944444444444445e-06,
"loss": 0.271,
"step": 1875
},
{
"epoch": 0.61,
"grad_norm": 16.182022094726562,
"learning_rate": 6.88888888888889e-06,
"loss": 0.3039,
"step": 1900
},
{
"epoch": 0.62,
"grad_norm": 6.995250225067139,
"learning_rate": 6.833333333333334e-06,
"loss": 0.2867,
"step": 1925
},
{
"epoch": 0.63,
"grad_norm": 11.519378662109375,
"learning_rate": 6.777777777777779e-06,
"loss": 0.3339,
"step": 1950
},
{
"epoch": 0.64,
"grad_norm": 10.880196571350098,
"learning_rate": 6.7222222222222235e-06,
"loss": 0.2873,
"step": 1975
},
{
"epoch": 0.64,
"grad_norm": 16.335350036621094,
"learning_rate": 6.666666666666667e-06,
"loss": 0.2977,
"step": 2000
},
{
"epoch": 0.64,
"eval_loss": 0.2855232059955597,
"eval_runtime": 2086.2971,
"eval_samples_per_second": 3.72,
"eval_steps_per_second": 0.465,
"eval_wer": 0.5296694986877808,
"step": 2000
},
{
"epoch": 0.65,
"grad_norm": 12.254558563232422,
"learning_rate": 6.6111111111111115e-06,
"loss": 0.2668,
"step": 2025
},
{
"epoch": 0.66,
"grad_norm": 14.6058931350708,
"learning_rate": 6.555555555555556e-06,
"loss": 0.3302,
"step": 2050
},
{
"epoch": 0.67,
"grad_norm": 9.349474906921387,
"learning_rate": 6.5000000000000004e-06,
"loss": 0.3228,
"step": 2075
},
{
"epoch": 0.68,
"grad_norm": 10.22000789642334,
"learning_rate": 6.444444444444445e-06,
"loss": 0.2944,
"step": 2100
},
{
"epoch": 0.68,
"grad_norm": 7.740598201751709,
"learning_rate": 6.3888888888888885e-06,
"loss": 0.2875,
"step": 2125
},
{
"epoch": 0.69,
"grad_norm": 5.818449974060059,
"learning_rate": 6.333333333333333e-06,
"loss": 0.2918,
"step": 2150
},
{
"epoch": 0.7,
"grad_norm": 17.033401489257812,
"learning_rate": 6.277777777777778e-06,
"loss": 0.2803,
"step": 2175
},
{
"epoch": 0.71,
"grad_norm": 8.139738082885742,
"learning_rate": 6.222222222222223e-06,
"loss": 0.2964,
"step": 2200
},
{
"epoch": 0.72,
"grad_norm": 4.335383892059326,
"learning_rate": 6.166666666666667e-06,
"loss": 0.2548,
"step": 2225
},
{
"epoch": 0.72,
"grad_norm": 8.209503173828125,
"learning_rate": 6.111111111111112e-06,
"loss": 0.2768,
"step": 2250
},
{
"epoch": 0.73,
"grad_norm": 8.060270309448242,
"learning_rate": 6.055555555555555e-06,
"loss": 0.3085,
"step": 2275
},
{
"epoch": 0.74,
"grad_norm": 7.265747547149658,
"learning_rate": 6e-06,
"loss": 0.2695,
"step": 2300
},
{
"epoch": 0.75,
"grad_norm": 14.466809272766113,
"learning_rate": 5.944444444444445e-06,
"loss": 0.2711,
"step": 2325
},
{
"epoch": 0.76,
"grad_norm": 9.156736373901367,
"learning_rate": 5.88888888888889e-06,
"loss": 0.2893,
"step": 2350
},
{
"epoch": 0.76,
"grad_norm": 11.08906364440918,
"learning_rate": 5.833333333333334e-06,
"loss": 0.2621,
"step": 2375
},
{
"epoch": 0.77,
"grad_norm": 14.466202735900879,
"learning_rate": 5.777777777777778e-06,
"loss": 0.2988,
"step": 2400
},
{
"epoch": 0.78,
"grad_norm": 11.54973030090332,
"learning_rate": 5.722222222222222e-06,
"loss": 0.2747,
"step": 2425
},
{
"epoch": 0.79,
"grad_norm": 9.84890079498291,
"learning_rate": 5.666666666666667e-06,
"loss": 0.2845,
"step": 2450
},
{
"epoch": 0.8,
"grad_norm": 7.711421966552734,
"learning_rate": 5.611111111111112e-06,
"loss": 0.2513,
"step": 2475
},
{
"epoch": 0.81,
"grad_norm": 13.456315994262695,
"learning_rate": 5.555555555555557e-06,
"loss": 0.2653,
"step": 2500
},
{
"epoch": 0.81,
"grad_norm": 12.786430358886719,
"learning_rate": 5.500000000000001e-06,
"loss": 0.2562,
"step": 2525
},
{
"epoch": 0.82,
"grad_norm": 5.972217559814453,
"learning_rate": 5.444444444444445e-06,
"loss": 0.2499,
"step": 2550
},
{
"epoch": 0.83,
"grad_norm": 4.539140224456787,
"learning_rate": 5.388888888888889e-06,
"loss": 0.2587,
"step": 2575
},
{
"epoch": 0.84,
"grad_norm": 8.782011032104492,
"learning_rate": 5.333333333333334e-06,
"loss": 0.2563,
"step": 2600
},
{
"epoch": 0.85,
"grad_norm": 9.317708015441895,
"learning_rate": 5.2777777777777785e-06,
"loss": 0.2669,
"step": 2625
},
{
"epoch": 0.85,
"grad_norm": 7.227994441986084,
"learning_rate": 5.2222222222222226e-06,
"loss": 0.2516,
"step": 2650
},
{
"epoch": 0.86,
"grad_norm": 7.91267728805542,
"learning_rate": 5.1666666666666675e-06,
"loss": 0.2708,
"step": 2675
},
{
"epoch": 0.87,
"grad_norm": 9.71472454071045,
"learning_rate": 5.1111111111111115e-06,
"loss": 0.2752,
"step": 2700
},
{
"epoch": 0.88,
"grad_norm": 9.538637161254883,
"learning_rate": 5.0555555555555555e-06,
"loss": 0.2541,
"step": 2725
},
{
"epoch": 0.89,
"grad_norm": 10.09244441986084,
"learning_rate": 5e-06,
"loss": 0.2667,
"step": 2750
},
{
"epoch": 0.89,
"grad_norm": 7.292573928833008,
"learning_rate": 4.944444444444445e-06,
"loss": 0.3007,
"step": 2775
},
{
"epoch": 0.9,
"grad_norm": 7.321602821350098,
"learning_rate": 4.888888888888889e-06,
"loss": 0.245,
"step": 2800
},
{
"epoch": 0.91,
"grad_norm": 9.209299087524414,
"learning_rate": 4.833333333333333e-06,
"loss": 0.2337,
"step": 2825
},
{
"epoch": 0.92,
"grad_norm": 11.65815258026123,
"learning_rate": 4.777777777777778e-06,
"loss": 0.2553,
"step": 2850
},
{
"epoch": 0.93,
"grad_norm": 9.273401260375977,
"learning_rate": 4.722222222222222e-06,
"loss": 0.247,
"step": 2875
},
{
"epoch": 0.93,
"grad_norm": 9.994607925415039,
"learning_rate": 4.666666666666667e-06,
"loss": 0.2761,
"step": 2900
},
{
"epoch": 0.94,
"grad_norm": 6.799399375915527,
"learning_rate": 4.611111111111112e-06,
"loss": 0.2637,
"step": 2925
},
{
"epoch": 0.95,
"grad_norm": 9.812761306762695,
"learning_rate": 4.555555555555556e-06,
"loss": 0.2926,
"step": 2950
},
{
"epoch": 0.96,
"grad_norm": 9.971693992614746,
"learning_rate": 4.5e-06,
"loss": 0.2571,
"step": 2975
},
{
"epoch": 0.97,
"grad_norm": 9.44893741607666,
"learning_rate": 4.444444444444444e-06,
"loss": 0.2541,
"step": 3000
},
{
"epoch": 0.97,
"eval_loss": 0.254996120929718,
"eval_runtime": 2083.5916,
"eval_samples_per_second": 3.725,
"eval_steps_per_second": 0.466,
"eval_wer": 0.5009118811440773,
"step": 3000
},
{
"epoch": 0.97,
"grad_norm": 8.639594078063965,
"learning_rate": 4.388888888888889e-06,
"loss": 0.2846,
"step": 3025
},
{
"epoch": 0.98,
"grad_norm": 9.723342895507812,
"learning_rate": 4.333333333333334e-06,
"loss": 0.2692,
"step": 3050
},
{
"epoch": 0.99,
"grad_norm": 7.255222320556641,
"learning_rate": 4.277777777777778e-06,
"loss": 0.2406,
"step": 3075
},
{
"epoch": 1.0,
"grad_norm": 9.54223346710205,
"learning_rate": 4.222222222222223e-06,
"loss": 0.2596,
"step": 3100
},
{
"epoch": 1.01,
"grad_norm": 7.074874401092529,
"learning_rate": 4.166666666666667e-06,
"loss": 0.2201,
"step": 3125
},
{
"epoch": 1.01,
"grad_norm": 5.8523640632629395,
"learning_rate": 4.111111111111111e-06,
"loss": 0.1695,
"step": 3150
},
{
"epoch": 1.02,
"grad_norm": 6.729598522186279,
"learning_rate": 4.055555555555556e-06,
"loss": 0.1396,
"step": 3175
},
{
"epoch": 1.03,
"grad_norm": 7.361201286315918,
"learning_rate": 4.000000000000001e-06,
"loss": 0.1583,
"step": 3200
},
{
"epoch": 1.04,
"grad_norm": 5.176375389099121,
"learning_rate": 3.944444444444445e-06,
"loss": 0.1543,
"step": 3225
},
{
"epoch": 1.05,
"grad_norm": 10.241930961608887,
"learning_rate": 3.88888888888889e-06,
"loss": 0.1555,
"step": 3250
},
{
"epoch": 1.05,
"grad_norm": 6.182642936706543,
"learning_rate": 3.833333333333334e-06,
"loss": 0.1441,
"step": 3275
},
{
"epoch": 1.06,
"grad_norm": 13.024508476257324,
"learning_rate": 3.777777777777778e-06,
"loss": 0.1503,
"step": 3300
},
{
"epoch": 1.07,
"grad_norm": 7.090968608856201,
"learning_rate": 3.7222222222222225e-06,
"loss": 0.1714,
"step": 3325
},
{
"epoch": 1.08,
"grad_norm": 8.179136276245117,
"learning_rate": 3.6666666666666666e-06,
"loss": 0.1663,
"step": 3350
},
{
"epoch": 1.09,
"grad_norm": 9.39560604095459,
"learning_rate": 3.6111111111111115e-06,
"loss": 0.1643,
"step": 3375
},
{
"epoch": 1.1,
"grad_norm": 6.854971408843994,
"learning_rate": 3.555555555555556e-06,
"loss": 0.1767,
"step": 3400
},
{
"epoch": 1.1,
"grad_norm": 7.17493200302124,
"learning_rate": 3.5e-06,
"loss": 0.1408,
"step": 3425
},
{
"epoch": 1.11,
"grad_norm": 5.211622714996338,
"learning_rate": 3.444444444444445e-06,
"loss": 0.1592,
"step": 3450
},
{
"epoch": 1.12,
"grad_norm": 4.639506816864014,
"learning_rate": 3.3888888888888893e-06,
"loss": 0.1428,
"step": 3475
},
{
"epoch": 1.13,
"grad_norm": 7.832518100738525,
"learning_rate": 3.3333333333333333e-06,
"loss": 0.1564,
"step": 3500
},
{
"epoch": 1.14,
"grad_norm": 5.339199542999268,
"learning_rate": 3.277777777777778e-06,
"loss": 0.1722,
"step": 3525
},
{
"epoch": 1.14,
"grad_norm": 7.612499713897705,
"learning_rate": 3.2222222222222227e-06,
"loss": 0.1667,
"step": 3550
},
{
"epoch": 1.15,
"grad_norm": 10.03056812286377,
"learning_rate": 3.1666666666666667e-06,
"loss": 0.1636,
"step": 3575
},
{
"epoch": 1.16,
"grad_norm": 6.799817085266113,
"learning_rate": 3.1111111111111116e-06,
"loss": 0.1454,
"step": 3600
},
{
"epoch": 1.17,
"grad_norm": 6.477158069610596,
"learning_rate": 3.055555555555556e-06,
"loss": 0.1604,
"step": 3625
},
{
"epoch": 1.18,
"grad_norm": 5.446293354034424,
"learning_rate": 3e-06,
"loss": 0.1545,
"step": 3650
},
{
"epoch": 1.18,
"grad_norm": 5.985503673553467,
"learning_rate": 2.944444444444445e-06,
"loss": 0.1508,
"step": 3675
},
{
"epoch": 1.19,
"grad_norm": 6.837751388549805,
"learning_rate": 2.888888888888889e-06,
"loss": 0.1587,
"step": 3700
},
{
"epoch": 1.2,
"grad_norm": 5.51887845993042,
"learning_rate": 2.8333333333333335e-06,
"loss": 0.1485,
"step": 3725
},
{
"epoch": 1.21,
"grad_norm": 8.242449760437012,
"learning_rate": 2.7777777777777783e-06,
"loss": 0.1656,
"step": 3750
},
{
"epoch": 1.22,
"grad_norm": 8.832524299621582,
"learning_rate": 2.7222222222222224e-06,
"loss": 0.1594,
"step": 3775
},
{
"epoch": 1.22,
"grad_norm": 5.77171516418457,
"learning_rate": 2.666666666666667e-06,
"loss": 0.1495,
"step": 3800
},
{
"epoch": 1.23,
"grad_norm": 6.725485801696777,
"learning_rate": 2.6111111111111113e-06,
"loss": 0.1473,
"step": 3825
},
{
"epoch": 1.24,
"grad_norm": 10.141379356384277,
"learning_rate": 2.5555555555555557e-06,
"loss": 0.1421,
"step": 3850
},
{
"epoch": 1.25,
"grad_norm": 5.970383167266846,
"learning_rate": 2.5e-06,
"loss": 0.1423,
"step": 3875
},
{
"epoch": 1.26,
"grad_norm": 6.767889499664307,
"learning_rate": 2.4444444444444447e-06,
"loss": 0.1741,
"step": 3900
},
{
"epoch": 1.26,
"grad_norm": 7.690521240234375,
"learning_rate": 2.388888888888889e-06,
"loss": 0.1614,
"step": 3925
},
{
"epoch": 1.27,
"grad_norm": 4.897395133972168,
"learning_rate": 2.3333333333333336e-06,
"loss": 0.153,
"step": 3950
},
{
"epoch": 1.28,
"grad_norm": 7.075612545013428,
"learning_rate": 2.277777777777778e-06,
"loss": 0.1563,
"step": 3975
},
{
"epoch": 1.29,
"grad_norm": 4.673244953155518,
"learning_rate": 2.222222222222222e-06,
"loss": 0.1512,
"step": 4000
},
{
"epoch": 1.29,
"eval_loss": 0.2438921481370926,
"eval_runtime": 2435.1055,
"eval_samples_per_second": 3.187,
"eval_steps_per_second": 0.399,
"eval_wer": 0.47462301499043635,
"step": 4000
},
{
"epoch": 1.3,
"grad_norm": 12.350273132324219,
"learning_rate": 2.166666666666667e-06,
"loss": 0.1631,
"step": 4025
},
{
"epoch": 1.3,
"grad_norm": 10.359116554260254,
"learning_rate": 2.1111111111111114e-06,
"loss": 0.149,
"step": 4050
},
{
"epoch": 1.31,
"grad_norm": 6.066259860992432,
"learning_rate": 2.0555555555555555e-06,
"loss": 0.1642,
"step": 4075
},
{
"epoch": 1.32,
"grad_norm": 5.781357765197754,
"learning_rate": 2.0000000000000003e-06,
"loss": 0.1451,
"step": 4100
},
{
"epoch": 1.33,
"grad_norm": 8.008292198181152,
"learning_rate": 1.944444444444445e-06,
"loss": 0.1498,
"step": 4125
},
{
"epoch": 1.34,
"grad_norm": 8.501078605651855,
"learning_rate": 1.888888888888889e-06,
"loss": 0.1614,
"step": 4150
},
{
"epoch": 1.34,
"grad_norm": 5.542455673217773,
"learning_rate": 1.8333333333333333e-06,
"loss": 0.1538,
"step": 4175
},
{
"epoch": 1.35,
"grad_norm": 6.979973316192627,
"learning_rate": 1.777777777777778e-06,
"loss": 0.151,
"step": 4200
},
{
"epoch": 1.36,
"grad_norm": 8.080177307128906,
"learning_rate": 1.7222222222222224e-06,
"loss": 0.1462,
"step": 4225
},
{
"epoch": 1.37,
"grad_norm": 7.01210355758667,
"learning_rate": 1.6666666666666667e-06,
"loss": 0.1578,
"step": 4250
},
{
"epoch": 1.38,
"grad_norm": 8.484158515930176,
"learning_rate": 1.6111111111111113e-06,
"loss": 0.1468,
"step": 4275
},
{
"epoch": 1.38,
"grad_norm": 4.0456132888793945,
"learning_rate": 1.5555555555555558e-06,
"loss": 0.1509,
"step": 4300
},
{
"epoch": 1.39,
"grad_norm": 10.766803741455078,
"learning_rate": 1.5e-06,
"loss": 0.1495,
"step": 4325
},
{
"epoch": 1.4,
"grad_norm": 9.22208023071289,
"learning_rate": 1.4444444444444445e-06,
"loss": 0.1621,
"step": 4350
},
{
"epoch": 1.41,
"grad_norm": 8.47338581085205,
"learning_rate": 1.3888888888888892e-06,
"loss": 0.1585,
"step": 4375
},
{
"epoch": 1.42,
"grad_norm": 7.198250770568848,
"learning_rate": 1.3333333333333334e-06,
"loss": 0.1726,
"step": 4400
},
{
"epoch": 1.43,
"grad_norm": 5.910817623138428,
"learning_rate": 1.2777777777777779e-06,
"loss": 0.1451,
"step": 4425
},
{
"epoch": 1.43,
"grad_norm": 5.221580982208252,
"learning_rate": 1.2222222222222223e-06,
"loss": 0.1469,
"step": 4450
},
{
"epoch": 1.44,
"grad_norm": 7.902259826660156,
"learning_rate": 1.1666666666666668e-06,
"loss": 0.1522,
"step": 4475
},
{
"epoch": 1.45,
"grad_norm": 7.130344390869141,
"learning_rate": 1.111111111111111e-06,
"loss": 0.1477,
"step": 4500
},
{
"epoch": 1.46,
"grad_norm": 4.117496490478516,
"learning_rate": 1.0555555555555557e-06,
"loss": 0.1308,
"step": 4525
},
{
"epoch": 1.47,
"grad_norm": 4.9565935134887695,
"learning_rate": 1.0000000000000002e-06,
"loss": 0.143,
"step": 4550
},
{
"epoch": 1.47,
"grad_norm": 9.52632999420166,
"learning_rate": 9.444444444444445e-07,
"loss": 0.1488,
"step": 4575
},
{
"epoch": 1.48,
"grad_norm": 7.335168838500977,
"learning_rate": 8.88888888888889e-07,
"loss": 0.1467,
"step": 4600
},
{
"epoch": 1.49,
"grad_norm": 4.118008613586426,
"learning_rate": 8.333333333333333e-07,
"loss": 0.1602,
"step": 4625
},
{
"epoch": 1.5,
"grad_norm": 5.684025764465332,
"learning_rate": 7.777777777777779e-07,
"loss": 0.1767,
"step": 4650
},
{
"epoch": 1.51,
"grad_norm": 7.431414604187012,
"learning_rate": 7.222222222222222e-07,
"loss": 0.1622,
"step": 4675
},
{
"epoch": 1.51,
"grad_norm": 11.669175148010254,
"learning_rate": 6.666666666666667e-07,
"loss": 0.1587,
"step": 4700
},
{
"epoch": 1.52,
"grad_norm": 4.074918270111084,
"learning_rate": 6.111111111111112e-07,
"loss": 0.1389,
"step": 4725
},
{
"epoch": 1.53,
"grad_norm": 4.829873561859131,
"learning_rate": 5.555555555555555e-07,
"loss": 0.1273,
"step": 4750
},
{
"epoch": 1.54,
"grad_norm": 6.028909683227539,
"learning_rate": 5.000000000000001e-07,
"loss": 0.1522,
"step": 4775
},
{
"epoch": 1.55,
"grad_norm": 6.762262344360352,
"learning_rate": 4.444444444444445e-07,
"loss": 0.1484,
"step": 4800
},
{
"epoch": 1.55,
"grad_norm": 6.13216495513916,
"learning_rate": 3.8888888888888895e-07,
"loss": 0.1423,
"step": 4825
},
{
"epoch": 1.56,
"grad_norm": 8.262957572937012,
"learning_rate": 3.3333333333333335e-07,
"loss": 0.1463,
"step": 4850
},
{
"epoch": 1.57,
"grad_norm": 3.5338327884674072,
"learning_rate": 2.7777777777777776e-07,
"loss": 0.1241,
"step": 4875
},
{
"epoch": 1.58,
"grad_norm": 8.249900817871094,
"learning_rate": 2.2222222222222224e-07,
"loss": 0.1419,
"step": 4900
},
{
"epoch": 1.59,
"grad_norm": 6.7409539222717285,
"learning_rate": 1.6666666666666668e-07,
"loss": 0.1308,
"step": 4925
},
{
"epoch": 1.59,
"grad_norm": 5.887229919433594,
"learning_rate": 1.1111111111111112e-07,
"loss": 0.1581,
"step": 4950
},
{
"epoch": 1.6,
"grad_norm": 6.0119099617004395,
"learning_rate": 5.555555555555556e-08,
"loss": 0.1407,
"step": 4975
},
{
"epoch": 1.61,
"grad_norm": 5.794888019561768,
"learning_rate": 0.0,
"loss": 0.1415,
"step": 5000
},
{
"epoch": 1.61,
"eval_loss": 0.23594366014003754,
"eval_runtime": 2284.612,
"eval_samples_per_second": 3.397,
"eval_steps_per_second": 0.425,
"eval_wer": 0.4691739691294871,
"step": 5000
}
],
"logging_steps": 25,
"max_steps": 5000,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 1000,
"total_flos": 1.15419730857984e+19,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}