{ "best_metric": null, "best_model_checkpoint": null, "epoch": 2.3342670401493932, "eval_steps": 1000, "global_step": 5000, "is_hyper_param_search": false, "is_local_process_zero": true, "is_world_process_zero": true, "log_history": [ { "epoch": 0.01, "grad_norm": 42.80295181274414, "learning_rate": 5e-09, "loss": 1.0881, "step": 25 }, { "epoch": 0.02, "grad_norm": 36.09112548828125, "learning_rate": 1e-08, "loss": 1.0284, "step": 50 }, { "epoch": 0.04, "grad_norm": 42.47553634643555, "learning_rate": 1.5e-08, "loss": 1.0556, "step": 75 }, { "epoch": 0.05, "grad_norm": 40.723106384277344, "learning_rate": 2e-08, "loss": 1.0326, "step": 100 }, { "epoch": 0.06, "grad_norm": 39.514530181884766, "learning_rate": 2.5e-08, "loss": 1.0675, "step": 125 }, { "epoch": 0.07, "grad_norm": 40.45370864868164, "learning_rate": 3e-08, "loss": 1.0513, "step": 150 }, { "epoch": 0.08, "grad_norm": 38.05487823486328, "learning_rate": 3.5e-08, "loss": 1.0533, "step": 175 }, { "epoch": 0.09, "grad_norm": 36.58195114135742, "learning_rate": 4e-08, "loss": 1.05, "step": 200 }, { "epoch": 0.11, "grad_norm": 41.44173049926758, "learning_rate": 4.5e-08, "loss": 1.0471, "step": 225 }, { "epoch": 0.12, "grad_norm": 37.410099029541016, "learning_rate": 5e-08, "loss": 1.0451, "step": 250 }, { "epoch": 0.13, "grad_norm": 38.047340393066406, "learning_rate": 5.4999999999999996e-08, "loss": 1.0402, "step": 275 }, { "epoch": 0.14, "grad_norm": 38.6348762512207, "learning_rate": 6e-08, "loss": 1.0654, "step": 300 }, { "epoch": 0.15, "grad_norm": 38.60363006591797, "learning_rate": 6.5e-08, "loss": 1.0348, "step": 325 }, { "epoch": 0.16, "grad_norm": 37.33879089355469, "learning_rate": 7e-08, "loss": 1.072, "step": 350 }, { "epoch": 0.18, "grad_norm": 36.03329086303711, "learning_rate": 7.5e-08, "loss": 0.9851, "step": 375 }, { "epoch": 0.19, "grad_norm": 34.04118347167969, "learning_rate": 8e-08, "loss": 1.0417, "step": 400 }, { "epoch": 0.2, "grad_norm": 34.55414581298828, "learning_rate": 8.500000000000001e-08, "loss": 0.9978, "step": 425 }, { "epoch": 0.21, "grad_norm": 35.677894592285156, "learning_rate": 9e-08, "loss": 1.0084, "step": 450 }, { "epoch": 0.22, "grad_norm": 36.7861213684082, "learning_rate": 9.499999999999999e-08, "loss": 0.9627, "step": 475 }, { "epoch": 0.23, "grad_norm": 32.090797424316406, "learning_rate": 1e-07, "loss": 0.9709, "step": 500 }, { "epoch": 0.25, "grad_norm": 35.564849853515625, "learning_rate": 1.0499999999999999e-07, "loss": 0.9682, "step": 525 }, { "epoch": 0.26, "grad_norm": 27.45682716369629, "learning_rate": 1.0999999999999999e-07, "loss": 0.9262, "step": 550 }, { "epoch": 0.27, "grad_norm": 32.76565933227539, "learning_rate": 1.15e-07, "loss": 0.8343, "step": 575 }, { "epoch": 0.28, "grad_norm": 25.909934997558594, "learning_rate": 1.2e-07, "loss": 0.8247, "step": 600 }, { "epoch": 0.29, "grad_norm": 28.087352752685547, "learning_rate": 1.25e-07, "loss": 0.8033, "step": 625 }, { "epoch": 0.3, "grad_norm": 28.653064727783203, "learning_rate": 1.3e-07, "loss": 0.742, "step": 650 }, { "epoch": 0.32, "grad_norm": 25.42856216430664, "learning_rate": 1.35e-07, "loss": 0.6943, "step": 675 }, { "epoch": 0.33, "grad_norm": 33.012351989746094, "learning_rate": 1.4e-07, "loss": 0.6754, "step": 700 }, { "epoch": 0.34, "grad_norm": 30.743101119995117, "learning_rate": 1.45e-07, "loss": 0.5455, "step": 725 }, { "epoch": 0.35, "grad_norm": 26.959556579589844, "learning_rate": 1.5e-07, "loss": 0.5058, "step": 750 }, { "epoch": 0.36, "grad_norm": 21.917985916137695, "learning_rate": 1.55e-07, "loss": 0.4522, "step": 775 }, { "epoch": 0.37, "grad_norm": 25.912960052490234, "learning_rate": 1.6e-07, "loss": 0.4506, "step": 800 }, { "epoch": 0.39, "grad_norm": 25.664531707763672, "learning_rate": 1.65e-07, "loss": 0.4513, "step": 825 }, { "epoch": 0.4, "grad_norm": 22.275697708129883, "learning_rate": 1.7000000000000001e-07, "loss": 0.4161, "step": 850 }, { "epoch": 0.41, "grad_norm": 24.786251068115234, "learning_rate": 1.75e-07, "loss": 0.4154, "step": 875 }, { "epoch": 0.42, "grad_norm": 25.422107696533203, "learning_rate": 1.8e-07, "loss": 0.3859, "step": 900 }, { "epoch": 0.43, "grad_norm": 17.307992935180664, "learning_rate": 1.85e-07, "loss": 0.3672, "step": 925 }, { "epoch": 0.44, "grad_norm": 21.586824417114258, "learning_rate": 1.8999999999999998e-07, "loss": 0.3655, "step": 950 }, { "epoch": 0.46, "grad_norm": 20.625280380249023, "learning_rate": 1.9499999999999999e-07, "loss": 0.3717, "step": 975 }, { "epoch": 0.47, "grad_norm": 24.026769638061523, "learning_rate": 2e-07, "loss": 0.3915, "step": 1000 }, { "epoch": 0.47, "eval_loss": 0.29781457781791687, "eval_runtime": 1337.151, "eval_samples_per_second": 7.04, "eval_steps_per_second": 0.44, "eval_wer": 0.18578591812714637, "step": 1000 }, { "epoch": 0.48, "grad_norm": 19.427650451660156, "learning_rate": 2.0499999999999997e-07, "loss": 0.3242, "step": 1025 }, { "epoch": 0.49, "grad_norm": 17.370065689086914, "learning_rate": 2.0999999999999997e-07, "loss": 0.3605, "step": 1050 }, { "epoch": 0.5, "grad_norm": 17.335697174072266, "learning_rate": 2.1499999999999998e-07, "loss": 0.3689, "step": 1075 }, { "epoch": 0.51, "grad_norm": 18.611391067504883, "learning_rate": 2.1999999999999998e-07, "loss": 0.4011, "step": 1100 }, { "epoch": 0.53, "grad_norm": 19.87590217590332, "learning_rate": 2.25e-07, "loss": 0.3401, "step": 1125 }, { "epoch": 0.54, "grad_norm": 14.708137512207031, "learning_rate": 2.3e-07, "loss": 0.353, "step": 1150 }, { "epoch": 0.55, "grad_norm": 19.14278793334961, "learning_rate": 2.3499999999999997e-07, "loss": 0.3472, "step": 1175 }, { "epoch": 0.56, "grad_norm": 15.516341209411621, "learning_rate": 2.4e-07, "loss": 0.3431, "step": 1200 }, { "epoch": 0.57, "grad_norm": 20.186065673828125, "learning_rate": 2.45e-07, "loss": 0.3244, "step": 1225 }, { "epoch": 0.58, "grad_norm": 21.650047302246094, "learning_rate": 2.5e-07, "loss": 0.341, "step": 1250 }, { "epoch": 0.6, "grad_norm": 17.734304428100586, "learning_rate": 2.55e-07, "loss": 0.3598, "step": 1275 }, { "epoch": 0.61, "grad_norm": 19.580183029174805, "learning_rate": 2.6e-07, "loss": 0.3312, "step": 1300 }, { "epoch": 0.62, "grad_norm": 19.631921768188477, "learning_rate": 2.65e-07, "loss": 0.3442, "step": 1325 }, { "epoch": 0.63, "grad_norm": 18.21005630493164, "learning_rate": 2.7e-07, "loss": 0.3318, "step": 1350 }, { "epoch": 0.64, "grad_norm": 20.728309631347656, "learning_rate": 2.75e-07, "loss": 0.3255, "step": 1375 }, { "epoch": 0.65, "grad_norm": 15.139528274536133, "learning_rate": 2.8e-07, "loss": 0.3195, "step": 1400 }, { "epoch": 0.67, "grad_norm": 14.433816909790039, "learning_rate": 2.8499999999999997e-07, "loss": 0.2906, "step": 1425 }, { "epoch": 0.68, "grad_norm": 17.25659942626953, "learning_rate": 2.9e-07, "loss": 0.3455, "step": 1450 }, { "epoch": 0.69, "grad_norm": 21.09754180908203, "learning_rate": 2.95e-07, "loss": 0.3386, "step": 1475 }, { "epoch": 0.7, "grad_norm": 15.790301322937012, "learning_rate": 3e-07, "loss": 0.3348, "step": 1500 }, { "epoch": 0.71, "grad_norm": 23.314483642578125, "learning_rate": 3.05e-07, "loss": 0.3079, "step": 1525 }, { "epoch": 0.72, "grad_norm": 16.656147003173828, "learning_rate": 3.1e-07, "loss": 0.3132, "step": 1550 }, { "epoch": 0.74, "grad_norm": 13.165972709655762, "learning_rate": 3.15e-07, "loss": 0.2978, "step": 1575 }, { "epoch": 0.75, "grad_norm": 18.286815643310547, "learning_rate": 3.2e-07, "loss": 0.3112, "step": 1600 }, { "epoch": 0.76, "grad_norm": 16.560523986816406, "learning_rate": 3.25e-07, "loss": 0.3051, "step": 1625 }, { "epoch": 0.77, "grad_norm": 18.13733673095703, "learning_rate": 3.3e-07, "loss": 0.3098, "step": 1650 }, { "epoch": 0.78, "grad_norm": 19.807390213012695, "learning_rate": 3.35e-07, "loss": 0.2979, "step": 1675 }, { "epoch": 0.79, "grad_norm": 15.216619491577148, "learning_rate": 3.4000000000000003e-07, "loss": 0.2979, "step": 1700 }, { "epoch": 0.81, "grad_norm": 17.19868278503418, "learning_rate": 3.45e-07, "loss": 0.3039, "step": 1725 }, { "epoch": 0.82, "grad_norm": 16.618858337402344, "learning_rate": 3.5e-07, "loss": 0.2991, "step": 1750 }, { "epoch": 0.83, "grad_norm": 14.719141006469727, "learning_rate": 3.55e-07, "loss": 0.2914, "step": 1775 }, { "epoch": 0.84, "grad_norm": 18.658220291137695, "learning_rate": 3.6e-07, "loss": 0.2904, "step": 1800 }, { "epoch": 0.85, "grad_norm": 16.73942756652832, "learning_rate": 3.65e-07, "loss": 0.2933, "step": 1825 }, { "epoch": 0.86, "grad_norm": 17.588727951049805, "learning_rate": 3.7e-07, "loss": 0.3185, "step": 1850 }, { "epoch": 0.88, "grad_norm": 14.940112113952637, "learning_rate": 3.75e-07, "loss": 0.3014, "step": 1875 }, { "epoch": 0.89, "grad_norm": 16.759981155395508, "learning_rate": 3.7999999999999996e-07, "loss": 0.29, "step": 1900 }, { "epoch": 0.9, "grad_norm": 16.736419677734375, "learning_rate": 3.8499999999999997e-07, "loss": 0.3015, "step": 1925 }, { "epoch": 0.91, "grad_norm": 13.355605125427246, "learning_rate": 3.8999999999999997e-07, "loss": 0.3035, "step": 1950 }, { "epoch": 0.92, "grad_norm": 22.4385929107666, "learning_rate": 3.95e-07, "loss": 0.2983, "step": 1975 }, { "epoch": 0.93, "grad_norm": 15.556875228881836, "learning_rate": 4e-07, "loss": 0.2942, "step": 2000 }, { "epoch": 0.93, "eval_loss": 0.24996864795684814, "eval_runtime": 1326.5807, "eval_samples_per_second": 7.096, "eval_steps_per_second": 0.444, "eval_wer": 0.1626493397610564, "step": 2000 }, { "epoch": 0.95, "grad_norm": 17.8362979888916, "learning_rate": 4.05e-07, "loss": 0.2912, "step": 2025 }, { "epoch": 0.96, "grad_norm": 21.89351463317871, "learning_rate": 4.0999999999999994e-07, "loss": 0.3063, "step": 2050 }, { "epoch": 0.97, "grad_norm": 13.810240745544434, "learning_rate": 4.1499999999999994e-07, "loss": 0.3019, "step": 2075 }, { "epoch": 0.98, "grad_norm": 17.076282501220703, "learning_rate": 4.1999999999999995e-07, "loss": 0.3091, "step": 2100 }, { "epoch": 0.99, "grad_norm": 15.350356101989746, "learning_rate": 4.2499999999999995e-07, "loss": 0.2994, "step": 2125 }, { "epoch": 1.0, "grad_norm": 7.815631866455078, "learning_rate": 4.2999999999999996e-07, "loss": 0.2792, "step": 2150 }, { "epoch": 1.02, "grad_norm": 8.397562026977539, "learning_rate": 4.3499999999999996e-07, "loss": 0.277, "step": 2175 }, { "epoch": 1.03, "grad_norm": 7.092517852783203, "learning_rate": 4.3999999999999997e-07, "loss": 0.2732, "step": 2200 }, { "epoch": 1.04, "grad_norm": 8.4801607131958, "learning_rate": 4.45e-07, "loss": 0.2876, "step": 2225 }, { "epoch": 1.05, "grad_norm": 9.114388465881348, "learning_rate": 4.5e-07, "loss": 0.2702, "step": 2250 }, { "epoch": 1.06, "grad_norm": 7.309689521789551, "learning_rate": 4.55e-07, "loss": 0.2994, "step": 2275 }, { "epoch": 1.07, "grad_norm": 7.109182357788086, "learning_rate": 4.6e-07, "loss": 0.2752, "step": 2300 }, { "epoch": 1.09, "grad_norm": 8.527280807495117, "learning_rate": 4.65e-07, "loss": 0.2493, "step": 2325 }, { "epoch": 1.1, "grad_norm": 5.584944725036621, "learning_rate": 4.6999999999999995e-07, "loss": 0.2618, "step": 2350 }, { "epoch": 1.11, "grad_norm": 7.848605632781982, "learning_rate": 4.7499999999999995e-07, "loss": 0.2748, "step": 2375 }, { "epoch": 1.12, "grad_norm": 6.040729999542236, "learning_rate": 4.8e-07, "loss": 0.2736, "step": 2400 }, { "epoch": 1.13, "grad_norm": 7.142736434936523, "learning_rate": 4.85e-07, "loss": 0.258, "step": 2425 }, { "epoch": 1.14, "grad_norm": 7.892075061798096, "learning_rate": 4.9e-07, "loss": 0.3096, "step": 2450 }, { "epoch": 1.16, "grad_norm": 8.93137264251709, "learning_rate": 4.95e-07, "loss": 0.2722, "step": 2475 }, { "epoch": 1.17, "grad_norm": 8.166301727294922, "learning_rate": 5e-07, "loss": 0.2924, "step": 2500 }, { "epoch": 1.18, "grad_norm": 6.910344123840332, "learning_rate": 5.049999999999999e-07, "loss": 0.2543, "step": 2525 }, { "epoch": 1.19, "grad_norm": 9.456042289733887, "learning_rate": 5.1e-07, "loss": 0.29, "step": 2550 }, { "epoch": 1.2, "grad_norm": 6.1278395652771, "learning_rate": 5.149999999999999e-07, "loss": 0.2752, "step": 2575 }, { "epoch": 1.21, "grad_norm": 8.067246437072754, "learning_rate": 5.2e-07, "loss": 0.2821, "step": 2600 }, { "epoch": 1.23, "grad_norm": 7.603886127471924, "learning_rate": 5.25e-07, "loss": 0.2838, "step": 2625 }, { "epoch": 1.24, "grad_norm": 8.258223533630371, "learning_rate": 5.3e-07, "loss": 0.2736, "step": 2650 }, { "epoch": 1.25, "grad_norm": 6.4209065437316895, "learning_rate": 5.35e-07, "loss": 0.2536, "step": 2675 }, { "epoch": 1.26, "grad_norm": 7.871333122253418, "learning_rate": 5.4e-07, "loss": 0.2779, "step": 2700 }, { "epoch": 1.27, "grad_norm": 7.8993306159973145, "learning_rate": 5.45e-07, "loss": 0.2718, "step": 2725 }, { "epoch": 1.28, "grad_norm": 7.633549690246582, "learning_rate": 5.5e-07, "loss": 0.2635, "step": 2750 }, { "epoch": 1.3, "grad_norm": 5.408315658569336, "learning_rate": 5.55e-07, "loss": 0.2602, "step": 2775 }, { "epoch": 1.31, "grad_norm": 8.084305763244629, "learning_rate": 5.6e-07, "loss": 0.2495, "step": 2800 }, { "epoch": 1.32, "grad_norm": 6.785496234893799, "learning_rate": 5.649999999999999e-07, "loss": 0.2797, "step": 2825 }, { "epoch": 1.33, "grad_norm": 8.085886001586914, "learning_rate": 5.699999999999999e-07, "loss": 0.2562, "step": 2850 }, { "epoch": 1.34, "grad_norm": 8.028220176696777, "learning_rate": 5.749999999999999e-07, "loss": 0.2534, "step": 2875 }, { "epoch": 1.35, "grad_norm": 7.391637802124023, "learning_rate": 5.8e-07, "loss": 0.2794, "step": 2900 }, { "epoch": 1.37, "grad_norm": 7.541261672973633, "learning_rate": 5.849999999999999e-07, "loss": 0.2776, "step": 2925 }, { "epoch": 1.38, "grad_norm": 7.456998348236084, "learning_rate": 5.9e-07, "loss": 0.2883, "step": 2950 }, { "epoch": 1.39, "grad_norm": 4.759747505187988, "learning_rate": 5.949999999999999e-07, "loss": 0.2468, "step": 2975 }, { "epoch": 1.4, "grad_norm": 6.258363723754883, "learning_rate": 6e-07, "loss": 0.2877, "step": 3000 }, { "epoch": 1.4, "eval_loss": 0.2336360216140747, "eval_runtime": 1334.64, "eval_samples_per_second": 7.054, "eval_steps_per_second": 0.441, "eval_wer": 0.15355593892588235, "step": 3000 }, { "epoch": 1.41, "grad_norm": 6.2899651527404785, "learning_rate": 6.049999999999999e-07, "loss": 0.2729, "step": 3025 }, { "epoch": 1.42, "grad_norm": 8.125208854675293, "learning_rate": 6.1e-07, "loss": 0.2598, "step": 3050 }, { "epoch": 1.44, "grad_norm": 7.969179630279541, "learning_rate": 6.149999999999999e-07, "loss": 0.2401, "step": 3075 }, { "epoch": 1.45, "grad_norm": 5.177332878112793, "learning_rate": 6.2e-07, "loss": 0.2491, "step": 3100 }, { "epoch": 1.46, "grad_norm": 9.723845481872559, "learning_rate": 6.249999999999999e-07, "loss": 0.2651, "step": 3125 }, { "epoch": 1.47, "grad_norm": 7.722317218780518, "learning_rate": 6.3e-07, "loss": 0.2669, "step": 3150 }, { "epoch": 1.48, "grad_norm": 5.817755222320557, "learning_rate": 6.35e-07, "loss": 0.2632, "step": 3175 }, { "epoch": 1.49, "grad_norm": 7.001987934112549, "learning_rate": 6.4e-07, "loss": 0.2419, "step": 3200 }, { "epoch": 1.51, "grad_norm": 6.027250289916992, "learning_rate": 6.45e-07, "loss": 0.2975, "step": 3225 }, { "epoch": 1.52, "grad_norm": 7.742289066314697, "learning_rate": 6.5e-07, "loss": 0.2574, "step": 3250 }, { "epoch": 1.53, "grad_norm": 7.380095958709717, "learning_rate": 6.55e-07, "loss": 0.2447, "step": 3275 }, { "epoch": 1.54, "grad_norm": 7.48017692565918, "learning_rate": 6.6e-07, "loss": 0.2903, "step": 3300 }, { "epoch": 1.55, "grad_norm": 5.8539299964904785, "learning_rate": 6.65e-07, "loss": 0.2622, "step": 3325 }, { "epoch": 1.56, "grad_norm": 5.69394063949585, "learning_rate": 6.7e-07, "loss": 0.2582, "step": 3350 }, { "epoch": 1.58, "grad_norm": 7.406588077545166, "learning_rate": 6.75e-07, "loss": 0.2804, "step": 3375 }, { "epoch": 1.59, "grad_norm": 5.877038955688477, "learning_rate": 6.800000000000001e-07, "loss": 0.2399, "step": 3400 }, { "epoch": 1.6, "grad_norm": 6.2857184410095215, "learning_rate": 6.85e-07, "loss": 0.2732, "step": 3425 }, { "epoch": 1.61, "grad_norm": 4.856884956359863, "learning_rate": 6.9e-07, "loss": 0.2459, "step": 3450 }, { "epoch": 1.62, "grad_norm": 5.824467658996582, "learning_rate": 6.949999999999999e-07, "loss": 0.2487, "step": 3475 }, { "epoch": 1.63, "grad_norm": 6.234816074371338, "learning_rate": 7e-07, "loss": 0.2594, "step": 3500 }, { "epoch": 1.65, "grad_norm": 8.982284545898438, "learning_rate": 7.049999999999999e-07, "loss": 0.2624, "step": 3525 }, { "epoch": 1.66, "grad_norm": 8.273755073547363, "learning_rate": 7.1e-07, "loss": 0.267, "step": 3550 }, { "epoch": 1.67, "grad_norm": 4.855440616607666, "learning_rate": 7.149999999999999e-07, "loss": 0.2444, "step": 3575 }, { "epoch": 1.68, "grad_norm": 7.118260383605957, "learning_rate": 7.2e-07, "loss": 0.2453, "step": 3600 }, { "epoch": 1.69, "grad_norm": 7.056091785430908, "learning_rate": 7.249999999999999e-07, "loss": 0.2602, "step": 3625 }, { "epoch": 1.7, "grad_norm": 8.34477424621582, "learning_rate": 7.3e-07, "loss": 0.2748, "step": 3650 }, { "epoch": 1.72, "grad_norm": 6.903721809387207, "learning_rate": 7.35e-07, "loss": 0.2287, "step": 3675 }, { "epoch": 1.73, "grad_norm": 6.3288092613220215, "learning_rate": 7.4e-07, "loss": 0.2604, "step": 3700 }, { "epoch": 1.74, "grad_norm": 9.17168140411377, "learning_rate": 7.45e-07, "loss": 0.2772, "step": 3725 }, { "epoch": 1.75, "grad_norm": 6.454259395599365, "learning_rate": 7.5e-07, "loss": 0.2862, "step": 3750 }, { "epoch": 1.76, "grad_norm": 7.24948787689209, "learning_rate": 7.55e-07, "loss": 0.254, "step": 3775 }, { "epoch": 1.77, "grad_norm": 6.764554977416992, "learning_rate": 7.599999999999999e-07, "loss": 0.2287, "step": 3800 }, { "epoch": 1.79, "grad_norm": 7.178032875061035, "learning_rate": 7.65e-07, "loss": 0.2592, "step": 3825 }, { "epoch": 1.8, "grad_norm": 5.957449436187744, "learning_rate": 7.699999999999999e-07, "loss": 0.2601, "step": 3850 }, { "epoch": 1.81, "grad_norm": 6.306526184082031, "learning_rate": 7.75e-07, "loss": 0.2499, "step": 3875 }, { "epoch": 1.82, "grad_norm": 6.489150524139404, "learning_rate": 7.799999999999999e-07, "loss": 0.2427, "step": 3900 }, { "epoch": 1.83, "grad_norm": 8.371627807617188, "learning_rate": 7.85e-07, "loss": 0.2586, "step": 3925 }, { "epoch": 1.84, "grad_norm": 6.425754547119141, "learning_rate": 7.9e-07, "loss": 0.2383, "step": 3950 }, { "epoch": 1.86, "grad_norm": 8.453587532043457, "learning_rate": 7.95e-07, "loss": 0.2364, "step": 3975 }, { "epoch": 1.87, "grad_norm": 5.8362717628479, "learning_rate": 8e-07, "loss": 0.2303, "step": 4000 }, { "epoch": 1.87, "eval_loss": 0.22252435982227325, "eval_runtime": 1357.5217, "eval_samples_per_second": 6.935, "eval_steps_per_second": 0.434, "eval_wer": 0.14815471679860698, "step": 4000 }, { "epoch": 1.88, "grad_norm": 7.4534735679626465, "learning_rate": 8.05e-07, "loss": 0.2451, "step": 4025 }, { "epoch": 1.89, "grad_norm": 5.907018661499023, "learning_rate": 8.1e-07, "loss": 0.2392, "step": 4050 }, { "epoch": 1.9, "grad_norm": 6.427994728088379, "learning_rate": 8.149999999999999e-07, "loss": 0.2318, "step": 4075 }, { "epoch": 1.91, "grad_norm": 7.975613117218018, "learning_rate": 8.199999999999999e-07, "loss": 0.2368, "step": 4100 }, { "epoch": 1.93, "grad_norm": 6.577622413635254, "learning_rate": 8.249999999999999e-07, "loss": 0.2458, "step": 4125 }, { "epoch": 1.94, "grad_norm": 7.6984429359436035, "learning_rate": 8.299999999999999e-07, "loss": 0.2425, "step": 4150 }, { "epoch": 1.95, "grad_norm": 6.880575180053711, "learning_rate": 8.349999999999999e-07, "loss": 0.2376, "step": 4175 }, { "epoch": 1.96, "grad_norm": 8.09802532196045, "learning_rate": 8.399999999999999e-07, "loss": 0.2668, "step": 4200 }, { "epoch": 1.97, "grad_norm": 5.784348011016846, "learning_rate": 8.45e-07, "loss": 0.2772, "step": 4225 }, { "epoch": 1.98, "grad_norm": 6.168551921844482, "learning_rate": 8.499999999999999e-07, "loss": 0.2394, "step": 4250 }, { "epoch": 2.0, "grad_norm": 8.498140335083008, "learning_rate": 8.55e-07, "loss": 0.2599, "step": 4275 }, { "epoch": 2.01, "grad_norm": 9.183964729309082, "learning_rate": 8.599999999999999e-07, "loss": 0.234, "step": 4300 }, { "epoch": 2.02, "grad_norm": 7.023620128631592, "learning_rate": 8.65e-07, "loss": 0.2208, "step": 4325 }, { "epoch": 2.03, "grad_norm": 6.869289398193359, "learning_rate": 8.699999999999999e-07, "loss": 0.2199, "step": 4350 }, { "epoch": 2.04, "grad_norm": 7.651854991912842, "learning_rate": 8.75e-07, "loss": 0.2153, "step": 4375 }, { "epoch": 2.05, "grad_norm": 10.046488761901855, "learning_rate": 8.799999999999999e-07, "loss": 0.2511, "step": 4400 }, { "epoch": 2.07, "grad_norm": 7.229065895080566, "learning_rate": 8.85e-07, "loss": 0.2172, "step": 4425 }, { "epoch": 2.08, "grad_norm": 12.570436477661133, "learning_rate": 8.9e-07, "loss": 0.2315, "step": 4450 }, { "epoch": 2.09, "grad_norm": 7.68054723739624, "learning_rate": 8.95e-07, "loss": 0.2073, "step": 4475 }, { "epoch": 2.1, "grad_norm": 4.375064373016357, "learning_rate": 9e-07, "loss": 0.2313, "step": 4500 }, { "epoch": 2.11, "grad_norm": 8.125436782836914, "learning_rate": 9.05e-07, "loss": 0.2324, "step": 4525 }, { "epoch": 2.12, "grad_norm": 7.027578353881836, "learning_rate": 9.1e-07, "loss": 0.2251, "step": 4550 }, { "epoch": 2.14, "grad_norm": 7.733719348907471, "learning_rate": 9.15e-07, "loss": 0.2192, "step": 4575 }, { "epoch": 2.15, "grad_norm": 5.648609638214111, "learning_rate": 9.2e-07, "loss": 0.2274, "step": 4600 }, { "epoch": 2.16, "grad_norm": 10.744574546813965, "learning_rate": 9.25e-07, "loss": 0.2301, "step": 4625 }, { "epoch": 2.17, "grad_norm": 8.391999244689941, "learning_rate": 9.3e-07, "loss": 0.2348, "step": 4650 }, { "epoch": 2.18, "grad_norm": 8.003060340881348, "learning_rate": 9.35e-07, "loss": 0.2147, "step": 4675 }, { "epoch": 2.19, "grad_norm": 7.240922927856445, "learning_rate": 9.399999999999999e-07, "loss": 0.2177, "step": 4700 }, { "epoch": 2.21, "grad_norm": 4.615699291229248, "learning_rate": 9.45e-07, "loss": 0.224, "step": 4725 }, { "epoch": 2.22, "grad_norm": 5.88244104385376, "learning_rate": 9.499999999999999e-07, "loss": 0.2119, "step": 4750 }, { "epoch": 2.23, "grad_norm": 8.743966102600098, "learning_rate": 9.55e-07, "loss": 0.2057, "step": 4775 }, { "epoch": 2.24, "grad_norm": 6.3263325691223145, "learning_rate": 9.6e-07, "loss": 0.208, "step": 4800 }, { "epoch": 2.25, "grad_norm": 8.424788475036621, "learning_rate": 9.649999999999999e-07, "loss": 0.2147, "step": 4825 }, { "epoch": 2.26, "grad_norm": 10.14958381652832, "learning_rate": 9.7e-07, "loss": 0.2065, "step": 4850 }, { "epoch": 2.28, "grad_norm": 7.147301197052002, "learning_rate": 9.75e-07, "loss": 0.2161, "step": 4875 }, { "epoch": 2.29, "grad_norm": 7.870689868927002, "learning_rate": 9.8e-07, "loss": 0.2138, "step": 4900 }, { "epoch": 2.3, "grad_norm": 8.061651229858398, "learning_rate": 9.849999999999999e-07, "loss": 0.2225, "step": 4925 }, { "epoch": 2.31, "grad_norm": 9.418707847595215, "learning_rate": 9.9e-07, "loss": 0.2114, "step": 4950 }, { "epoch": 2.32, "grad_norm": 7.523941516876221, "learning_rate": 9.95e-07, "loss": 0.2067, "step": 4975 }, { "epoch": 2.33, "grad_norm": 10.119476318359375, "learning_rate": 0.0, "loss": 0.2192, "step": 5000 }, { "epoch": 2.33, "eval_loss": 0.21544259786605835, "eval_runtime": 1354.6965, "eval_samples_per_second": 6.949, "eval_steps_per_second": 0.435, "eval_wer": 0.14422069232381535, "step": 5000 }, { "epoch": 2.33, "step": 5000, "total_flos": 2.308683219546931e+19, "train_loss": 0.3779804196357727, "train_runtime": 15838.189, "train_samples_per_second": 5.051, "train_steps_per_second": 0.316 } ], "logging_steps": 25, "max_steps": 5000, "num_input_tokens_seen": 0, "num_train_epochs": 3, "save_steps": 1000, "total_flos": 2.308683219546931e+19, "train_batch_size": 8, "trial_name": null, "trial_params": null }