whisper-small-wer35-ekg / trainer_state.json
rikeshsilwalekg's picture
Upload folder using huggingface_hub
1dc5d3a verified
{
"best_metric": 39.565217391304344,
"best_model_checkpoint": "./whisper-small-wer35-ekg\\checkpoint-6000",
"epoch": 95.58823529411765,
"eval_steps": 500,
"global_step": 6500,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.15,
"grad_norm": 28.624893188476562,
"learning_rate": 9.433962264150944e-08,
"loss": 2.7579,
"step": 10
},
{
"epoch": 0.29,
"grad_norm": 28.752416610717773,
"learning_rate": 2.1226415094339622e-07,
"loss": 2.6811,
"step": 20
},
{
"epoch": 0.44,
"grad_norm": 25.40899658203125,
"learning_rate": 3.3018867924528305e-07,
"loss": 2.5367,
"step": 30
},
{
"epoch": 0.59,
"grad_norm": 22.991146087646484,
"learning_rate": 4.4811320754716983e-07,
"loss": 2.2512,
"step": 40
},
{
"epoch": 0.74,
"grad_norm": 18.11638832092285,
"learning_rate": 5.660377358490567e-07,
"loss": 2.0029,
"step": 50
},
{
"epoch": 0.88,
"grad_norm": 17.4566650390625,
"learning_rate": 6.839622641509434e-07,
"loss": 1.7014,
"step": 60
},
{
"epoch": 1.03,
"grad_norm": 12.19814395904541,
"learning_rate": 8.018867924528302e-07,
"loss": 1.4435,
"step": 70
},
{
"epoch": 1.18,
"grad_norm": 9.45124626159668,
"learning_rate": 9.19811320754717e-07,
"loss": 1.2141,
"step": 80
},
{
"epoch": 1.32,
"grad_norm": 8.163089752197266,
"learning_rate": 1.037735849056604e-06,
"loss": 1.0621,
"step": 90
},
{
"epoch": 1.47,
"grad_norm": 6.771365642547607,
"learning_rate": 1.1556603773584908e-06,
"loss": 0.9325,
"step": 100
},
{
"epoch": 1.62,
"grad_norm": 5.819299697875977,
"learning_rate": 1.2735849056603775e-06,
"loss": 0.8767,
"step": 110
},
{
"epoch": 1.76,
"grad_norm": 5.547601222991943,
"learning_rate": 1.3915094339622643e-06,
"loss": 0.7872,
"step": 120
},
{
"epoch": 1.91,
"grad_norm": 4.7814154624938965,
"learning_rate": 1.509433962264151e-06,
"loss": 0.7148,
"step": 130
},
{
"epoch": 2.06,
"grad_norm": 4.755609512329102,
"learning_rate": 1.6273584905660379e-06,
"loss": 0.6742,
"step": 140
},
{
"epoch": 2.21,
"grad_norm": 4.354093074798584,
"learning_rate": 1.7452830188679247e-06,
"loss": 0.6191,
"step": 150
},
{
"epoch": 2.35,
"grad_norm": 4.038461685180664,
"learning_rate": 1.8632075471698114e-06,
"loss": 0.5673,
"step": 160
},
{
"epoch": 2.5,
"grad_norm": 4.2933526039123535,
"learning_rate": 1.981132075471698e-06,
"loss": 0.5534,
"step": 170
},
{
"epoch": 2.65,
"grad_norm": 4.423107147216797,
"learning_rate": 2.099056603773585e-06,
"loss": 0.5162,
"step": 180
},
{
"epoch": 2.79,
"grad_norm": 3.7280497550964355,
"learning_rate": 2.2169811320754718e-06,
"loss": 0.4796,
"step": 190
},
{
"epoch": 2.94,
"grad_norm": 3.990518808364868,
"learning_rate": 2.3349056603773588e-06,
"loss": 0.4976,
"step": 200
},
{
"epoch": 3.09,
"grad_norm": 3.825350046157837,
"learning_rate": 2.4528301886792453e-06,
"loss": 0.4493,
"step": 210
},
{
"epoch": 3.24,
"grad_norm": 4.116565704345703,
"learning_rate": 2.5707547169811327e-06,
"loss": 0.429,
"step": 220
},
{
"epoch": 3.38,
"grad_norm": 3.6445629596710205,
"learning_rate": 2.688679245283019e-06,
"loss": 0.3898,
"step": 230
},
{
"epoch": 3.53,
"grad_norm": 3.741975784301758,
"learning_rate": 2.8066037735849063e-06,
"loss": 0.394,
"step": 240
},
{
"epoch": 3.68,
"grad_norm": 3.4814412593841553,
"learning_rate": 2.9245283018867924e-06,
"loss": 0.4115,
"step": 250
},
{
"epoch": 3.82,
"grad_norm": 3.914616823196411,
"learning_rate": 3.04245283018868e-06,
"loss": 0.3858,
"step": 260
},
{
"epoch": 3.97,
"grad_norm": 3.9170775413513184,
"learning_rate": 3.160377358490566e-06,
"loss": 0.3731,
"step": 270
},
{
"epoch": 4.12,
"grad_norm": 3.542541980743408,
"learning_rate": 3.2783018867924534e-06,
"loss": 0.347,
"step": 280
},
{
"epoch": 4.26,
"grad_norm": 3.5728800296783447,
"learning_rate": 3.3962264150943395e-06,
"loss": 0.3171,
"step": 290
},
{
"epoch": 4.41,
"grad_norm": 3.627119541168213,
"learning_rate": 3.514150943396227e-06,
"loss": 0.3105,
"step": 300
},
{
"epoch": 4.56,
"grad_norm": 3.763631820678711,
"learning_rate": 3.632075471698113e-06,
"loss": 0.315,
"step": 310
},
{
"epoch": 4.71,
"grad_norm": 3.1514198780059814,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.3116,
"step": 320
},
{
"epoch": 4.85,
"grad_norm": 3.847184181213379,
"learning_rate": 3.8679245283018875e-06,
"loss": 0.3076,
"step": 330
},
{
"epoch": 5.0,
"grad_norm": 3.8552136421203613,
"learning_rate": 3.985849056603774e-06,
"loss": 0.3106,
"step": 340
},
{
"epoch": 5.15,
"grad_norm": 3.341456174850464,
"learning_rate": 4.103773584905661e-06,
"loss": 0.2699,
"step": 350
},
{
"epoch": 5.29,
"grad_norm": 3.3753416538238525,
"learning_rate": 4.221698113207548e-06,
"loss": 0.2442,
"step": 360
},
{
"epoch": 5.44,
"grad_norm": 3.4466023445129395,
"learning_rate": 4.339622641509435e-06,
"loss": 0.2436,
"step": 370
},
{
"epoch": 5.59,
"grad_norm": 3.2262954711914062,
"learning_rate": 4.457547169811321e-06,
"loss": 0.2524,
"step": 380
},
{
"epoch": 5.74,
"grad_norm": 3.167746067047119,
"learning_rate": 4.575471698113208e-06,
"loss": 0.2551,
"step": 390
},
{
"epoch": 5.88,
"grad_norm": 3.338438034057617,
"learning_rate": 4.693396226415095e-06,
"loss": 0.2525,
"step": 400
},
{
"epoch": 6.03,
"grad_norm": 2.4594788551330566,
"learning_rate": 4.811320754716982e-06,
"loss": 0.2243,
"step": 410
},
{
"epoch": 6.18,
"grad_norm": 3.1020987033843994,
"learning_rate": 4.929245283018868e-06,
"loss": 0.2055,
"step": 420
},
{
"epoch": 6.32,
"grad_norm": 2.919455051422119,
"learning_rate": 5.047169811320756e-06,
"loss": 0.2107,
"step": 430
},
{
"epoch": 6.47,
"grad_norm": 3.4662413597106934,
"learning_rate": 5.165094339622642e-06,
"loss": 0.1946,
"step": 440
},
{
"epoch": 6.62,
"grad_norm": 3.209831476211548,
"learning_rate": 5.283018867924529e-06,
"loss": 0.2049,
"step": 450
},
{
"epoch": 6.76,
"grad_norm": 3.5380163192749023,
"learning_rate": 5.400943396226416e-06,
"loss": 0.1865,
"step": 460
},
{
"epoch": 6.91,
"grad_norm": 3.4075424671173096,
"learning_rate": 5.518867924528303e-06,
"loss": 0.184,
"step": 470
},
{
"epoch": 7.06,
"grad_norm": 2.4087564945220947,
"learning_rate": 5.636792452830189e-06,
"loss": 0.1725,
"step": 480
},
{
"epoch": 7.21,
"grad_norm": 3.172959566116333,
"learning_rate": 5.754716981132076e-06,
"loss": 0.1505,
"step": 490
},
{
"epoch": 7.35,
"grad_norm": 3.0081095695495605,
"learning_rate": 5.872641509433963e-06,
"loss": 0.149,
"step": 500
},
{
"epoch": 7.35,
"eval_loss": 0.32423341274261475,
"eval_runtime": 476.393,
"eval_samples_per_second": 1.583,
"eval_steps_per_second": 0.017,
"eval_wer": 50.19762845849802,
"step": 500
},
{
"epoch": 7.5,
"grad_norm": 2.85640811920166,
"learning_rate": 5.99056603773585e-06,
"loss": 0.1478,
"step": 510
},
{
"epoch": 7.65,
"grad_norm": 3.185556650161743,
"learning_rate": 6.108490566037736e-06,
"loss": 0.1495,
"step": 520
},
{
"epoch": 7.79,
"grad_norm": 2.893033266067505,
"learning_rate": 6.226415094339623e-06,
"loss": 0.1508,
"step": 530
},
{
"epoch": 7.94,
"grad_norm": 3.044069290161133,
"learning_rate": 6.34433962264151e-06,
"loss": 0.155,
"step": 540
},
{
"epoch": 8.09,
"grad_norm": 2.3645215034484863,
"learning_rate": 6.462264150943397e-06,
"loss": 0.1191,
"step": 550
},
{
"epoch": 8.24,
"grad_norm": 2.373063564300537,
"learning_rate": 6.580188679245284e-06,
"loss": 0.1092,
"step": 560
},
{
"epoch": 8.38,
"grad_norm": 2.313464641571045,
"learning_rate": 6.69811320754717e-06,
"loss": 0.1108,
"step": 570
},
{
"epoch": 8.53,
"grad_norm": 2.749617099761963,
"learning_rate": 6.816037735849057e-06,
"loss": 0.108,
"step": 580
},
{
"epoch": 8.68,
"grad_norm": 2.794914960861206,
"learning_rate": 6.933962264150944e-06,
"loss": 0.1102,
"step": 590
},
{
"epoch": 8.82,
"grad_norm": 2.7622482776641846,
"learning_rate": 7.051886792452831e-06,
"loss": 0.1121,
"step": 600
},
{
"epoch": 8.97,
"grad_norm": 2.6082873344421387,
"learning_rate": 7.169811320754717e-06,
"loss": 0.1171,
"step": 610
},
{
"epoch": 9.12,
"grad_norm": 2.379115581512451,
"learning_rate": 7.287735849056604e-06,
"loss": 0.0824,
"step": 620
},
{
"epoch": 9.26,
"grad_norm": 3.37858510017395,
"learning_rate": 7.405660377358491e-06,
"loss": 0.0834,
"step": 630
},
{
"epoch": 9.41,
"grad_norm": 2.881742000579834,
"learning_rate": 7.523584905660378e-06,
"loss": 0.0809,
"step": 640
},
{
"epoch": 9.56,
"grad_norm": 2.6887032985687256,
"learning_rate": 7.641509433962266e-06,
"loss": 0.0832,
"step": 650
},
{
"epoch": 9.71,
"grad_norm": 2.685237407684326,
"learning_rate": 7.75943396226415e-06,
"loss": 0.0815,
"step": 660
},
{
"epoch": 9.85,
"grad_norm": 2.297445774078369,
"learning_rate": 7.877358490566038e-06,
"loss": 0.078,
"step": 670
},
{
"epoch": 10.0,
"grad_norm": 3.055098056793213,
"learning_rate": 7.995283018867925e-06,
"loss": 0.0753,
"step": 680
},
{
"epoch": 10.15,
"grad_norm": 2.498236656188965,
"learning_rate": 8.113207547169812e-06,
"loss": 0.054,
"step": 690
},
{
"epoch": 10.29,
"grad_norm": 2.1496338844299316,
"learning_rate": 8.231132075471699e-06,
"loss": 0.0583,
"step": 700
},
{
"epoch": 10.44,
"grad_norm": 2.9474973678588867,
"learning_rate": 8.349056603773585e-06,
"loss": 0.0559,
"step": 710
},
{
"epoch": 10.59,
"grad_norm": 3.049436569213867,
"learning_rate": 8.466981132075472e-06,
"loss": 0.0557,
"step": 720
},
{
"epoch": 10.74,
"grad_norm": 2.463704824447632,
"learning_rate": 8.58490566037736e-06,
"loss": 0.058,
"step": 730
},
{
"epoch": 10.88,
"grad_norm": 2.369274139404297,
"learning_rate": 8.702830188679245e-06,
"loss": 0.0614,
"step": 740
},
{
"epoch": 11.03,
"grad_norm": 2.1053879261016846,
"learning_rate": 8.820754716981133e-06,
"loss": 0.0545,
"step": 750
},
{
"epoch": 11.18,
"grad_norm": 1.6645501852035522,
"learning_rate": 8.938679245283019e-06,
"loss": 0.0404,
"step": 760
},
{
"epoch": 11.32,
"grad_norm": 2.5582962036132812,
"learning_rate": 9.056603773584907e-06,
"loss": 0.0407,
"step": 770
},
{
"epoch": 11.47,
"grad_norm": 2.079361915588379,
"learning_rate": 9.174528301886794e-06,
"loss": 0.0392,
"step": 780
},
{
"epoch": 11.62,
"grad_norm": 2.74023699760437,
"learning_rate": 9.292452830188679e-06,
"loss": 0.0437,
"step": 790
},
{
"epoch": 11.76,
"grad_norm": 2.8623881340026855,
"learning_rate": 9.410377358490567e-06,
"loss": 0.0416,
"step": 800
},
{
"epoch": 11.91,
"grad_norm": 2.3341763019561768,
"learning_rate": 9.528301886792455e-06,
"loss": 0.0435,
"step": 810
},
{
"epoch": 12.06,
"grad_norm": 2.0387160778045654,
"learning_rate": 9.64622641509434e-06,
"loss": 0.04,
"step": 820
},
{
"epoch": 12.21,
"grad_norm": 2.4834096431732178,
"learning_rate": 9.764150943396227e-06,
"loss": 0.029,
"step": 830
},
{
"epoch": 12.35,
"grad_norm": 2.721518039703369,
"learning_rate": 9.882075471698113e-06,
"loss": 0.0297,
"step": 840
},
{
"epoch": 12.5,
"grad_norm": 2.684687376022339,
"learning_rate": 1e-05,
"loss": 0.0316,
"step": 850
},
{
"epoch": 12.65,
"grad_norm": 2.5442607402801514,
"learning_rate": 9.983153638814017e-06,
"loss": 0.031,
"step": 860
},
{
"epoch": 12.79,
"grad_norm": 3.274472236633301,
"learning_rate": 9.966307277628034e-06,
"loss": 0.0328,
"step": 870
},
{
"epoch": 12.94,
"grad_norm": 2.6189253330230713,
"learning_rate": 9.94946091644205e-06,
"loss": 0.0328,
"step": 880
},
{
"epoch": 13.09,
"grad_norm": 2.8672521114349365,
"learning_rate": 9.932614555256066e-06,
"loss": 0.0299,
"step": 890
},
{
"epoch": 13.24,
"grad_norm": 2.2015247344970703,
"learning_rate": 9.915768194070081e-06,
"loss": 0.0224,
"step": 900
},
{
"epoch": 13.38,
"grad_norm": 1.7655293941497803,
"learning_rate": 9.898921832884097e-06,
"loss": 0.0222,
"step": 910
},
{
"epoch": 13.53,
"grad_norm": 2.5391361713409424,
"learning_rate": 9.882075471698113e-06,
"loss": 0.0232,
"step": 920
},
{
"epoch": 13.68,
"grad_norm": 2.2108376026153564,
"learning_rate": 9.86522911051213e-06,
"loss": 0.0249,
"step": 930
},
{
"epoch": 13.82,
"grad_norm": 2.0707106590270996,
"learning_rate": 9.848382749326146e-06,
"loss": 0.0241,
"step": 940
},
{
"epoch": 13.97,
"grad_norm": 2.2643473148345947,
"learning_rate": 9.831536388140162e-06,
"loss": 0.0276,
"step": 950
},
{
"epoch": 14.12,
"grad_norm": 1.739284634590149,
"learning_rate": 9.81469002695418e-06,
"loss": 0.017,
"step": 960
},
{
"epoch": 14.26,
"grad_norm": 1.8066262006759644,
"learning_rate": 9.797843665768195e-06,
"loss": 0.0157,
"step": 970
},
{
"epoch": 14.41,
"grad_norm": 2.5150985717773438,
"learning_rate": 9.780997304582211e-06,
"loss": 0.0196,
"step": 980
},
{
"epoch": 14.56,
"grad_norm": 1.7605470418930054,
"learning_rate": 9.764150943396227e-06,
"loss": 0.0169,
"step": 990
},
{
"epoch": 14.71,
"grad_norm": 2.1510426998138428,
"learning_rate": 9.747304582210243e-06,
"loss": 0.0186,
"step": 1000
},
{
"epoch": 14.71,
"eval_loss": 0.4280255138874054,
"eval_runtime": 408.4481,
"eval_samples_per_second": 1.846,
"eval_steps_per_second": 0.02,
"eval_wer": 46.126482213438734,
"step": 1000
},
{
"epoch": 14.85,
"grad_norm": 2.3640592098236084,
"learning_rate": 9.73045822102426e-06,
"loss": 0.0187,
"step": 1010
},
{
"epoch": 15.0,
"grad_norm": 1.8428599834442139,
"learning_rate": 9.713611859838276e-06,
"loss": 0.0213,
"step": 1020
},
{
"epoch": 15.15,
"grad_norm": 1.8431464433670044,
"learning_rate": 9.696765498652292e-06,
"loss": 0.0127,
"step": 1030
},
{
"epoch": 15.29,
"grad_norm": 1.7996199131011963,
"learning_rate": 9.679919137466307e-06,
"loss": 0.0137,
"step": 1040
},
{
"epoch": 15.44,
"grad_norm": 1.7150269746780396,
"learning_rate": 9.663072776280325e-06,
"loss": 0.0148,
"step": 1050
},
{
"epoch": 15.59,
"grad_norm": 1.8035845756530762,
"learning_rate": 9.64622641509434e-06,
"loss": 0.0137,
"step": 1060
},
{
"epoch": 15.74,
"grad_norm": 2.240055799484253,
"learning_rate": 9.629380053908356e-06,
"loss": 0.0146,
"step": 1070
},
{
"epoch": 15.88,
"grad_norm": 2.2169549465179443,
"learning_rate": 9.612533692722372e-06,
"loss": 0.0129,
"step": 1080
},
{
"epoch": 16.03,
"grad_norm": 1.7007489204406738,
"learning_rate": 9.595687331536388e-06,
"loss": 0.0143,
"step": 1090
},
{
"epoch": 16.18,
"grad_norm": 1.3945379257202148,
"learning_rate": 9.578840970350406e-06,
"loss": 0.0111,
"step": 1100
},
{
"epoch": 16.32,
"grad_norm": 1.6498446464538574,
"learning_rate": 9.561994609164421e-06,
"loss": 0.0108,
"step": 1110
},
{
"epoch": 16.47,
"grad_norm": 1.7959011793136597,
"learning_rate": 9.545148247978437e-06,
"loss": 0.012,
"step": 1120
},
{
"epoch": 16.62,
"grad_norm": 1.7820062637329102,
"learning_rate": 9.528301886792455e-06,
"loss": 0.0133,
"step": 1130
},
{
"epoch": 16.76,
"grad_norm": 1.8330453634262085,
"learning_rate": 9.51145552560647e-06,
"loss": 0.0134,
"step": 1140
},
{
"epoch": 16.91,
"grad_norm": 1.6996911764144897,
"learning_rate": 9.494609164420486e-06,
"loss": 0.0121,
"step": 1150
},
{
"epoch": 17.06,
"grad_norm": 1.2411192655563354,
"learning_rate": 9.477762803234502e-06,
"loss": 0.0112,
"step": 1160
},
{
"epoch": 17.21,
"grad_norm": 1.4605441093444824,
"learning_rate": 9.460916442048518e-06,
"loss": 0.0108,
"step": 1170
},
{
"epoch": 17.35,
"grad_norm": 1.6172523498535156,
"learning_rate": 9.444070080862533e-06,
"loss": 0.0093,
"step": 1180
},
{
"epoch": 17.5,
"grad_norm": 1.367756724357605,
"learning_rate": 9.427223719676551e-06,
"loss": 0.0124,
"step": 1190
},
{
"epoch": 17.65,
"grad_norm": 2.769695997238159,
"learning_rate": 9.410377358490567e-06,
"loss": 0.0111,
"step": 1200
},
{
"epoch": 17.79,
"grad_norm": 1.4088915586471558,
"learning_rate": 9.393530997304582e-06,
"loss": 0.0118,
"step": 1210
},
{
"epoch": 17.94,
"grad_norm": 2.075134038925171,
"learning_rate": 9.3766846361186e-06,
"loss": 0.0113,
"step": 1220
},
{
"epoch": 18.09,
"grad_norm": 1.1849188804626465,
"learning_rate": 9.359838274932616e-06,
"loss": 0.0095,
"step": 1230
},
{
"epoch": 18.24,
"grad_norm": 0.7864483594894409,
"learning_rate": 9.342991913746632e-06,
"loss": 0.008,
"step": 1240
},
{
"epoch": 18.38,
"grad_norm": 1.687619924545288,
"learning_rate": 9.326145552560647e-06,
"loss": 0.0082,
"step": 1250
},
{
"epoch": 18.53,
"grad_norm": 2.207996368408203,
"learning_rate": 9.309299191374663e-06,
"loss": 0.0105,
"step": 1260
},
{
"epoch": 18.68,
"grad_norm": 1.876765489578247,
"learning_rate": 9.292452830188679e-06,
"loss": 0.009,
"step": 1270
},
{
"epoch": 18.82,
"grad_norm": 2.1015830039978027,
"learning_rate": 9.275606469002696e-06,
"loss": 0.0113,
"step": 1280
},
{
"epoch": 18.97,
"grad_norm": 1.7208313941955566,
"learning_rate": 9.258760107816712e-06,
"loss": 0.0116,
"step": 1290
},
{
"epoch": 19.12,
"grad_norm": 1.086099624633789,
"learning_rate": 9.241913746630728e-06,
"loss": 0.0085,
"step": 1300
},
{
"epoch": 19.26,
"grad_norm": 1.8027433156967163,
"learning_rate": 9.225067385444745e-06,
"loss": 0.0088,
"step": 1310
},
{
"epoch": 19.41,
"grad_norm": 1.5787283182144165,
"learning_rate": 9.208221024258761e-06,
"loss": 0.0075,
"step": 1320
},
{
"epoch": 19.56,
"grad_norm": 1.3783093690872192,
"learning_rate": 9.191374663072777e-06,
"loss": 0.0086,
"step": 1330
},
{
"epoch": 19.71,
"grad_norm": 1.2076382637023926,
"learning_rate": 9.174528301886794e-06,
"loss": 0.0084,
"step": 1340
},
{
"epoch": 19.85,
"grad_norm": 1.688698649406433,
"learning_rate": 9.157681940700809e-06,
"loss": 0.008,
"step": 1350
},
{
"epoch": 20.0,
"grad_norm": 1.9429757595062256,
"learning_rate": 9.140835579514824e-06,
"loss": 0.0081,
"step": 1360
},
{
"epoch": 20.15,
"grad_norm": 1.4460784196853638,
"learning_rate": 9.123989218328842e-06,
"loss": 0.0067,
"step": 1370
},
{
"epoch": 20.29,
"grad_norm": 1.672726035118103,
"learning_rate": 9.107142857142858e-06,
"loss": 0.0068,
"step": 1380
},
{
"epoch": 20.44,
"grad_norm": 0.6751158833503723,
"learning_rate": 9.090296495956873e-06,
"loss": 0.0066,
"step": 1390
},
{
"epoch": 20.59,
"grad_norm": 1.696090817451477,
"learning_rate": 9.07345013477089e-06,
"loss": 0.0066,
"step": 1400
},
{
"epoch": 20.74,
"grad_norm": 1.0858286619186401,
"learning_rate": 9.056603773584907e-06,
"loss": 0.0065,
"step": 1410
},
{
"epoch": 20.88,
"grad_norm": 1.6216528415679932,
"learning_rate": 9.039757412398922e-06,
"loss": 0.0069,
"step": 1420
},
{
"epoch": 21.03,
"grad_norm": 1.192878246307373,
"learning_rate": 9.02291105121294e-06,
"loss": 0.0063,
"step": 1430
},
{
"epoch": 21.18,
"grad_norm": 1.2905162572860718,
"learning_rate": 9.006064690026954e-06,
"loss": 0.0062,
"step": 1440
},
{
"epoch": 21.32,
"grad_norm": 1.0860906839370728,
"learning_rate": 8.989218328840971e-06,
"loss": 0.0058,
"step": 1450
},
{
"epoch": 21.47,
"grad_norm": 1.6400948762893677,
"learning_rate": 8.972371967654987e-06,
"loss": 0.0066,
"step": 1460
},
{
"epoch": 21.62,
"grad_norm": 2.1465609073638916,
"learning_rate": 8.955525606469003e-06,
"loss": 0.0097,
"step": 1470
},
{
"epoch": 21.76,
"grad_norm": 0.9771747589111328,
"learning_rate": 8.938679245283019e-06,
"loss": 0.0074,
"step": 1480
},
{
"epoch": 21.91,
"grad_norm": 1.7411590814590454,
"learning_rate": 8.921832884097036e-06,
"loss": 0.0068,
"step": 1490
},
{
"epoch": 22.06,
"grad_norm": 1.491974949836731,
"learning_rate": 8.904986522911052e-06,
"loss": 0.006,
"step": 1500
},
{
"epoch": 22.06,
"eval_loss": 0.44684624671936035,
"eval_runtime": 395.7518,
"eval_samples_per_second": 1.905,
"eval_steps_per_second": 0.02,
"eval_wer": 43.537549407114625,
"step": 1500
},
{
"epoch": 22.21,
"grad_norm": 1.5780655145645142,
"learning_rate": 8.888140161725068e-06,
"loss": 0.0047,
"step": 1510
},
{
"epoch": 22.35,
"grad_norm": 1.6800497770309448,
"learning_rate": 8.871293800539085e-06,
"loss": 0.0057,
"step": 1520
},
{
"epoch": 22.5,
"grad_norm": 1.0676138401031494,
"learning_rate": 8.8544474393531e-06,
"loss": 0.0055,
"step": 1530
},
{
"epoch": 22.65,
"grad_norm": 2.0500409603118896,
"learning_rate": 8.837601078167117e-06,
"loss": 0.0059,
"step": 1540
},
{
"epoch": 22.79,
"grad_norm": 1.3901283740997314,
"learning_rate": 8.820754716981133e-06,
"loss": 0.0056,
"step": 1550
},
{
"epoch": 22.94,
"grad_norm": 1.7657601833343506,
"learning_rate": 8.803908355795148e-06,
"loss": 0.0069,
"step": 1560
},
{
"epoch": 23.09,
"grad_norm": 0.7901756763458252,
"learning_rate": 8.787061994609166e-06,
"loss": 0.0048,
"step": 1570
},
{
"epoch": 23.24,
"grad_norm": 1.3437069654464722,
"learning_rate": 8.770215633423182e-06,
"loss": 0.0052,
"step": 1580
},
{
"epoch": 23.38,
"grad_norm": 1.1439257860183716,
"learning_rate": 8.753369272237197e-06,
"loss": 0.0048,
"step": 1590
},
{
"epoch": 23.53,
"grad_norm": 2.013002872467041,
"learning_rate": 8.736522911051213e-06,
"loss": 0.005,
"step": 1600
},
{
"epoch": 23.68,
"grad_norm": 1.2465882301330566,
"learning_rate": 8.71967654986523e-06,
"loss": 0.0067,
"step": 1610
},
{
"epoch": 23.82,
"grad_norm": 2.2301642894744873,
"learning_rate": 8.702830188679245e-06,
"loss": 0.0058,
"step": 1620
},
{
"epoch": 23.97,
"grad_norm": 1.269247055053711,
"learning_rate": 8.685983827493262e-06,
"loss": 0.0055,
"step": 1630
},
{
"epoch": 24.12,
"grad_norm": 0.8886606693267822,
"learning_rate": 8.669137466307278e-06,
"loss": 0.0037,
"step": 1640
},
{
"epoch": 24.26,
"grad_norm": 1.7440249919891357,
"learning_rate": 8.652291105121294e-06,
"loss": 0.0049,
"step": 1650
},
{
"epoch": 24.41,
"grad_norm": 0.9428110122680664,
"learning_rate": 8.635444743935311e-06,
"loss": 0.0052,
"step": 1660
},
{
"epoch": 24.56,
"grad_norm": 1.336026906967163,
"learning_rate": 8.618598382749327e-06,
"loss": 0.0066,
"step": 1670
},
{
"epoch": 24.71,
"grad_norm": 0.7566521763801575,
"learning_rate": 8.601752021563343e-06,
"loss": 0.004,
"step": 1680
},
{
"epoch": 24.85,
"grad_norm": 1.2722612619400024,
"learning_rate": 8.58490566037736e-06,
"loss": 0.0039,
"step": 1690
},
{
"epoch": 25.0,
"grad_norm": 0.7349913120269775,
"learning_rate": 8.568059299191376e-06,
"loss": 0.0048,
"step": 1700
},
{
"epoch": 25.15,
"grad_norm": 0.8350051641464233,
"learning_rate": 8.55121293800539e-06,
"loss": 0.004,
"step": 1710
},
{
"epoch": 25.29,
"grad_norm": 1.1201083660125732,
"learning_rate": 8.534366576819408e-06,
"loss": 0.0044,
"step": 1720
},
{
"epoch": 25.44,
"grad_norm": 0.9569421410560608,
"learning_rate": 8.517520215633423e-06,
"loss": 0.0048,
"step": 1730
},
{
"epoch": 25.59,
"grad_norm": 0.9667839407920837,
"learning_rate": 8.50067385444744e-06,
"loss": 0.0051,
"step": 1740
},
{
"epoch": 25.74,
"grad_norm": 0.6828237175941467,
"learning_rate": 8.483827493261457e-06,
"loss": 0.0061,
"step": 1750
},
{
"epoch": 25.88,
"grad_norm": 1.4523017406463623,
"learning_rate": 8.466981132075472e-06,
"loss": 0.0054,
"step": 1760
},
{
"epoch": 26.03,
"grad_norm": 1.0334619283676147,
"learning_rate": 8.450134770889488e-06,
"loss": 0.0034,
"step": 1770
},
{
"epoch": 26.18,
"grad_norm": 1.5895966291427612,
"learning_rate": 8.433288409703506e-06,
"loss": 0.0036,
"step": 1780
},
{
"epoch": 26.32,
"grad_norm": 1.0893361568450928,
"learning_rate": 8.416442048517522e-06,
"loss": 0.0041,
"step": 1790
},
{
"epoch": 26.47,
"grad_norm": 1.2191766500473022,
"learning_rate": 8.399595687331537e-06,
"loss": 0.0038,
"step": 1800
},
{
"epoch": 26.62,
"grad_norm": 1.8165545463562012,
"learning_rate": 8.382749326145553e-06,
"loss": 0.0038,
"step": 1810
},
{
"epoch": 26.76,
"grad_norm": 1.477317452430725,
"learning_rate": 8.365902964959569e-06,
"loss": 0.0044,
"step": 1820
},
{
"epoch": 26.91,
"grad_norm": 0.8306496739387512,
"learning_rate": 8.349056603773585e-06,
"loss": 0.0036,
"step": 1830
},
{
"epoch": 27.06,
"grad_norm": 0.9612523317337036,
"learning_rate": 8.332210242587602e-06,
"loss": 0.0031,
"step": 1840
},
{
"epoch": 27.21,
"grad_norm": 1.7324286699295044,
"learning_rate": 8.315363881401618e-06,
"loss": 0.0029,
"step": 1850
},
{
"epoch": 27.35,
"grad_norm": 1.0282984972000122,
"learning_rate": 8.298517520215634e-06,
"loss": 0.0038,
"step": 1860
},
{
"epoch": 27.5,
"grad_norm": 0.7670619487762451,
"learning_rate": 8.281671159029651e-06,
"loss": 0.0044,
"step": 1870
},
{
"epoch": 27.65,
"grad_norm": 0.9554408192634583,
"learning_rate": 8.264824797843667e-06,
"loss": 0.0048,
"step": 1880
},
{
"epoch": 27.79,
"grad_norm": 1.2249259948730469,
"learning_rate": 8.247978436657683e-06,
"loss": 0.0038,
"step": 1890
},
{
"epoch": 27.94,
"grad_norm": 1.6148368120193481,
"learning_rate": 8.231132075471699e-06,
"loss": 0.0034,
"step": 1900
},
{
"epoch": 28.09,
"grad_norm": 1.3190560340881348,
"learning_rate": 8.214285714285714e-06,
"loss": 0.005,
"step": 1910
},
{
"epoch": 28.24,
"grad_norm": 1.7175958156585693,
"learning_rate": 8.197439353099732e-06,
"loss": 0.0037,
"step": 1920
},
{
"epoch": 28.38,
"grad_norm": 1.4584541320800781,
"learning_rate": 8.180592991913748e-06,
"loss": 0.0032,
"step": 1930
},
{
"epoch": 28.53,
"grad_norm": 1.2781018018722534,
"learning_rate": 8.163746630727763e-06,
"loss": 0.0038,
"step": 1940
},
{
"epoch": 28.68,
"grad_norm": 1.805674433708191,
"learning_rate": 8.146900269541779e-06,
"loss": 0.0044,
"step": 1950
},
{
"epoch": 28.82,
"grad_norm": 0.8535717129707336,
"learning_rate": 8.130053908355797e-06,
"loss": 0.0037,
"step": 1960
},
{
"epoch": 28.97,
"grad_norm": 1.6355313062667847,
"learning_rate": 8.113207547169812e-06,
"loss": 0.0033,
"step": 1970
},
{
"epoch": 29.12,
"grad_norm": 0.6453380584716797,
"learning_rate": 8.096361185983828e-06,
"loss": 0.0031,
"step": 1980
},
{
"epoch": 29.26,
"grad_norm": 1.806770920753479,
"learning_rate": 8.079514824797844e-06,
"loss": 0.0038,
"step": 1990
},
{
"epoch": 29.41,
"grad_norm": 0.49916043877601624,
"learning_rate": 8.06266846361186e-06,
"loss": 0.0032,
"step": 2000
},
{
"epoch": 29.41,
"eval_loss": 0.4711141288280487,
"eval_runtime": 394.2364,
"eval_samples_per_second": 1.913,
"eval_steps_per_second": 0.02,
"eval_wer": 42.29249011857708,
"step": 2000
},
{
"epoch": 29.56,
"grad_norm": 1.2711254358291626,
"learning_rate": 8.045822102425877e-06,
"loss": 0.0036,
"step": 2010
},
{
"epoch": 29.71,
"grad_norm": 1.2178030014038086,
"learning_rate": 8.028975741239893e-06,
"loss": 0.0042,
"step": 2020
},
{
"epoch": 29.85,
"grad_norm": 1.2590820789337158,
"learning_rate": 8.012129380053909e-06,
"loss": 0.0033,
"step": 2030
},
{
"epoch": 30.0,
"grad_norm": 0.7739225625991821,
"learning_rate": 7.995283018867925e-06,
"loss": 0.0053,
"step": 2040
},
{
"epoch": 30.15,
"grad_norm": 1.882087230682373,
"learning_rate": 7.978436657681942e-06,
"loss": 0.0043,
"step": 2050
},
{
"epoch": 30.29,
"grad_norm": 1.1462137699127197,
"learning_rate": 7.961590296495958e-06,
"loss": 0.0027,
"step": 2060
},
{
"epoch": 30.44,
"grad_norm": 1.1864248514175415,
"learning_rate": 7.944743935309974e-06,
"loss": 0.003,
"step": 2070
},
{
"epoch": 30.59,
"grad_norm": 0.4730267822742462,
"learning_rate": 7.92789757412399e-06,
"loss": 0.0026,
"step": 2080
},
{
"epoch": 30.74,
"grad_norm": 0.8079692125320435,
"learning_rate": 7.911051212938005e-06,
"loss": 0.002,
"step": 2090
},
{
"epoch": 30.88,
"grad_norm": 0.9878515005111694,
"learning_rate": 7.894204851752023e-06,
"loss": 0.0039,
"step": 2100
},
{
"epoch": 31.03,
"grad_norm": 0.26896196603775024,
"learning_rate": 7.877358490566038e-06,
"loss": 0.0033,
"step": 2110
},
{
"epoch": 31.18,
"grad_norm": 0.5570561289787292,
"learning_rate": 7.860512129380054e-06,
"loss": 0.0041,
"step": 2120
},
{
"epoch": 31.32,
"grad_norm": 1.9286779165267944,
"learning_rate": 7.843665768194072e-06,
"loss": 0.0031,
"step": 2130
},
{
"epoch": 31.47,
"grad_norm": 0.5159671306610107,
"learning_rate": 7.826819407008087e-06,
"loss": 0.002,
"step": 2140
},
{
"epoch": 31.62,
"grad_norm": 0.811177670955658,
"learning_rate": 7.809973045822103e-06,
"loss": 0.003,
"step": 2150
},
{
"epoch": 31.76,
"grad_norm": 0.8233473300933838,
"learning_rate": 7.793126684636119e-06,
"loss": 0.0023,
"step": 2160
},
{
"epoch": 31.91,
"grad_norm": 0.3896019756793976,
"learning_rate": 7.776280323450135e-06,
"loss": 0.0021,
"step": 2170
},
{
"epoch": 32.06,
"grad_norm": 0.37738358974456787,
"learning_rate": 7.75943396226415e-06,
"loss": 0.0024,
"step": 2180
},
{
"epoch": 32.21,
"grad_norm": 1.1048047542572021,
"learning_rate": 7.742587601078168e-06,
"loss": 0.003,
"step": 2190
},
{
"epoch": 32.35,
"grad_norm": 0.4888085126876831,
"learning_rate": 7.725741239892184e-06,
"loss": 0.0018,
"step": 2200
},
{
"epoch": 32.5,
"grad_norm": 0.8060516119003296,
"learning_rate": 7.7088948787062e-06,
"loss": 0.0019,
"step": 2210
},
{
"epoch": 32.65,
"grad_norm": 0.8470051884651184,
"learning_rate": 7.692048517520217e-06,
"loss": 0.0019,
"step": 2220
},
{
"epoch": 32.79,
"grad_norm": 0.9551669955253601,
"learning_rate": 7.675202156334233e-06,
"loss": 0.0027,
"step": 2230
},
{
"epoch": 32.94,
"grad_norm": 1.0600543022155762,
"learning_rate": 7.658355795148249e-06,
"loss": 0.0025,
"step": 2240
},
{
"epoch": 33.09,
"grad_norm": 0.4787401258945465,
"learning_rate": 7.641509433962266e-06,
"loss": 0.0027,
"step": 2250
},
{
"epoch": 33.24,
"grad_norm": 0.7312725782394409,
"learning_rate": 7.62466307277628e-06,
"loss": 0.0024,
"step": 2260
},
{
"epoch": 33.38,
"grad_norm": 1.3341702222824097,
"learning_rate": 7.607816711590297e-06,
"loss": 0.0022,
"step": 2270
},
{
"epoch": 33.53,
"grad_norm": 0.6375266909599304,
"learning_rate": 7.5909703504043134e-06,
"loss": 0.0029,
"step": 2280
},
{
"epoch": 33.68,
"grad_norm": 0.8027037978172302,
"learning_rate": 7.574123989218329e-06,
"loss": 0.0019,
"step": 2290
},
{
"epoch": 33.82,
"grad_norm": 0.45866358280181885,
"learning_rate": 7.557277628032346e-06,
"loss": 0.002,
"step": 2300
},
{
"epoch": 33.97,
"grad_norm": 0.4257550835609436,
"learning_rate": 7.540431266846362e-06,
"loss": 0.0022,
"step": 2310
},
{
"epoch": 34.12,
"grad_norm": 0.6710132360458374,
"learning_rate": 7.523584905660378e-06,
"loss": 0.0027,
"step": 2320
},
{
"epoch": 34.26,
"grad_norm": 0.7800111770629883,
"learning_rate": 7.506738544474395e-06,
"loss": 0.0022,
"step": 2330
},
{
"epoch": 34.41,
"grad_norm": 0.5604603290557861,
"learning_rate": 7.489892183288411e-06,
"loss": 0.0028,
"step": 2340
},
{
"epoch": 34.56,
"grad_norm": 1.3979426622390747,
"learning_rate": 7.473045822102426e-06,
"loss": 0.0024,
"step": 2350
},
{
"epoch": 34.71,
"grad_norm": 1.4295878410339355,
"learning_rate": 7.456199460916442e-06,
"loss": 0.0025,
"step": 2360
},
{
"epoch": 34.85,
"grad_norm": 0.8446325659751892,
"learning_rate": 7.439353099730459e-06,
"loss": 0.0039,
"step": 2370
},
{
"epoch": 35.0,
"grad_norm": 0.7062538266181946,
"learning_rate": 7.422506738544475e-06,
"loss": 0.0024,
"step": 2380
},
{
"epoch": 35.15,
"grad_norm": 0.42034047842025757,
"learning_rate": 7.405660377358491e-06,
"loss": 0.0015,
"step": 2390
},
{
"epoch": 35.29,
"grad_norm": 0.21419081091880798,
"learning_rate": 7.388814016172508e-06,
"loss": 0.0015,
"step": 2400
},
{
"epoch": 35.44,
"grad_norm": 0.2849736213684082,
"learning_rate": 7.371967654986524e-06,
"loss": 0.0008,
"step": 2410
},
{
"epoch": 35.59,
"grad_norm": 0.4012085795402527,
"learning_rate": 7.35512129380054e-06,
"loss": 0.0015,
"step": 2420
},
{
"epoch": 35.74,
"grad_norm": 0.12295886129140854,
"learning_rate": 7.338274932614556e-06,
"loss": 0.0016,
"step": 2430
},
{
"epoch": 35.88,
"grad_norm": 0.5100615620613098,
"learning_rate": 7.321428571428572e-06,
"loss": 0.0017,
"step": 2440
},
{
"epoch": 36.03,
"grad_norm": 0.19803810119628906,
"learning_rate": 7.304582210242588e-06,
"loss": 0.0013,
"step": 2450
},
{
"epoch": 36.18,
"grad_norm": 0.45984306931495667,
"learning_rate": 7.287735849056604e-06,
"loss": 0.0007,
"step": 2460
},
{
"epoch": 36.32,
"grad_norm": 1.1022437810897827,
"learning_rate": 7.27088948787062e-06,
"loss": 0.0012,
"step": 2470
},
{
"epoch": 36.47,
"grad_norm": 0.20512618124485016,
"learning_rate": 7.254043126684637e-06,
"loss": 0.0017,
"step": 2480
},
{
"epoch": 36.62,
"grad_norm": 0.11672288924455643,
"learning_rate": 7.237196765498653e-06,
"loss": 0.0009,
"step": 2490
},
{
"epoch": 36.76,
"grad_norm": 0.41004320979118347,
"learning_rate": 7.220350404312669e-06,
"loss": 0.0005,
"step": 2500
},
{
"epoch": 36.76,
"eval_loss": 0.48888495564460754,
"eval_runtime": 395.4163,
"eval_samples_per_second": 1.907,
"eval_steps_per_second": 0.02,
"eval_wer": 41.22529644268775,
"step": 2500
},
{
"epoch": 36.91,
"grad_norm": 0.41586098074913025,
"learning_rate": 7.203504043126686e-06,
"loss": 0.0006,
"step": 2510
},
{
"epoch": 37.06,
"grad_norm": 0.05540100112557411,
"learning_rate": 7.1866576819407015e-06,
"loss": 0.0004,
"step": 2520
},
{
"epoch": 37.21,
"grad_norm": 0.20431630313396454,
"learning_rate": 7.169811320754717e-06,
"loss": 0.0005,
"step": 2530
},
{
"epoch": 37.35,
"grad_norm": 0.13198405504226685,
"learning_rate": 7.152964959568733e-06,
"loss": 0.0005,
"step": 2540
},
{
"epoch": 37.5,
"grad_norm": 1.1269527673721313,
"learning_rate": 7.13611859838275e-06,
"loss": 0.0006,
"step": 2550
},
{
"epoch": 37.65,
"grad_norm": 0.06169961765408516,
"learning_rate": 7.119272237196766e-06,
"loss": 0.0003,
"step": 2560
},
{
"epoch": 37.79,
"grad_norm": 1.2373998165130615,
"learning_rate": 7.102425876010782e-06,
"loss": 0.0007,
"step": 2570
},
{
"epoch": 37.94,
"grad_norm": 0.730436384677887,
"learning_rate": 7.085579514824799e-06,
"loss": 0.0006,
"step": 2580
},
{
"epoch": 38.09,
"grad_norm": 0.2517056465148926,
"learning_rate": 7.0687331536388145e-06,
"loss": 0.0009,
"step": 2590
},
{
"epoch": 38.24,
"grad_norm": 0.24311576783657074,
"learning_rate": 7.051886792452831e-06,
"loss": 0.0008,
"step": 2600
},
{
"epoch": 38.38,
"grad_norm": 0.46015554666519165,
"learning_rate": 7.035040431266848e-06,
"loss": 0.0008,
"step": 2610
},
{
"epoch": 38.53,
"grad_norm": 0.10832543671131134,
"learning_rate": 7.018194070080863e-06,
"loss": 0.0011,
"step": 2620
},
{
"epoch": 38.68,
"grad_norm": 0.06158292293548584,
"learning_rate": 7.0013477088948785e-06,
"loss": 0.0011,
"step": 2630
},
{
"epoch": 38.82,
"grad_norm": 0.8122970461845398,
"learning_rate": 6.984501347708895e-06,
"loss": 0.0009,
"step": 2640
},
{
"epoch": 38.97,
"grad_norm": 0.7663437128067017,
"learning_rate": 6.967654986522912e-06,
"loss": 0.0008,
"step": 2650
},
{
"epoch": 39.12,
"grad_norm": 0.145246684551239,
"learning_rate": 6.9508086253369275e-06,
"loss": 0.0007,
"step": 2660
},
{
"epoch": 39.26,
"grad_norm": 0.5487465858459473,
"learning_rate": 6.933962264150944e-06,
"loss": 0.0016,
"step": 2670
},
{
"epoch": 39.41,
"grad_norm": 0.9387257695198059,
"learning_rate": 6.917115902964961e-06,
"loss": 0.0009,
"step": 2680
},
{
"epoch": 39.56,
"grad_norm": 0.31192561984062195,
"learning_rate": 6.9002695417789766e-06,
"loss": 0.0008,
"step": 2690
},
{
"epoch": 39.71,
"grad_norm": 0.0915888324379921,
"learning_rate": 6.883423180592993e-06,
"loss": 0.0005,
"step": 2700
},
{
"epoch": 39.85,
"grad_norm": 0.10339212417602539,
"learning_rate": 6.866576819407008e-06,
"loss": 0.0005,
"step": 2710
},
{
"epoch": 40.0,
"grad_norm": 0.07240493595600128,
"learning_rate": 6.849730458221025e-06,
"loss": 0.0008,
"step": 2720
},
{
"epoch": 40.15,
"grad_norm": 0.08370707184076309,
"learning_rate": 6.8328840970350405e-06,
"loss": 0.0009,
"step": 2730
},
{
"epoch": 40.29,
"grad_norm": 0.07619204372167587,
"learning_rate": 6.816037735849057e-06,
"loss": 0.0005,
"step": 2740
},
{
"epoch": 40.44,
"grad_norm": 0.4352577030658722,
"learning_rate": 6.799191374663073e-06,
"loss": 0.0016,
"step": 2750
},
{
"epoch": 40.59,
"grad_norm": 0.6862916946411133,
"learning_rate": 6.78234501347709e-06,
"loss": 0.0006,
"step": 2760
},
{
"epoch": 40.74,
"grad_norm": 0.2726142406463623,
"learning_rate": 6.765498652291106e-06,
"loss": 0.0008,
"step": 2770
},
{
"epoch": 40.88,
"grad_norm": 0.1969071477651596,
"learning_rate": 6.748652291105122e-06,
"loss": 0.0005,
"step": 2780
},
{
"epoch": 41.03,
"grad_norm": 0.341925710439682,
"learning_rate": 6.731805929919139e-06,
"loss": 0.0007,
"step": 2790
},
{
"epoch": 41.18,
"grad_norm": 1.3179571628570557,
"learning_rate": 6.7149595687331536e-06,
"loss": 0.0005,
"step": 2800
},
{
"epoch": 41.32,
"grad_norm": 0.5500785112380981,
"learning_rate": 6.69811320754717e-06,
"loss": 0.0014,
"step": 2810
},
{
"epoch": 41.47,
"grad_norm": 0.30600807070732117,
"learning_rate": 6.681266846361186e-06,
"loss": 0.0008,
"step": 2820
},
{
"epoch": 41.62,
"grad_norm": 0.2631785571575165,
"learning_rate": 6.664420485175203e-06,
"loss": 0.001,
"step": 2830
},
{
"epoch": 41.76,
"grad_norm": 0.8868472576141357,
"learning_rate": 6.647574123989219e-06,
"loss": 0.0009,
"step": 2840
},
{
"epoch": 41.91,
"grad_norm": 1.278052806854248,
"learning_rate": 6.630727762803235e-06,
"loss": 0.0011,
"step": 2850
},
{
"epoch": 42.06,
"grad_norm": 0.1165945902466774,
"learning_rate": 6.613881401617252e-06,
"loss": 0.0014,
"step": 2860
},
{
"epoch": 42.21,
"grad_norm": 0.37418603897094727,
"learning_rate": 6.597035040431267e-06,
"loss": 0.0009,
"step": 2870
},
{
"epoch": 42.35,
"grad_norm": 2.285724401473999,
"learning_rate": 6.580188679245284e-06,
"loss": 0.0013,
"step": 2880
},
{
"epoch": 42.5,
"grad_norm": 0.48481008410453796,
"learning_rate": 6.563342318059299e-06,
"loss": 0.0012,
"step": 2890
},
{
"epoch": 42.65,
"grad_norm": 0.330135703086853,
"learning_rate": 6.546495956873316e-06,
"loss": 0.0007,
"step": 2900
},
{
"epoch": 42.79,
"grad_norm": 0.5341326594352722,
"learning_rate": 6.529649595687332e-06,
"loss": 0.0008,
"step": 2910
},
{
"epoch": 42.94,
"grad_norm": 1.2509163618087769,
"learning_rate": 6.512803234501348e-06,
"loss": 0.0016,
"step": 2920
},
{
"epoch": 43.09,
"grad_norm": 1.3307183980941772,
"learning_rate": 6.495956873315365e-06,
"loss": 0.0015,
"step": 2930
},
{
"epoch": 43.24,
"grad_norm": 0.4615732729434967,
"learning_rate": 6.47911051212938e-06,
"loss": 0.0011,
"step": 2940
},
{
"epoch": 43.38,
"grad_norm": 0.3208690583705902,
"learning_rate": 6.462264150943397e-06,
"loss": 0.001,
"step": 2950
},
{
"epoch": 43.53,
"grad_norm": 0.6121777892112732,
"learning_rate": 6.445417789757414e-06,
"loss": 0.0012,
"step": 2960
},
{
"epoch": 43.68,
"grad_norm": 0.8947989344596863,
"learning_rate": 6.4285714285714295e-06,
"loss": 0.0021,
"step": 2970
},
{
"epoch": 43.82,
"grad_norm": 1.4063951969146729,
"learning_rate": 6.411725067385444e-06,
"loss": 0.0022,
"step": 2980
},
{
"epoch": 43.97,
"grad_norm": 0.9614599347114563,
"learning_rate": 6.394878706199461e-06,
"loss": 0.0022,
"step": 2990
},
{
"epoch": 44.12,
"grad_norm": 0.3544367551803589,
"learning_rate": 6.378032345013478e-06,
"loss": 0.0012,
"step": 3000
},
{
"epoch": 44.12,
"eval_loss": 0.4919167459011078,
"eval_runtime": 412.7274,
"eval_samples_per_second": 1.827,
"eval_steps_per_second": 0.019,
"eval_wer": 42.03557312252965,
"step": 3000
},
{
"epoch": 44.26,
"grad_norm": 0.3670799732208252,
"learning_rate": 6.3611859838274934e-06,
"loss": 0.0011,
"step": 3010
},
{
"epoch": 44.41,
"grad_norm": 0.8639935255050659,
"learning_rate": 6.34433962264151e-06,
"loss": 0.0021,
"step": 3020
},
{
"epoch": 44.56,
"grad_norm": 0.8535746932029724,
"learning_rate": 6.327493261455526e-06,
"loss": 0.0023,
"step": 3030
},
{
"epoch": 44.71,
"grad_norm": 0.791280210018158,
"learning_rate": 6.3106469002695425e-06,
"loss": 0.0017,
"step": 3040
},
{
"epoch": 44.85,
"grad_norm": 1.2309428453445435,
"learning_rate": 6.293800539083559e-06,
"loss": 0.0022,
"step": 3050
},
{
"epoch": 45.0,
"grad_norm": 2.02470326423645,
"learning_rate": 6.276954177897575e-06,
"loss": 0.0023,
"step": 3060
},
{
"epoch": 45.15,
"grad_norm": 0.6730286478996277,
"learning_rate": 6.260107816711591e-06,
"loss": 0.0014,
"step": 3070
},
{
"epoch": 45.29,
"grad_norm": 0.6371138095855713,
"learning_rate": 6.2432614555256064e-06,
"loss": 0.0013,
"step": 3080
},
{
"epoch": 45.44,
"grad_norm": 0.8862447738647461,
"learning_rate": 6.226415094339623e-06,
"loss": 0.002,
"step": 3090
},
{
"epoch": 45.59,
"grad_norm": 0.600082516670227,
"learning_rate": 6.209568733153639e-06,
"loss": 0.0015,
"step": 3100
},
{
"epoch": 45.74,
"grad_norm": 1.0069173574447632,
"learning_rate": 6.1927223719676555e-06,
"loss": 0.0011,
"step": 3110
},
{
"epoch": 45.88,
"grad_norm": 0.8306257724761963,
"learning_rate": 6.175876010781672e-06,
"loss": 0.0017,
"step": 3120
},
{
"epoch": 46.03,
"grad_norm": 0.0885746031999588,
"learning_rate": 6.159029649595688e-06,
"loss": 0.002,
"step": 3130
},
{
"epoch": 46.18,
"grad_norm": 0.2680250108242035,
"learning_rate": 6.1421832884097045e-06,
"loss": 0.0011,
"step": 3140
},
{
"epoch": 46.32,
"grad_norm": 0.5739225745201111,
"learning_rate": 6.12533692722372e-06,
"loss": 0.0006,
"step": 3150
},
{
"epoch": 46.47,
"grad_norm": 0.09838727116584778,
"learning_rate": 6.108490566037736e-06,
"loss": 0.001,
"step": 3160
},
{
"epoch": 46.62,
"grad_norm": 0.6324687600135803,
"learning_rate": 6.091644204851752e-06,
"loss": 0.0012,
"step": 3170
},
{
"epoch": 46.76,
"grad_norm": 0.8236867189407349,
"learning_rate": 6.0747978436657685e-06,
"loss": 0.0018,
"step": 3180
},
{
"epoch": 46.91,
"grad_norm": 1.935333251953125,
"learning_rate": 6.057951482479785e-06,
"loss": 0.0028,
"step": 3190
},
{
"epoch": 47.06,
"grad_norm": 0.9879902601242065,
"learning_rate": 6.041105121293801e-06,
"loss": 0.0025,
"step": 3200
},
{
"epoch": 47.21,
"grad_norm": 0.7782288193702698,
"learning_rate": 6.0242587601078175e-06,
"loss": 0.0017,
"step": 3210
},
{
"epoch": 47.35,
"grad_norm": 0.9317180514335632,
"learning_rate": 6.007412398921833e-06,
"loss": 0.0014,
"step": 3220
},
{
"epoch": 47.5,
"grad_norm": 0.4948224127292633,
"learning_rate": 5.99056603773585e-06,
"loss": 0.0012,
"step": 3230
},
{
"epoch": 47.65,
"grad_norm": 0.5893387198448181,
"learning_rate": 5.9737196765498666e-06,
"loss": 0.0012,
"step": 3240
},
{
"epoch": 47.79,
"grad_norm": 0.22002609074115753,
"learning_rate": 5.9568733153638815e-06,
"loss": 0.0015,
"step": 3250
},
{
"epoch": 47.94,
"grad_norm": 0.4735713005065918,
"learning_rate": 5.940026954177897e-06,
"loss": 0.0008,
"step": 3260
},
{
"epoch": 48.09,
"grad_norm": 0.8969672918319702,
"learning_rate": 5.923180592991914e-06,
"loss": 0.0014,
"step": 3270
},
{
"epoch": 48.24,
"grad_norm": 0.6368213295936584,
"learning_rate": 5.9063342318059305e-06,
"loss": 0.0014,
"step": 3280
},
{
"epoch": 48.38,
"grad_norm": 0.1679704338312149,
"learning_rate": 5.889487870619946e-06,
"loss": 0.0013,
"step": 3290
},
{
"epoch": 48.53,
"grad_norm": 0.9470486044883728,
"learning_rate": 5.872641509433963e-06,
"loss": 0.0017,
"step": 3300
},
{
"epoch": 48.68,
"grad_norm": 0.08789900690317154,
"learning_rate": 5.855795148247979e-06,
"loss": 0.0007,
"step": 3310
},
{
"epoch": 48.82,
"grad_norm": 0.24826857447624207,
"learning_rate": 5.838948787061995e-06,
"loss": 0.0009,
"step": 3320
},
{
"epoch": 48.97,
"grad_norm": 0.11028200387954712,
"learning_rate": 5.822102425876012e-06,
"loss": 0.0009,
"step": 3330
},
{
"epoch": 49.12,
"grad_norm": 0.11782591044902802,
"learning_rate": 5.805256064690027e-06,
"loss": 0.0008,
"step": 3340
},
{
"epoch": 49.26,
"grad_norm": 0.3089038133621216,
"learning_rate": 5.7884097035040435e-06,
"loss": 0.0007,
"step": 3350
},
{
"epoch": 49.41,
"grad_norm": 0.5864781737327576,
"learning_rate": 5.771563342318059e-06,
"loss": 0.0006,
"step": 3360
},
{
"epoch": 49.56,
"grad_norm": 0.6767354607582092,
"learning_rate": 5.754716981132076e-06,
"loss": 0.0004,
"step": 3370
},
{
"epoch": 49.71,
"grad_norm": 0.7366251349449158,
"learning_rate": 5.737870619946092e-06,
"loss": 0.0006,
"step": 3380
},
{
"epoch": 49.85,
"grad_norm": 0.460439532995224,
"learning_rate": 5.721024258760108e-06,
"loss": 0.0006,
"step": 3390
},
{
"epoch": 50.0,
"grad_norm": 0.2541469633579254,
"learning_rate": 5.704177897574125e-06,
"loss": 0.0011,
"step": 3400
},
{
"epoch": 50.15,
"grad_norm": 0.25263655185699463,
"learning_rate": 5.687331536388141e-06,
"loss": 0.0009,
"step": 3410
},
{
"epoch": 50.29,
"grad_norm": 0.3167230188846588,
"learning_rate": 5.670485175202157e-06,
"loss": 0.0005,
"step": 3420
},
{
"epoch": 50.44,
"grad_norm": 0.051995839923620224,
"learning_rate": 5.653638814016172e-06,
"loss": 0.0005,
"step": 3430
},
{
"epoch": 50.59,
"grad_norm": 0.23166793584823608,
"learning_rate": 5.636792452830189e-06,
"loss": 0.0012,
"step": 3440
},
{
"epoch": 50.74,
"grad_norm": 0.46085381507873535,
"learning_rate": 5.619946091644205e-06,
"loss": 0.0011,
"step": 3450
},
{
"epoch": 50.88,
"grad_norm": 0.07778172194957733,
"learning_rate": 5.603099730458221e-06,
"loss": 0.0006,
"step": 3460
},
{
"epoch": 51.03,
"grad_norm": 0.23257611691951752,
"learning_rate": 5.586253369272238e-06,
"loss": 0.0009,
"step": 3470
},
{
"epoch": 51.18,
"grad_norm": 0.6834172606468201,
"learning_rate": 5.569407008086254e-06,
"loss": 0.0006,
"step": 3480
},
{
"epoch": 51.32,
"grad_norm": 0.028312204405665398,
"learning_rate": 5.55256064690027e-06,
"loss": 0.0008,
"step": 3490
},
{
"epoch": 51.47,
"grad_norm": 0.053247835487127304,
"learning_rate": 5.535714285714286e-06,
"loss": 0.0004,
"step": 3500
},
{
"epoch": 51.47,
"eval_loss": 0.5013700127601624,
"eval_runtime": 393.4347,
"eval_samples_per_second": 1.916,
"eval_steps_per_second": 0.02,
"eval_wer": 40.39525691699605,
"step": 3500
},
{
"epoch": 51.62,
"grad_norm": 0.09007065743207932,
"learning_rate": 5.518867924528303e-06,
"loss": 0.0003,
"step": 3510
},
{
"epoch": 51.76,
"grad_norm": 0.03709765151143074,
"learning_rate": 5.502021563342318e-06,
"loss": 0.0003,
"step": 3520
},
{
"epoch": 51.91,
"grad_norm": 0.024563109502196312,
"learning_rate": 5.485175202156334e-06,
"loss": 0.0002,
"step": 3530
},
{
"epoch": 52.06,
"grad_norm": 0.38505399227142334,
"learning_rate": 5.46832884097035e-06,
"loss": 0.0007,
"step": 3540
},
{
"epoch": 52.21,
"grad_norm": 0.041454803198575974,
"learning_rate": 5.451482479784367e-06,
"loss": 0.0002,
"step": 3550
},
{
"epoch": 52.35,
"grad_norm": 0.03118882142007351,
"learning_rate": 5.4346361185983834e-06,
"loss": 0.0002,
"step": 3560
},
{
"epoch": 52.5,
"grad_norm": 0.1050623282790184,
"learning_rate": 5.417789757412399e-06,
"loss": 0.0006,
"step": 3570
},
{
"epoch": 52.65,
"grad_norm": 0.019064374268054962,
"learning_rate": 5.400943396226416e-06,
"loss": 0.0001,
"step": 3580
},
{
"epoch": 52.79,
"grad_norm": 0.020824899896979332,
"learning_rate": 5.384097035040432e-06,
"loss": 0.0001,
"step": 3590
},
{
"epoch": 52.94,
"grad_norm": 0.019172310829162598,
"learning_rate": 5.367250673854448e-06,
"loss": 0.0002,
"step": 3600
},
{
"epoch": 53.09,
"grad_norm": 0.015175268054008484,
"learning_rate": 5.350404312668463e-06,
"loss": 0.0003,
"step": 3610
},
{
"epoch": 53.24,
"grad_norm": 0.03325885161757469,
"learning_rate": 5.33355795148248e-06,
"loss": 0.0004,
"step": 3620
},
{
"epoch": 53.38,
"grad_norm": 0.02436959184706211,
"learning_rate": 5.3167115902964964e-06,
"loss": 0.0002,
"step": 3630
},
{
"epoch": 53.53,
"grad_norm": 0.017924895510077477,
"learning_rate": 5.299865229110512e-06,
"loss": 0.0001,
"step": 3640
},
{
"epoch": 53.68,
"grad_norm": 0.02767987549304962,
"learning_rate": 5.283018867924529e-06,
"loss": 0.0001,
"step": 3650
},
{
"epoch": 53.82,
"grad_norm": 0.01485258899629116,
"learning_rate": 5.266172506738545e-06,
"loss": 0.0003,
"step": 3660
},
{
"epoch": 53.97,
"grad_norm": 0.013557419180870056,
"learning_rate": 5.249326145552561e-06,
"loss": 0.0001,
"step": 3670
},
{
"epoch": 54.12,
"grad_norm": 0.010153830982744694,
"learning_rate": 5.232479784366578e-06,
"loss": 0.0002,
"step": 3680
},
{
"epoch": 54.26,
"grad_norm": 0.01170294638723135,
"learning_rate": 5.215633423180594e-06,
"loss": 0.0001,
"step": 3690
},
{
"epoch": 54.41,
"grad_norm": 0.008615361526608467,
"learning_rate": 5.19878706199461e-06,
"loss": 0.0001,
"step": 3700
},
{
"epoch": 54.56,
"grad_norm": 0.00960615649819374,
"learning_rate": 5.181940700808625e-06,
"loss": 0.0001,
"step": 3710
},
{
"epoch": 54.71,
"grad_norm": 0.008886042051017284,
"learning_rate": 5.165094339622642e-06,
"loss": 0.0001,
"step": 3720
},
{
"epoch": 54.85,
"grad_norm": 0.014573791064321995,
"learning_rate": 5.148247978436658e-06,
"loss": 0.0001,
"step": 3730
},
{
"epoch": 55.0,
"grad_norm": 0.008440797217190266,
"learning_rate": 5.131401617250674e-06,
"loss": 0.0001,
"step": 3740
},
{
"epoch": 55.15,
"grad_norm": 0.010030324570834637,
"learning_rate": 5.114555256064691e-06,
"loss": 0.0001,
"step": 3750
},
{
"epoch": 55.29,
"grad_norm": 0.0625586286187172,
"learning_rate": 5.097708894878707e-06,
"loss": 0.0001,
"step": 3760
},
{
"epoch": 55.44,
"grad_norm": 0.011248340830206871,
"learning_rate": 5.080862533692723e-06,
"loss": 0.0001,
"step": 3770
},
{
"epoch": 55.59,
"grad_norm": 0.006955728866159916,
"learning_rate": 5.064016172506739e-06,
"loss": 0.0001,
"step": 3780
},
{
"epoch": 55.74,
"grad_norm": 0.00837327353656292,
"learning_rate": 5.047169811320756e-06,
"loss": 0.0001,
"step": 3790
},
{
"epoch": 55.88,
"grad_norm": 0.008046143688261509,
"learning_rate": 5.030323450134771e-06,
"loss": 0.0001,
"step": 3800
},
{
"epoch": 56.03,
"grad_norm": 0.007838546298444271,
"learning_rate": 5.013477088948787e-06,
"loss": 0.0001,
"step": 3810
},
{
"epoch": 56.18,
"grad_norm": 0.006525520700961351,
"learning_rate": 4.996630727762803e-06,
"loss": 0.0001,
"step": 3820
},
{
"epoch": 56.32,
"grad_norm": 0.005834328010678291,
"learning_rate": 4.97978436657682e-06,
"loss": 0.0001,
"step": 3830
},
{
"epoch": 56.47,
"grad_norm": 0.01387355849146843,
"learning_rate": 4.962938005390836e-06,
"loss": 0.0003,
"step": 3840
},
{
"epoch": 56.62,
"grad_norm": 0.01750769093632698,
"learning_rate": 4.946091644204852e-06,
"loss": 0.0001,
"step": 3850
},
{
"epoch": 56.76,
"grad_norm": 0.018308816477656364,
"learning_rate": 4.929245283018868e-06,
"loss": 0.0001,
"step": 3860
},
{
"epoch": 56.91,
"grad_norm": 0.0117345554754138,
"learning_rate": 4.9123989218328845e-06,
"loss": 0.0001,
"step": 3870
},
{
"epoch": 57.06,
"grad_norm": 0.007436331827193499,
"learning_rate": 4.8955525606469e-06,
"loss": 0.0001,
"step": 3880
},
{
"epoch": 57.21,
"grad_norm": 0.00813203677535057,
"learning_rate": 4.878706199460917e-06,
"loss": 0.0001,
"step": 3890
},
{
"epoch": 57.35,
"grad_norm": 0.005871808156371117,
"learning_rate": 4.8618598382749335e-06,
"loss": 0.0001,
"step": 3900
},
{
"epoch": 57.5,
"grad_norm": 0.006669571157544851,
"learning_rate": 4.845013477088949e-06,
"loss": 0.0001,
"step": 3910
},
{
"epoch": 57.65,
"grad_norm": 0.004995842929929495,
"learning_rate": 4.828167115902965e-06,
"loss": 0.0001,
"step": 3920
},
{
"epoch": 57.79,
"grad_norm": 0.0052757407538592815,
"learning_rate": 4.811320754716982e-06,
"loss": 0.0001,
"step": 3930
},
{
"epoch": 57.94,
"grad_norm": 0.005870501976460218,
"learning_rate": 4.7944743935309975e-06,
"loss": 0.0001,
"step": 3940
},
{
"epoch": 58.09,
"grad_norm": 0.007945972494781017,
"learning_rate": 4.777628032345013e-06,
"loss": 0.0001,
"step": 3950
},
{
"epoch": 58.24,
"grad_norm": 0.005463455803692341,
"learning_rate": 4.76078167115903e-06,
"loss": 0.0001,
"step": 3960
},
{
"epoch": 58.38,
"grad_norm": 0.0040795220993459225,
"learning_rate": 4.7439353099730466e-06,
"loss": 0.0001,
"step": 3970
},
{
"epoch": 58.53,
"grad_norm": 0.004236205480992794,
"learning_rate": 4.727088948787062e-06,
"loss": 0.0001,
"step": 3980
},
{
"epoch": 58.68,
"grad_norm": 0.004899207036942244,
"learning_rate": 4.710242587601079e-06,
"loss": 0.0001,
"step": 3990
},
{
"epoch": 58.82,
"grad_norm": 0.006042791530489922,
"learning_rate": 4.693396226415095e-06,
"loss": 0.0001,
"step": 4000
},
{
"epoch": 58.82,
"eval_loss": 0.5202410221099854,
"eval_runtime": 391.5224,
"eval_samples_per_second": 1.926,
"eval_steps_per_second": 0.02,
"eval_wer": 40.07905138339921,
"step": 4000
},
{
"epoch": 58.97,
"grad_norm": 0.004714336711913347,
"learning_rate": 4.6765498652291105e-06,
"loss": 0.0001,
"step": 4010
},
{
"epoch": 59.12,
"grad_norm": 0.004793087020516396,
"learning_rate": 4.659703504043127e-06,
"loss": 0.0001,
"step": 4020
},
{
"epoch": 59.26,
"grad_norm": 0.0044836923480033875,
"learning_rate": 4.642857142857144e-06,
"loss": 0.0001,
"step": 4030
},
{
"epoch": 59.41,
"grad_norm": 0.004963328130543232,
"learning_rate": 4.626010781671159e-06,
"loss": 0.0001,
"step": 4040
},
{
"epoch": 59.56,
"grad_norm": 0.003358585759997368,
"learning_rate": 4.609164420485175e-06,
"loss": 0.0001,
"step": 4050
},
{
"epoch": 59.71,
"grad_norm": 0.004264355171471834,
"learning_rate": 4.592318059299192e-06,
"loss": 0.0001,
"step": 4060
},
{
"epoch": 59.85,
"grad_norm": 0.0040533156134188175,
"learning_rate": 4.575471698113208e-06,
"loss": 0.0001,
"step": 4070
},
{
"epoch": 60.0,
"grad_norm": 0.004796077497303486,
"learning_rate": 4.558625336927224e-06,
"loss": 0.0001,
"step": 4080
},
{
"epoch": 60.15,
"grad_norm": 0.0041580419056117535,
"learning_rate": 4.54177897574124e-06,
"loss": 0.0001,
"step": 4090
},
{
"epoch": 60.29,
"grad_norm": 0.004052949137985706,
"learning_rate": 4.524932614555256e-06,
"loss": 0.0,
"step": 4100
},
{
"epoch": 60.44,
"grad_norm": 0.004524200223386288,
"learning_rate": 4.5080862533692726e-06,
"loss": 0.0,
"step": 4110
},
{
"epoch": 60.59,
"grad_norm": 0.0040616546757519245,
"learning_rate": 4.491239892183289e-06,
"loss": 0.0,
"step": 4120
},
{
"epoch": 60.74,
"grad_norm": 0.004815140273422003,
"learning_rate": 4.474393530997305e-06,
"loss": 0.0001,
"step": 4130
},
{
"epoch": 60.88,
"grad_norm": 0.0043374099768698215,
"learning_rate": 4.457547169811321e-06,
"loss": 0.0,
"step": 4140
},
{
"epoch": 61.03,
"grad_norm": 0.003582689445465803,
"learning_rate": 4.440700808625337e-06,
"loss": 0.0,
"step": 4150
},
{
"epoch": 61.18,
"grad_norm": 0.00396451773121953,
"learning_rate": 4.423854447439353e-06,
"loss": 0.0,
"step": 4160
},
{
"epoch": 61.32,
"grad_norm": 0.004014734644442797,
"learning_rate": 4.40700808625337e-06,
"loss": 0.0,
"step": 4170
},
{
"epoch": 61.47,
"grad_norm": 0.0032315212301909924,
"learning_rate": 4.390161725067386e-06,
"loss": 0.0,
"step": 4180
},
{
"epoch": 61.62,
"grad_norm": 0.0032560701947659254,
"learning_rate": 4.373315363881402e-06,
"loss": 0.0001,
"step": 4190
},
{
"epoch": 61.76,
"grad_norm": 0.0036136566195636988,
"learning_rate": 4.356469002695418e-06,
"loss": 0.0,
"step": 4200
},
{
"epoch": 61.91,
"grad_norm": 0.004307006951421499,
"learning_rate": 4.339622641509435e-06,
"loss": 0.0,
"step": 4210
},
{
"epoch": 62.06,
"grad_norm": 0.00355698075145483,
"learning_rate": 4.32277628032345e-06,
"loss": 0.0,
"step": 4220
},
{
"epoch": 62.21,
"grad_norm": 0.003031729022040963,
"learning_rate": 4.305929919137466e-06,
"loss": 0.0,
"step": 4230
},
{
"epoch": 62.35,
"grad_norm": 0.003886875230818987,
"learning_rate": 4.289083557951483e-06,
"loss": 0.0,
"step": 4240
},
{
"epoch": 62.5,
"grad_norm": 0.003472729818895459,
"learning_rate": 4.2722371967654994e-06,
"loss": 0.0,
"step": 4250
},
{
"epoch": 62.65,
"grad_norm": 0.004264041781425476,
"learning_rate": 4.255390835579515e-06,
"loss": 0.0,
"step": 4260
},
{
"epoch": 62.79,
"grad_norm": 0.0033135716803371906,
"learning_rate": 4.238544474393531e-06,
"loss": 0.0,
"step": 4270
},
{
"epoch": 62.94,
"grad_norm": 0.003554994473233819,
"learning_rate": 4.221698113207548e-06,
"loss": 0.0,
"step": 4280
},
{
"epoch": 63.09,
"grad_norm": 0.0036291780415922403,
"learning_rate": 4.204851752021563e-06,
"loss": 0.0,
"step": 4290
},
{
"epoch": 63.24,
"grad_norm": 0.003123074769973755,
"learning_rate": 4.18800539083558e-06,
"loss": 0.0,
"step": 4300
},
{
"epoch": 63.38,
"grad_norm": 0.0030143638141453266,
"learning_rate": 4.171159029649597e-06,
"loss": 0.0,
"step": 4310
},
{
"epoch": 63.53,
"grad_norm": 0.0033236437011510134,
"learning_rate": 4.1543126684636125e-06,
"loss": 0.0,
"step": 4320
},
{
"epoch": 63.68,
"grad_norm": 0.003377540735527873,
"learning_rate": 4.137466307277628e-06,
"loss": 0.0,
"step": 4330
},
{
"epoch": 63.82,
"grad_norm": 0.0034374604001641273,
"learning_rate": 4.120619946091645e-06,
"loss": 0.0,
"step": 4340
},
{
"epoch": 63.97,
"grad_norm": 0.0032602883875370026,
"learning_rate": 4.103773584905661e-06,
"loss": 0.0,
"step": 4350
},
{
"epoch": 64.12,
"grad_norm": 0.003039925592020154,
"learning_rate": 4.086927223719676e-06,
"loss": 0.0,
"step": 4360
},
{
"epoch": 64.26,
"grad_norm": 0.003154992824420333,
"learning_rate": 4.070080862533693e-06,
"loss": 0.0,
"step": 4370
},
{
"epoch": 64.41,
"grad_norm": 0.0030352873727679253,
"learning_rate": 4.053234501347709e-06,
"loss": 0.0,
"step": 4380
},
{
"epoch": 64.56,
"grad_norm": 0.002541678724810481,
"learning_rate": 4.0363881401617255e-06,
"loss": 0.0,
"step": 4390
},
{
"epoch": 64.71,
"grad_norm": 0.0035515627823770046,
"learning_rate": 4.019541778975742e-06,
"loss": 0.0,
"step": 4400
},
{
"epoch": 64.85,
"grad_norm": 0.0032643494196236134,
"learning_rate": 4.002695417789758e-06,
"loss": 0.0,
"step": 4410
},
{
"epoch": 65.0,
"grad_norm": 0.0029830934945493937,
"learning_rate": 3.985849056603774e-06,
"loss": 0.0,
"step": 4420
},
{
"epoch": 65.15,
"grad_norm": 0.003434343496337533,
"learning_rate": 3.96900269541779e-06,
"loss": 0.0,
"step": 4430
},
{
"epoch": 65.29,
"grad_norm": 0.003297910327091813,
"learning_rate": 3.952156334231806e-06,
"loss": 0.0,
"step": 4440
},
{
"epoch": 65.44,
"grad_norm": 0.0029679101426154375,
"learning_rate": 3.935309973045822e-06,
"loss": 0.0,
"step": 4450
},
{
"epoch": 65.59,
"grad_norm": 0.002996984403580427,
"learning_rate": 3.9184636118598385e-06,
"loss": 0.0,
"step": 4460
},
{
"epoch": 65.74,
"grad_norm": 0.0031057405285537243,
"learning_rate": 3.901617250673855e-06,
"loss": 0.0,
"step": 4470
},
{
"epoch": 65.88,
"grad_norm": 0.0026742741465568542,
"learning_rate": 3.884770889487871e-06,
"loss": 0.0,
"step": 4480
},
{
"epoch": 66.03,
"grad_norm": 0.003236331744119525,
"learning_rate": 3.8679245283018875e-06,
"loss": 0.0,
"step": 4490
},
{
"epoch": 66.18,
"grad_norm": 0.002975077135488391,
"learning_rate": 3.851078167115903e-06,
"loss": 0.0,
"step": 4500
},
{
"epoch": 66.18,
"eval_loss": 0.5316298604011536,
"eval_runtime": 394.5635,
"eval_samples_per_second": 1.911,
"eval_steps_per_second": 0.02,
"eval_wer": 39.80237154150198,
"step": 4500
},
{
"epoch": 66.32,
"grad_norm": 0.0026845140382647514,
"learning_rate": 3.834231805929919e-06,
"loss": 0.0,
"step": 4510
},
{
"epoch": 66.47,
"grad_norm": 0.002751615596935153,
"learning_rate": 3.817385444743936e-06,
"loss": 0.0,
"step": 4520
},
{
"epoch": 66.62,
"grad_norm": 0.0025983687955886126,
"learning_rate": 3.800539083557952e-06,
"loss": 0.0,
"step": 4530
},
{
"epoch": 66.76,
"grad_norm": 0.0027817366644740105,
"learning_rate": 3.7836927223719677e-06,
"loss": 0.0,
"step": 4540
},
{
"epoch": 66.91,
"grad_norm": 0.0028080460615456104,
"learning_rate": 3.766846361185984e-06,
"loss": 0.0,
"step": 4550
},
{
"epoch": 67.06,
"grad_norm": 0.0026772802229970694,
"learning_rate": 3.7500000000000005e-06,
"loss": 0.0,
"step": 4560
},
{
"epoch": 67.21,
"grad_norm": 0.002854007761925459,
"learning_rate": 3.7331536388140167e-06,
"loss": 0.0,
"step": 4570
},
{
"epoch": 67.35,
"grad_norm": 0.0026527883019298315,
"learning_rate": 3.716307277628033e-06,
"loss": 0.0,
"step": 4580
},
{
"epoch": 67.5,
"grad_norm": 0.002448031213134527,
"learning_rate": 3.6994609164420487e-06,
"loss": 0.0,
"step": 4590
},
{
"epoch": 67.65,
"grad_norm": 0.002617811318486929,
"learning_rate": 3.682614555256065e-06,
"loss": 0.0,
"step": 4600
},
{
"epoch": 67.79,
"grad_norm": 0.0031759808771312237,
"learning_rate": 3.665768194070081e-06,
"loss": 0.0,
"step": 4610
},
{
"epoch": 67.94,
"grad_norm": 0.002507594181224704,
"learning_rate": 3.6489218328840973e-06,
"loss": 0.0,
"step": 4620
},
{
"epoch": 68.09,
"grad_norm": 0.0023514223285019398,
"learning_rate": 3.632075471698113e-06,
"loss": 0.0,
"step": 4630
},
{
"epoch": 68.24,
"grad_norm": 0.0027671242132782936,
"learning_rate": 3.6152291105121297e-06,
"loss": 0.0,
"step": 4640
},
{
"epoch": 68.38,
"grad_norm": 0.002477505709975958,
"learning_rate": 3.598382749326146e-06,
"loss": 0.0,
"step": 4650
},
{
"epoch": 68.53,
"grad_norm": 0.0028106619138270617,
"learning_rate": 3.581536388140162e-06,
"loss": 0.0,
"step": 4660
},
{
"epoch": 68.68,
"grad_norm": 0.0034189883153885603,
"learning_rate": 3.5646900269541783e-06,
"loss": 0.0,
"step": 4670
},
{
"epoch": 68.82,
"grad_norm": 0.0030117544811218977,
"learning_rate": 3.547843665768194e-06,
"loss": 0.0,
"step": 4680
},
{
"epoch": 68.97,
"grad_norm": 0.002722077304497361,
"learning_rate": 3.5309973045822103e-06,
"loss": 0.0,
"step": 4690
},
{
"epoch": 69.12,
"grad_norm": 0.00265983073040843,
"learning_rate": 3.514150943396227e-06,
"loss": 0.0,
"step": 4700
},
{
"epoch": 69.26,
"grad_norm": 0.002687458647415042,
"learning_rate": 3.497304582210243e-06,
"loss": 0.0,
"step": 4710
},
{
"epoch": 69.41,
"grad_norm": 0.002611985895782709,
"learning_rate": 3.480458221024259e-06,
"loss": 0.0,
"step": 4720
},
{
"epoch": 69.56,
"grad_norm": 0.002745421137660742,
"learning_rate": 3.463611859838275e-06,
"loss": 0.0,
"step": 4730
},
{
"epoch": 69.71,
"grad_norm": 0.00274544395506382,
"learning_rate": 3.4467654986522914e-06,
"loss": 0.0,
"step": 4740
},
{
"epoch": 69.85,
"grad_norm": 0.0032134081702679396,
"learning_rate": 3.4299191374663076e-06,
"loss": 0.0,
"step": 4750
},
{
"epoch": 70.0,
"grad_norm": 0.0027218926697969437,
"learning_rate": 3.4130727762803238e-06,
"loss": 0.0,
"step": 4760
},
{
"epoch": 70.15,
"grad_norm": 0.0029763532802462578,
"learning_rate": 3.3962264150943395e-06,
"loss": 0.0,
"step": 4770
},
{
"epoch": 70.29,
"grad_norm": 0.002971038920804858,
"learning_rate": 3.379380053908356e-06,
"loss": 0.0,
"step": 4780
},
{
"epoch": 70.44,
"grad_norm": 0.002593262353911996,
"learning_rate": 3.3625336927223724e-06,
"loss": 0.0,
"step": 4790
},
{
"epoch": 70.59,
"grad_norm": 0.0025443334598094225,
"learning_rate": 3.3456873315363886e-06,
"loss": 0.0,
"step": 4800
},
{
"epoch": 70.74,
"grad_norm": 0.002275052247568965,
"learning_rate": 3.3288409703504044e-06,
"loss": 0.0,
"step": 4810
},
{
"epoch": 70.88,
"grad_norm": 0.0028011808171868324,
"learning_rate": 3.3119946091644206e-06,
"loss": 0.0,
"step": 4820
},
{
"epoch": 71.03,
"grad_norm": 0.0028713145293295383,
"learning_rate": 3.2951482479784368e-06,
"loss": 0.0,
"step": 4830
},
{
"epoch": 71.18,
"grad_norm": 0.00250536622479558,
"learning_rate": 3.2783018867924534e-06,
"loss": 0.0,
"step": 4840
},
{
"epoch": 71.32,
"grad_norm": 0.002349557587876916,
"learning_rate": 3.2614555256064696e-06,
"loss": 0.0,
"step": 4850
},
{
"epoch": 71.47,
"grad_norm": 0.002672821283340454,
"learning_rate": 3.2446091644204854e-06,
"loss": 0.0,
"step": 4860
},
{
"epoch": 71.62,
"grad_norm": 0.002757046138867736,
"learning_rate": 3.2277628032345016e-06,
"loss": 0.0,
"step": 4870
},
{
"epoch": 71.76,
"grad_norm": 0.002534884260967374,
"learning_rate": 3.210916442048518e-06,
"loss": 0.0,
"step": 4880
},
{
"epoch": 71.91,
"grad_norm": 0.0023479722440242767,
"learning_rate": 3.194070080862534e-06,
"loss": 0.0,
"step": 4890
},
{
"epoch": 72.06,
"grad_norm": 0.0025629138108342886,
"learning_rate": 3.1772237196765498e-06,
"loss": 0.0,
"step": 4900
},
{
"epoch": 72.21,
"grad_norm": 0.002501903334632516,
"learning_rate": 3.160377358490566e-06,
"loss": 0.0,
"step": 4910
},
{
"epoch": 72.35,
"grad_norm": 0.002238241257146001,
"learning_rate": 3.1435309973045826e-06,
"loss": 0.0,
"step": 4920
},
{
"epoch": 72.5,
"grad_norm": 0.0019617443904280663,
"learning_rate": 3.126684636118599e-06,
"loss": 0.0,
"step": 4930
},
{
"epoch": 72.65,
"grad_norm": 0.0025888625532388687,
"learning_rate": 3.109838274932615e-06,
"loss": 0.0,
"step": 4940
},
{
"epoch": 72.79,
"grad_norm": 0.0022458680905401707,
"learning_rate": 3.092991913746631e-06,
"loss": 0.0,
"step": 4950
},
{
"epoch": 72.94,
"grad_norm": 0.0028502140194177628,
"learning_rate": 3.076145552560647e-06,
"loss": 0.0,
"step": 4960
},
{
"epoch": 73.09,
"grad_norm": 0.0024815101642161608,
"learning_rate": 3.0592991913746632e-06,
"loss": 0.0,
"step": 4970
},
{
"epoch": 73.24,
"grad_norm": 0.0020565763115882874,
"learning_rate": 3.04245283018868e-06,
"loss": 0.0,
"step": 4980
},
{
"epoch": 73.38,
"grad_norm": 0.0019868898671120405,
"learning_rate": 3.025606469002695e-06,
"loss": 0.0,
"step": 4990
},
{
"epoch": 73.53,
"grad_norm": 0.002612064126878977,
"learning_rate": 3.008760107816712e-06,
"loss": 0.0,
"step": 5000
},
{
"epoch": 73.53,
"eval_loss": 0.5392932891845703,
"eval_runtime": 438.3941,
"eval_samples_per_second": 1.72,
"eval_steps_per_second": 0.018,
"eval_wer": 39.901185770750985,
"step": 5000
},
{
"epoch": 73.68,
"grad_norm": 0.0019985686521977186,
"learning_rate": 2.991913746630728e-06,
"loss": 0.0,
"step": 5010
},
{
"epoch": 73.82,
"grad_norm": 0.0021473197266459465,
"learning_rate": 2.9750673854447442e-06,
"loss": 0.0,
"step": 5020
},
{
"epoch": 73.97,
"grad_norm": 0.0023225159384310246,
"learning_rate": 2.9582210242587605e-06,
"loss": 0.0,
"step": 5030
},
{
"epoch": 74.12,
"grad_norm": 0.0022743339650332928,
"learning_rate": 2.9413746630727762e-06,
"loss": 0.0,
"step": 5040
},
{
"epoch": 74.26,
"grad_norm": 0.002134987385943532,
"learning_rate": 2.9245283018867924e-06,
"loss": 0.0,
"step": 5050
},
{
"epoch": 74.41,
"grad_norm": 0.0021126819774508476,
"learning_rate": 2.907681940700809e-06,
"loss": 0.0,
"step": 5060
},
{
"epoch": 74.56,
"grad_norm": 0.002264765091240406,
"learning_rate": 2.8908355795148253e-06,
"loss": 0.0,
"step": 5070
},
{
"epoch": 74.71,
"grad_norm": 0.0028144866228103638,
"learning_rate": 2.8739892183288415e-06,
"loss": 0.0,
"step": 5080
},
{
"epoch": 74.85,
"grad_norm": 0.0017289798706769943,
"learning_rate": 2.8571428571428573e-06,
"loss": 0.0,
"step": 5090
},
{
"epoch": 75.0,
"grad_norm": 0.00249221152625978,
"learning_rate": 2.8402964959568735e-06,
"loss": 0.0,
"step": 5100
},
{
"epoch": 75.15,
"grad_norm": 0.0018814760260283947,
"learning_rate": 2.8234501347708897e-06,
"loss": 0.0,
"step": 5110
},
{
"epoch": 75.29,
"grad_norm": 0.0022003494668751955,
"learning_rate": 2.8066037735849063e-06,
"loss": 0.0,
"step": 5120
},
{
"epoch": 75.44,
"grad_norm": 0.0023006245028227568,
"learning_rate": 2.7897574123989217e-06,
"loss": 0.0,
"step": 5130
},
{
"epoch": 75.59,
"grad_norm": 0.0018226418178528547,
"learning_rate": 2.7729110512129383e-06,
"loss": 0.0,
"step": 5140
},
{
"epoch": 75.74,
"grad_norm": 0.0019806609489023685,
"learning_rate": 2.7560646900269545e-06,
"loss": 0.0,
"step": 5150
},
{
"epoch": 75.88,
"grad_norm": 0.0021330551244318485,
"learning_rate": 2.7392183288409707e-06,
"loss": 0.0,
"step": 5160
},
{
"epoch": 76.03,
"grad_norm": 0.0021852573845535517,
"learning_rate": 2.722371967654987e-06,
"loss": 0.0,
"step": 5170
},
{
"epoch": 76.18,
"grad_norm": 0.0021261894144117832,
"learning_rate": 2.7055256064690027e-06,
"loss": 0.0,
"step": 5180
},
{
"epoch": 76.32,
"grad_norm": 0.0022390589583665133,
"learning_rate": 2.688679245283019e-06,
"loss": 0.0,
"step": 5190
},
{
"epoch": 76.47,
"grad_norm": 0.002223338931798935,
"learning_rate": 2.6718328840970355e-06,
"loss": 0.0,
"step": 5200
},
{
"epoch": 76.62,
"grad_norm": 0.0019250946352258325,
"learning_rate": 2.6549865229110517e-06,
"loss": 0.0,
"step": 5210
},
{
"epoch": 76.76,
"grad_norm": 0.001935865031555295,
"learning_rate": 2.6381401617250675e-06,
"loss": 0.0,
"step": 5220
},
{
"epoch": 76.91,
"grad_norm": 0.0018222652142867446,
"learning_rate": 2.6212938005390837e-06,
"loss": 0.0,
"step": 5230
},
{
"epoch": 77.06,
"grad_norm": 0.0015805804869160056,
"learning_rate": 2.6044474393531e-06,
"loss": 0.0,
"step": 5240
},
{
"epoch": 77.21,
"grad_norm": 0.0018061785958707333,
"learning_rate": 2.587601078167116e-06,
"loss": 0.0,
"step": 5250
},
{
"epoch": 77.35,
"grad_norm": 0.0019492105348035693,
"learning_rate": 2.5707547169811327e-06,
"loss": 0.0,
"step": 5260
},
{
"epoch": 77.5,
"grad_norm": 0.0020225332118570805,
"learning_rate": 2.553908355795148e-06,
"loss": 0.0,
"step": 5270
},
{
"epoch": 77.65,
"grad_norm": 0.0021102512255311012,
"learning_rate": 2.5370619946091647e-06,
"loss": 0.0,
"step": 5280
},
{
"epoch": 77.79,
"grad_norm": 0.001952395774424076,
"learning_rate": 2.520215633423181e-06,
"loss": 0.0,
"step": 5290
},
{
"epoch": 77.94,
"grad_norm": 0.0022354356478899717,
"learning_rate": 2.503369272237197e-06,
"loss": 0.0,
"step": 5300
},
{
"epoch": 78.09,
"grad_norm": 0.0015943964244797826,
"learning_rate": 2.4865229110512133e-06,
"loss": 0.0,
"step": 5310
},
{
"epoch": 78.24,
"grad_norm": 0.00202794186770916,
"learning_rate": 2.469676549865229e-06,
"loss": 0.0,
"step": 5320
},
{
"epoch": 78.38,
"grad_norm": 0.002233139704912901,
"learning_rate": 2.4528301886792453e-06,
"loss": 0.0,
"step": 5330
},
{
"epoch": 78.53,
"grad_norm": 0.0019945132080465555,
"learning_rate": 2.435983827493262e-06,
"loss": 0.0,
"step": 5340
},
{
"epoch": 78.68,
"grad_norm": 0.001916647655889392,
"learning_rate": 2.4191374663072777e-06,
"loss": 0.0,
"step": 5350
},
{
"epoch": 78.82,
"grad_norm": 0.0020198444835841656,
"learning_rate": 2.402291105121294e-06,
"loss": 0.0,
"step": 5360
},
{
"epoch": 78.97,
"grad_norm": 0.0019901052583009005,
"learning_rate": 2.38544474393531e-06,
"loss": 0.0,
"step": 5370
},
{
"epoch": 79.12,
"grad_norm": 0.0020235173869878054,
"learning_rate": 2.3685983827493263e-06,
"loss": 0.0,
"step": 5380
},
{
"epoch": 79.26,
"grad_norm": 0.0018330395687371492,
"learning_rate": 2.3517520215633426e-06,
"loss": 0.0,
"step": 5390
},
{
"epoch": 79.41,
"grad_norm": 0.0023276926949620247,
"learning_rate": 2.3349056603773588e-06,
"loss": 0.0,
"step": 5400
},
{
"epoch": 79.56,
"grad_norm": 0.0019117222400382161,
"learning_rate": 2.3180592991913745e-06,
"loss": 0.0,
"step": 5410
},
{
"epoch": 79.71,
"grad_norm": 0.0019178889924660325,
"learning_rate": 2.301212938005391e-06,
"loss": 0.0,
"step": 5420
},
{
"epoch": 79.85,
"grad_norm": 0.002448749728500843,
"learning_rate": 2.2843665768194074e-06,
"loss": 0.0,
"step": 5430
},
{
"epoch": 80.0,
"grad_norm": 0.0022572469897568226,
"learning_rate": 2.267520215633423e-06,
"loss": 0.0,
"step": 5440
},
{
"epoch": 80.15,
"grad_norm": 0.001967059215530753,
"learning_rate": 2.2506738544474398e-06,
"loss": 0.0,
"step": 5450
},
{
"epoch": 80.29,
"grad_norm": 0.0022357290145009756,
"learning_rate": 2.2338274932614556e-06,
"loss": 0.0,
"step": 5460
},
{
"epoch": 80.44,
"grad_norm": 0.002172647975385189,
"learning_rate": 2.2169811320754718e-06,
"loss": 0.0,
"step": 5470
},
{
"epoch": 80.59,
"grad_norm": 0.0016459524631500244,
"learning_rate": 2.200134770889488e-06,
"loss": 0.0,
"step": 5480
},
{
"epoch": 80.74,
"grad_norm": 0.00204639439471066,
"learning_rate": 2.183288409703504e-06,
"loss": 0.0,
"step": 5490
},
{
"epoch": 80.88,
"grad_norm": 0.0017498856177553535,
"learning_rate": 2.1664420485175204e-06,
"loss": 0.0,
"step": 5500
},
{
"epoch": 80.88,
"eval_loss": 0.5448271632194519,
"eval_runtime": 430.4314,
"eval_samples_per_second": 1.752,
"eval_steps_per_second": 0.019,
"eval_wer": 39.66403162055336,
"step": 5500
},
{
"epoch": 81.03,
"grad_norm": 0.0018372549675405025,
"learning_rate": 2.1495956873315366e-06,
"loss": 0.0,
"step": 5510
},
{
"epoch": 81.18,
"grad_norm": 0.0016495827585458755,
"learning_rate": 2.132749326145553e-06,
"loss": 0.0,
"step": 5520
},
{
"epoch": 81.32,
"grad_norm": 0.0017999428091570735,
"learning_rate": 2.115902964959569e-06,
"loss": 0.0,
"step": 5530
},
{
"epoch": 81.47,
"grad_norm": 0.0018734214827418327,
"learning_rate": 2.099056603773585e-06,
"loss": 0.0,
"step": 5540
},
{
"epoch": 81.62,
"grad_norm": 0.0016185399144887924,
"learning_rate": 2.082210242587601e-06,
"loss": 0.0,
"step": 5550
},
{
"epoch": 81.76,
"grad_norm": 0.001710501266643405,
"learning_rate": 2.0653638814016176e-06,
"loss": 0.0,
"step": 5560
},
{
"epoch": 81.91,
"grad_norm": 0.002252147998660803,
"learning_rate": 2.0485175202156334e-06,
"loss": 0.0,
"step": 5570
},
{
"epoch": 82.06,
"grad_norm": 0.0018945408519357443,
"learning_rate": 2.0316711590296496e-06,
"loss": 0.0,
"step": 5580
},
{
"epoch": 82.21,
"grad_norm": 0.0016931135905906558,
"learning_rate": 2.014824797843666e-06,
"loss": 0.0,
"step": 5590
},
{
"epoch": 82.35,
"grad_norm": 0.002079603960737586,
"learning_rate": 1.997978436657682e-06,
"loss": 0.0,
"step": 5600
},
{
"epoch": 82.5,
"grad_norm": 0.0014379196800291538,
"learning_rate": 1.981132075471698e-06,
"loss": 0.0,
"step": 5610
},
{
"epoch": 82.65,
"grad_norm": 0.0017873428296297789,
"learning_rate": 1.9642857142857144e-06,
"loss": 0.0,
"step": 5620
},
{
"epoch": 82.79,
"grad_norm": 0.0016607646830379963,
"learning_rate": 1.9474393530997306e-06,
"loss": 0.0,
"step": 5630
},
{
"epoch": 82.94,
"grad_norm": 0.0018578683957457542,
"learning_rate": 1.930592991913747e-06,
"loss": 0.0,
"step": 5640
},
{
"epoch": 83.09,
"grad_norm": 0.001992453821003437,
"learning_rate": 1.913746630727763e-06,
"loss": 0.0,
"step": 5650
},
{
"epoch": 83.24,
"grad_norm": 0.001795714139007032,
"learning_rate": 1.896900269541779e-06,
"loss": 0.0,
"step": 5660
},
{
"epoch": 83.38,
"grad_norm": 0.0017513408092781901,
"learning_rate": 1.8800539083557952e-06,
"loss": 0.0,
"step": 5670
},
{
"epoch": 83.53,
"grad_norm": 0.001984973670914769,
"learning_rate": 1.8632075471698114e-06,
"loss": 0.0,
"step": 5680
},
{
"epoch": 83.68,
"grad_norm": 0.001725591835565865,
"learning_rate": 1.8463611859838276e-06,
"loss": 0.0,
"step": 5690
},
{
"epoch": 83.82,
"grad_norm": 0.001957892207428813,
"learning_rate": 1.8295148247978438e-06,
"loss": 0.0,
"step": 5700
},
{
"epoch": 83.97,
"grad_norm": 0.0014876240165904164,
"learning_rate": 1.8126684636118598e-06,
"loss": 0.0,
"step": 5710
},
{
"epoch": 84.12,
"grad_norm": 0.0017114444635808468,
"learning_rate": 1.7958221024258763e-06,
"loss": 0.0,
"step": 5720
},
{
"epoch": 84.26,
"grad_norm": 0.0018760694656521082,
"learning_rate": 1.7789757412398922e-06,
"loss": 0.0,
"step": 5730
},
{
"epoch": 84.41,
"grad_norm": 0.001803001156076789,
"learning_rate": 1.7621293800539085e-06,
"loss": 0.0,
"step": 5740
},
{
"epoch": 84.56,
"grad_norm": 0.0018211943097412586,
"learning_rate": 1.7452830188679247e-06,
"loss": 0.0,
"step": 5750
},
{
"epoch": 84.71,
"grad_norm": 0.001698785461485386,
"learning_rate": 1.7284366576819409e-06,
"loss": 0.0,
"step": 5760
},
{
"epoch": 84.85,
"grad_norm": 0.00188141327816993,
"learning_rate": 1.711590296495957e-06,
"loss": 0.0,
"step": 5770
},
{
"epoch": 85.0,
"grad_norm": 0.0018102971371263266,
"learning_rate": 1.6947439353099733e-06,
"loss": 0.0,
"step": 5780
},
{
"epoch": 85.15,
"grad_norm": 0.002033288823440671,
"learning_rate": 1.6778975741239895e-06,
"loss": 0.0,
"step": 5790
},
{
"epoch": 85.29,
"grad_norm": 0.0015100777382031083,
"learning_rate": 1.6610512129380055e-06,
"loss": 0.0,
"step": 5800
},
{
"epoch": 85.44,
"grad_norm": 0.0018259905045852065,
"learning_rate": 1.6442048517520217e-06,
"loss": 0.0,
"step": 5810
},
{
"epoch": 85.59,
"grad_norm": 0.0018028096528723836,
"learning_rate": 1.6273584905660379e-06,
"loss": 0.0,
"step": 5820
},
{
"epoch": 85.74,
"grad_norm": 0.0016898865578696132,
"learning_rate": 1.610512129380054e-06,
"loss": 0.0,
"step": 5830
},
{
"epoch": 85.88,
"grad_norm": 0.0020932299084961414,
"learning_rate": 1.59366576819407e-06,
"loss": 0.0,
"step": 5840
},
{
"epoch": 86.03,
"grad_norm": 0.0018589514074847102,
"learning_rate": 1.5768194070080865e-06,
"loss": 0.0,
"step": 5850
},
{
"epoch": 86.18,
"grad_norm": 0.001960631925612688,
"learning_rate": 1.5599730458221027e-06,
"loss": 0.0,
"step": 5860
},
{
"epoch": 86.32,
"grad_norm": 0.0015809714095667005,
"learning_rate": 1.5431266846361187e-06,
"loss": 0.0,
"step": 5870
},
{
"epoch": 86.47,
"grad_norm": 0.0015852567739784718,
"learning_rate": 1.5262803234501349e-06,
"loss": 0.0,
"step": 5880
},
{
"epoch": 86.62,
"grad_norm": 0.001482869847677648,
"learning_rate": 1.509433962264151e-06,
"loss": 0.0,
"step": 5890
},
{
"epoch": 86.76,
"grad_norm": 0.00168868328910321,
"learning_rate": 1.4925876010781673e-06,
"loss": 0.0,
"step": 5900
},
{
"epoch": 86.91,
"grad_norm": 0.0017209333600476384,
"learning_rate": 1.4757412398921833e-06,
"loss": 0.0,
"step": 5910
},
{
"epoch": 87.06,
"grad_norm": 0.0015247148694470525,
"learning_rate": 1.4588948787061997e-06,
"loss": 0.0,
"step": 5920
},
{
"epoch": 87.21,
"grad_norm": 0.0020303872879594564,
"learning_rate": 1.4420485175202157e-06,
"loss": 0.0,
"step": 5930
},
{
"epoch": 87.35,
"grad_norm": 0.001518064527772367,
"learning_rate": 1.425202156334232e-06,
"loss": 0.0,
"step": 5940
},
{
"epoch": 87.5,
"grad_norm": 0.0015475251711905003,
"learning_rate": 1.4083557951482481e-06,
"loss": 0.0,
"step": 5950
},
{
"epoch": 87.65,
"grad_norm": 0.0014926763251423836,
"learning_rate": 1.3915094339622643e-06,
"loss": 0.0,
"step": 5960
},
{
"epoch": 87.79,
"grad_norm": 0.0014638776192441583,
"learning_rate": 1.3746630727762805e-06,
"loss": 0.0,
"step": 5970
},
{
"epoch": 87.94,
"grad_norm": 0.0015593844000250101,
"learning_rate": 1.3578167115902965e-06,
"loss": 0.0,
"step": 5980
},
{
"epoch": 88.09,
"grad_norm": 0.001806886401027441,
"learning_rate": 1.340970350404313e-06,
"loss": 0.0,
"step": 5990
},
{
"epoch": 88.24,
"grad_norm": 0.001656343461945653,
"learning_rate": 1.324123989218329e-06,
"loss": 0.0,
"step": 6000
},
{
"epoch": 88.24,
"eval_loss": 0.5489587783813477,
"eval_runtime": 428.0298,
"eval_samples_per_second": 1.762,
"eval_steps_per_second": 0.019,
"eval_wer": 39.565217391304344,
"step": 6000
},
{
"epoch": 88.38,
"grad_norm": 0.0019978652708232403,
"learning_rate": 1.3072776280323451e-06,
"loss": 0.0,
"step": 6010
},
{
"epoch": 88.53,
"grad_norm": 0.0016022637719288468,
"learning_rate": 1.2904312668463611e-06,
"loss": 0.0,
"step": 6020
},
{
"epoch": 88.68,
"grad_norm": 0.0016824258491396904,
"learning_rate": 1.2735849056603775e-06,
"loss": 0.0,
"step": 6030
},
{
"epoch": 88.82,
"grad_norm": 0.0017654020339250565,
"learning_rate": 1.2567385444743937e-06,
"loss": 0.0,
"step": 6040
},
{
"epoch": 88.97,
"grad_norm": 0.001638589659705758,
"learning_rate": 1.2398921832884097e-06,
"loss": 0.0,
"step": 6050
},
{
"epoch": 89.12,
"grad_norm": 0.001573295914568007,
"learning_rate": 1.223045822102426e-06,
"loss": 0.0,
"step": 6060
},
{
"epoch": 89.26,
"grad_norm": 0.0014727225061506033,
"learning_rate": 1.2061994609164422e-06,
"loss": 0.0,
"step": 6070
},
{
"epoch": 89.41,
"grad_norm": 0.0016162623651325703,
"learning_rate": 1.1893530997304584e-06,
"loss": 0.0,
"step": 6080
},
{
"epoch": 89.56,
"grad_norm": 0.0016934397863224149,
"learning_rate": 1.1725067385444746e-06,
"loss": 0.0,
"step": 6090
},
{
"epoch": 89.71,
"grad_norm": 0.0017905846470966935,
"learning_rate": 1.1556603773584908e-06,
"loss": 0.0,
"step": 6100
},
{
"epoch": 89.85,
"grad_norm": 0.001646560151129961,
"learning_rate": 1.1388140161725068e-06,
"loss": 0.0,
"step": 6110
},
{
"epoch": 90.0,
"grad_norm": 0.001669664983637631,
"learning_rate": 1.121967654986523e-06,
"loss": 0.0,
"step": 6120
},
{
"epoch": 90.15,
"grad_norm": 0.0015495212282985449,
"learning_rate": 1.1051212938005392e-06,
"loss": 0.0,
"step": 6130
},
{
"epoch": 90.29,
"grad_norm": 0.0015938329743221402,
"learning_rate": 1.0882749326145554e-06,
"loss": 0.0,
"step": 6140
},
{
"epoch": 90.44,
"grad_norm": 0.0019517322070896626,
"learning_rate": 1.0714285714285714e-06,
"loss": 0.0,
"step": 6150
},
{
"epoch": 90.59,
"grad_norm": 0.0015430712373927236,
"learning_rate": 1.0545822102425878e-06,
"loss": 0.0,
"step": 6160
},
{
"epoch": 90.74,
"grad_norm": 0.0016941017238423228,
"learning_rate": 1.037735849056604e-06,
"loss": 0.0,
"step": 6170
},
{
"epoch": 90.88,
"grad_norm": 0.001742173102684319,
"learning_rate": 1.02088948787062e-06,
"loss": 0.0,
"step": 6180
},
{
"epoch": 91.03,
"grad_norm": 0.0018419664120301604,
"learning_rate": 1.0040431266846362e-06,
"loss": 0.0,
"step": 6190
},
{
"epoch": 91.18,
"grad_norm": 0.0017677543219178915,
"learning_rate": 9.871967654986524e-07,
"loss": 0.0,
"step": 6200
},
{
"epoch": 91.32,
"grad_norm": 0.0016228326130658388,
"learning_rate": 9.703504043126686e-07,
"loss": 0.0,
"step": 6210
},
{
"epoch": 91.47,
"grad_norm": 0.0014178988058120012,
"learning_rate": 9.535040431266847e-07,
"loss": 0.0,
"step": 6220
},
{
"epoch": 91.62,
"grad_norm": 0.0019479967886582017,
"learning_rate": 9.366576819407008e-07,
"loss": 0.0,
"step": 6230
},
{
"epoch": 91.76,
"grad_norm": 0.0013647832674905658,
"learning_rate": 9.19811320754717e-07,
"loss": 0.0,
"step": 6240
},
{
"epoch": 91.91,
"grad_norm": 0.001629057340323925,
"learning_rate": 9.029649595687333e-07,
"loss": 0.0,
"step": 6250
},
{
"epoch": 92.06,
"grad_norm": 0.0017082348931580782,
"learning_rate": 8.861185983827494e-07,
"loss": 0.0,
"step": 6260
},
{
"epoch": 92.21,
"grad_norm": 0.0017187732737511396,
"learning_rate": 8.692722371967656e-07,
"loss": 0.0,
"step": 6270
},
{
"epoch": 92.35,
"grad_norm": 0.0015693089226260781,
"learning_rate": 8.524258760107817e-07,
"loss": 0.0,
"step": 6280
},
{
"epoch": 92.5,
"grad_norm": 0.0016504173399880528,
"learning_rate": 8.355795148247979e-07,
"loss": 0.0,
"step": 6290
},
{
"epoch": 92.65,
"grad_norm": 0.0014378475025296211,
"learning_rate": 8.18733153638814e-07,
"loss": 0.0,
"step": 6300
},
{
"epoch": 92.79,
"grad_norm": 0.0018091071397066116,
"learning_rate": 8.018867924528302e-07,
"loss": 0.0,
"step": 6310
},
{
"epoch": 92.94,
"grad_norm": 0.0016386961797252297,
"learning_rate": 7.850404312668463e-07,
"loss": 0.0,
"step": 6320
},
{
"epoch": 93.09,
"grad_norm": 0.0017092525959014893,
"learning_rate": 7.681940700808626e-07,
"loss": 0.0,
"step": 6330
},
{
"epoch": 93.24,
"grad_norm": 0.0015886749606579542,
"learning_rate": 7.513477088948788e-07,
"loss": 0.0,
"step": 6340
},
{
"epoch": 93.38,
"grad_norm": 0.0018031731015071273,
"learning_rate": 7.345013477088949e-07,
"loss": 0.0,
"step": 6350
},
{
"epoch": 93.53,
"grad_norm": 0.0015261101070791483,
"learning_rate": 7.176549865229111e-07,
"loss": 0.0,
"step": 6360
},
{
"epoch": 93.68,
"grad_norm": 0.001621345872990787,
"learning_rate": 7.008086253369272e-07,
"loss": 0.0,
"step": 6370
},
{
"epoch": 93.82,
"grad_norm": 0.0018881525611504912,
"learning_rate": 6.839622641509434e-07,
"loss": 0.0,
"step": 6380
},
{
"epoch": 93.97,
"grad_norm": 0.0014266808284446597,
"learning_rate": 6.671159029649596e-07,
"loss": 0.0,
"step": 6390
},
{
"epoch": 94.12,
"grad_norm": 0.0017150483326986432,
"learning_rate": 6.502695417789757e-07,
"loss": 0.0,
"step": 6400
},
{
"epoch": 94.26,
"grad_norm": 0.001493273302912712,
"learning_rate": 6.33423180592992e-07,
"loss": 0.0,
"step": 6410
},
{
"epoch": 94.41,
"grad_norm": 0.0016202511033043265,
"learning_rate": 6.165768194070082e-07,
"loss": 0.0,
"step": 6420
},
{
"epoch": 94.56,
"grad_norm": 0.001579651259817183,
"learning_rate": 5.997304582210243e-07,
"loss": 0.0,
"step": 6430
},
{
"epoch": 94.71,
"grad_norm": 0.0014305037911981344,
"learning_rate": 5.828840970350405e-07,
"loss": 0.0,
"step": 6440
},
{
"epoch": 94.85,
"grad_norm": 0.0015384092694148421,
"learning_rate": 5.660377358490567e-07,
"loss": 0.0,
"step": 6450
},
{
"epoch": 95.0,
"grad_norm": 0.0018785832216963172,
"learning_rate": 5.491913746630729e-07,
"loss": 0.0,
"step": 6460
},
{
"epoch": 95.15,
"grad_norm": 0.001387153286486864,
"learning_rate": 5.32345013477089e-07,
"loss": 0.0,
"step": 6470
},
{
"epoch": 95.29,
"grad_norm": 0.0013170434394851327,
"learning_rate": 5.154986522911052e-07,
"loss": 0.0,
"step": 6480
},
{
"epoch": 95.44,
"grad_norm": 0.001672849990427494,
"learning_rate": 4.986522911051214e-07,
"loss": 0.0,
"step": 6490
},
{
"epoch": 95.59,
"grad_norm": 0.0014924348797649145,
"learning_rate": 4.818059299191375e-07,
"loss": 0.0,
"step": 6500
},
{
"epoch": 95.59,
"eval_loss": 0.5511065125465393,
"eval_runtime": 393.4846,
"eval_samples_per_second": 1.916,
"eval_steps_per_second": 0.02,
"eval_wer": 39.565217391304344,
"step": 6500
}
],
"logging_steps": 10,
"max_steps": 6784,
"num_input_tokens_seen": 0,
"num_train_epochs": 100,
"save_steps": 500,
"total_flos": 1.87169276012544e+20,
"train_batch_size": 100,
"trial_name": null,
"trial_params": null
}