diff --git "a/wandb/run-20220731_183904-2sjxhpmt/files/wandb-summary.json" "b/wandb/run-20220731_183904-2sjxhpmt/files/wandb-summary.json" --- "a/wandb/run-20220731_183904-2sjxhpmt/files/wandb-summary.json" +++ "b/wandb/run-20220731_183904-2sjxhpmt/files/wandb-summary.json" @@ -1 +1 @@ -{"train/grad_norm": 3.03125, "layer_grad_norm/": {"lm_head": {"bias": 0.015380859375, "kernel": 2.875}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 0.022216796875, "scale": 0.028564453125}, "layers": {"0": {"attention": {"k_proj": {"bias": 2.47955322265625e-05, "kernel": 0.00738525390625}, "out_proj": {"bias": 0.00830078125, "kernel": 0.05810546875}, "q_proj": {"bias": 0.00064849853515625, "kernel": 0.006683349609375}, "v_proj": {"bias": 0.010986328125, "kernel": 0.054443359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0130615234375, "kernel": 0.1513671875}, "output_dense": {"bias": 0.002777099609375, "kernel": 0.1171875}}, "final_layer_norm": {"bias": 0.054443359375, "scale": 0.1201171875}, "layer_norm": {"bias": 0.02685546875, "scale": 0.026611328125}}, "1": {"attention": {"k_proj": {"bias": 1.811981201171875e-05, "kernel": 0.0093994140625}, "out_proj": {"bias": 0.002960205078125, "kernel": 0.0269775390625}, "q_proj": {"bias": 0.001007080078125, "kernel": 0.0103759765625}, "v_proj": {"bias": 0.00579833984375, "kernel": 0.0206298828125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00531005859375, "kernel": 0.0791015625}, "output_dense": {"bias": 0.0028076171875, "kernel": 0.052490234375}}, "final_layer_norm": {"bias": 0.01171875, "scale": 0.0115966796875}, "layer_norm": {"bias": 0.0166015625, "scale": 0.0191650390625}}, "10": {"attention": {"k_proj": {"bias": 1.3947486877441406e-05, "kernel": 0.0260009765625}, "out_proj": {"bias": 0.0023193359375, "kernel": 0.03466796875}, "q_proj": {"bias": 0.00225830078125, "kernel": 0.02587890625}, "v_proj": {"bias": 0.004608154296875, "kernel": 0.032470703125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00408935546875, "kernel": 0.054443359375}, "output_dense": {"bias": 0.0022430419921875, "kernel": 0.041015625}}, "final_layer_norm": {"bias": 0.0086669921875, "scale": 0.0091552734375}, "layer_norm": {"bias": 0.014404296875, "scale": 0.017578125}}, "11": {"attention": {"k_proj": {"bias": 2.9921531677246094e-05, "kernel": 0.030517578125}, "out_proj": {"bias": 0.002288818359375, "kernel": 0.054931640625}, "q_proj": {"bias": 0.002593994140625, "kernel": 0.030029296875}, "v_proj": {"bias": 0.005126953125, "kernel": 0.04931640625}}, "feed_forward": {"intermediate_dense": {"bias": 0.003875732421875, "kernel": 0.053955078125}, "output_dense": {"bias": 0.0022430419921875, "kernel": 0.0400390625}}, "final_layer_norm": {"bias": 0.007781982421875, "scale": 0.00830078125}, "layer_norm": {"bias": 0.0152587890625, "scale": 0.020751953125}}, "12": {"attention": {"k_proj": {"bias": 2.0742416381835938e-05, "kernel": 0.0279541015625}, "out_proj": {"bias": 0.0023040771484375, "kernel": 0.04248046875}, "q_proj": {"bias": 0.0023193359375, "kernel": 0.02734375}, "v_proj": {"bias": 0.00482177734375, "kernel": 0.0390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.004119873046875, "kernel": 0.05419921875}, "output_dense": {"bias": 0.0022430419921875, "kernel": 0.0400390625}}, "final_layer_norm": {"bias": 0.00897216796875, "scale": 0.0101318359375}, "layer_norm": {"bias": 0.0137939453125, "scale": 0.01318359375}}, "13": {"attention": {"k_proj": {"bias": 3.5762786865234375e-05, "kernel": 0.03857421875}, "out_proj": {"bias": 0.0022735595703125, "kernel": 0.0576171875}, "q_proj": {"bias": 0.003143310546875, "kernel": 0.0390625}, "v_proj": {"bias": 0.004913330078125, "kernel": 0.0517578125}}, "feed_forward": {"intermediate_dense": {"bias": 0.004150390625, "kernel": 0.0556640625}, "output_dense": {"bias": 0.002227783203125, "kernel": 0.043701171875}}, "final_layer_norm": {"bias": 0.0087890625, "scale": 0.0087890625}, "layer_norm": {"bias": 0.0142822265625, "scale": 0.017578125}}, "14": {"attention": {"k_proj": {"bias": 2.7418136596679688e-05, "kernel": 0.0230712890625}, "out_proj": {"bias": 0.002288818359375, "kernel": 0.04443359375}, "q_proj": {"bias": 0.001922607421875, "kernel": 0.0233154296875}, "v_proj": {"bias": 0.00433349609375, "kernel": 0.0400390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.004638671875, "kernel": 0.061279296875}, "output_dense": {"bias": 0.002227783203125, "kernel": 0.047119140625}}, "final_layer_norm": {"bias": 0.011474609375, "scale": 0.0130615234375}, "layer_norm": {"bias": 0.010986328125, "scale": 0.00958251953125}}, "15": {"attention": {"k_proj": {"bias": 9.202957153320312e-05, "kernel": 0.04541015625}, "out_proj": {"bias": 0.0022430419921875, "kernel": 0.09326171875}, "q_proj": {"bias": 0.0034027099609375, "kernel": 0.043701171875}, "v_proj": {"bias": 0.00506591796875, "kernel": 0.068359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0045166015625, "kernel": 0.06103515625}, "output_dense": {"bias": 0.002197265625, "kernel": 0.052734375}}, "final_layer_norm": {"bias": 0.0093994140625, "scale": 0.0084228515625}, "layer_norm": {"bias": 0.01397705078125, "scale": 0.015869140625}}, "16": {"attention": {"k_proj": {"bias": 4.57763671875e-05, "kernel": 0.03076171875}, "out_proj": {"bias": 0.0022430419921875, "kernel": 0.05908203125}, "q_proj": {"bias": 0.0023651123046875, "kernel": 0.0296630859375}, "v_proj": {"bias": 0.00445556640625, "kernel": 0.0458984375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0040283203125, "kernel": 0.05712890625}, "output_dense": {"bias": 0.002197265625, "kernel": 0.048828125}}, "final_layer_norm": {"bias": 0.00848388671875, "scale": 0.00714111328125}, "layer_norm": {"bias": 0.0120849609375, "scale": 0.01055908203125}}, "17": {"attention": {"k_proj": {"bias": 1.4781951904296875e-05, "kernel": 0.0247802734375}, "out_proj": {"bias": 0.0022735595703125, "kernel": 0.03466796875}, "q_proj": {"bias": 0.00189208984375, "kernel": 0.0240478515625}, "v_proj": {"bias": 0.00457763671875, "kernel": 0.03466796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.004150390625, "kernel": 0.057861328125}, "output_dense": {"bias": 0.002197265625, "kernel": 0.0478515625}}, "final_layer_norm": {"bias": 0.0089111328125, "scale": 0.007232666015625}, "layer_norm": {"bias": 0.01287841796875, "scale": 0.01129150390625}}, "18": {"attention": {"k_proj": {"bias": 5.817413330078125e-05, "kernel": 0.038330078125}, "out_proj": {"bias": 0.002227783203125, "kernel": 0.064453125}, "q_proj": {"bias": 0.0025787353515625, "kernel": 0.035400390625}, "v_proj": {"bias": 0.004669189453125, "kernel": 0.0537109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.003875732421875, "kernel": 0.05712890625}, "output_dense": {"bias": 0.002166748046875, "kernel": 0.0478515625}}, "final_layer_norm": {"bias": 0.0081787109375, "scale": 0.00860595703125}, "layer_norm": {"bias": 0.012939453125, "scale": 0.01165771484375}}, "19": {"attention": {"k_proj": {"bias": 2.574920654296875e-05, "kernel": 0.02734375}, "out_proj": {"bias": 0.002197265625, "kernel": 0.040771484375}, "q_proj": {"bias": 0.00201416015625, "kernel": 0.0269775390625}, "v_proj": {"bias": 0.00433349609375, "kernel": 0.03662109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0038604736328125, "kernel": 0.059326171875}, "output_dense": {"bias": 0.00213623046875, "kernel": 0.05029296875}}, "final_layer_norm": {"bias": 0.00811767578125, "scale": 0.006500244140625}, "layer_norm": {"bias": 0.010986328125, "scale": 0.0106201171875}}, "2": {"attention": {"k_proj": {"bias": 1.6450881958007812e-05, "kernel": 0.014892578125}, "out_proj": {"bias": 0.0030364990234375, "kernel": 0.03466796875}, "q_proj": {"bias": 0.0017242431640625, "kernel": 0.015625}, "v_proj": {"bias": 0.0064697265625, "kernel": 0.02978515625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00579833984375, "kernel": 0.09033203125}, "output_dense": {"bias": 0.0027923583984375, "kernel": 0.053466796875}}, "final_layer_norm": {"bias": 0.01318359375, "scale": 0.01312255859375}, "layer_norm": {"bias": 0.0184326171875, "scale": 0.0283203125}}, "20": {"attention": {"k_proj": {"bias": 7.331371307373047e-06, "kernel": 0.0155029296875}, "out_proj": {"bias": 0.002227783203125, "kernel": 0.0234375}, "q_proj": {"bias": 0.0011444091796875, "kernel": 0.016357421875}, "v_proj": {"bias": 0.00396728515625, "kernel": 0.02197265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0040283203125, "kernel": 0.0634765625}, "output_dense": {"bias": 0.002166748046875, "kernel": 0.05029296875}}, "final_layer_norm": {"bias": 0.00823974609375, "scale": 0.00799560546875}, "layer_norm": {"bias": 0.009521484375, "scale": 0.0078125}}, "21": {"attention": {"k_proj": {"bias": 1.609325408935547e-05, "kernel": 0.0224609375}, "out_proj": {"bias": 0.002227783203125, "kernel": 0.0361328125}, "q_proj": {"bias": 0.0017242431640625, "kernel": 0.0228271484375}, "v_proj": {"bias": 0.00421142578125, "kernel": 0.03369140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0042724609375, "kernel": 0.068359375}, "output_dense": {"bias": 0.002166748046875, "kernel": 0.0517578125}}, "final_layer_norm": {"bias": 0.00909423828125, "scale": 0.0087890625}, "layer_norm": {"bias": 0.01031494140625, "scale": 0.01019287109375}}, "22": {"attention": {"k_proj": {"bias": 1.6689300537109375e-05, "kernel": 0.0308837890625}, "out_proj": {"bias": 0.00225830078125, "kernel": 0.0380859375}, "q_proj": {"bias": 0.0022125244140625, "kernel": 0.0306396484375}, "v_proj": {"bias": 0.0045166015625, "kernel": 0.03564453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00439453125, "kernel": 0.0703125}, "output_dense": {"bias": 0.002197265625, "kernel": 0.05126953125}}, "final_layer_norm": {"bias": 0.00958251953125, "scale": 0.00872802734375}, "layer_norm": {"bias": 0.012939453125, "scale": 0.010009765625}}, "23": {"attention": {"k_proj": {"bias": 7.05718994140625e-05, "kernel": 0.058349609375}, "out_proj": {"bias": 0.002197265625, "kernel": 0.076171875}, "q_proj": {"bias": 0.0035552978515625, "kernel": 0.055419921875}, "v_proj": {"bias": 0.005340576171875, "kernel": 0.068359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00433349609375, "kernel": 0.0751953125}, "output_dense": {"bias": 0.002105712890625, "kernel": 0.05224609375}}, "final_layer_norm": {"bias": 0.00982666015625, "scale": 0.0107421875}, "layer_norm": {"bias": 0.015869140625, "scale": 0.017578125}}, "24": {"attention": {"k_proj": {"bias": 5.340576171875e-05, "kernel": 0.059326171875}, "out_proj": {"bias": 0.00201416015625, "kernel": 0.072265625}, "q_proj": {"bias": 0.003814697265625, "kernel": 0.060546875}, "v_proj": {"bias": 0.005126953125, "kernel": 0.06787109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.004241943359375, "kernel": 0.0751953125}, "output_dense": {"bias": 0.001922607421875, "kernel": 0.04541015625}}, "final_layer_norm": {"bias": 0.0093994140625, "scale": 0.0087890625}, "layer_norm": {"bias": 0.016357421875, "scale": 0.0177001953125}}, "25": {"attention": {"k_proj": {"bias": 3.5762786865234375e-05, "kernel": 0.04296875}, "out_proj": {"bias": 0.0019378662109375, "kernel": 0.05322265625}, "q_proj": {"bias": 0.0028839111328125, "kernel": 0.042724609375}, "v_proj": {"bias": 0.00457763671875, "kernel": 0.052734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00396728515625, "kernel": 0.072265625}, "output_dense": {"bias": 0.00185394287109375, "kernel": 0.044677734375}}, "final_layer_norm": {"bias": 0.00909423828125, "scale": 0.0081787109375}, "layer_norm": {"bias": 0.0135498046875, "scale": 0.018310546875}}, "26": {"attention": {"k_proj": {"bias": 3.24249267578125e-05, "kernel": 0.045654296875}, "out_proj": {"bias": 0.00185394287109375, "kernel": 0.047607421875}, "q_proj": {"bias": 0.0030364990234375, "kernel": 0.04638671875}, "v_proj": {"bias": 0.004425048828125, "kernel": 0.04736328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0035858154296875, "kernel": 0.0634765625}, "output_dense": {"bias": 0.001800537109375, "kernel": 0.043212890625}}, "final_layer_norm": {"bias": 0.0081787109375, "scale": 0.008544921875}, "layer_norm": {"bias": 0.01312255859375, "scale": 0.01806640625}}, "27": {"attention": {"k_proj": {"bias": 6.580352783203125e-05, "kernel": 0.07421875}, "out_proj": {"bias": 0.00167083740234375, "kernel": 0.0703125}, "q_proj": {"bias": 0.00457763671875, "kernel": 0.072265625}, "v_proj": {"bias": 0.00469970703125, "kernel": 0.0693359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.003692626953125, "kernel": 0.0634765625}, "output_dense": {"bias": 0.0016326904296875, "kernel": 0.04150390625}}, "final_layer_norm": {"bias": 0.0087890625, "scale": 0.0084228515625}, "layer_norm": {"bias": 0.017822265625, "scale": 0.024658203125}}, "28": {"attention": {"k_proj": {"bias": 5.340576171875e-05, "kernel": 0.059814453125}, "out_proj": {"bias": 0.0015106201171875, "kernel": 0.0634765625}, "q_proj": {"bias": 0.0036468505859375, "kernel": 0.06103515625}, "v_proj": {"bias": 0.0042724609375, "kernel": 0.0634765625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0034027099609375, "kernel": 0.060302734375}, "output_dense": {"bias": 0.001495361328125, "kernel": 0.04052734375}}, "final_layer_norm": {"bias": 0.008544921875, "scale": 0.00836181640625}, "layer_norm": {"bias": 0.0147705078125, "scale": 0.0191650390625}}, "29": {"attention": {"k_proj": {"bias": 6.079673767089844e-05, "kernel": 0.05908203125}, "out_proj": {"bias": 0.00140380859375, "kernel": 0.05712890625}, "q_proj": {"bias": 0.0032958984375, "kernel": 0.06005859375}, "v_proj": {"bias": 0.004119873046875, "kernel": 0.059814453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.003509521484375, "kernel": 0.06689453125}, "output_dense": {"bias": 0.0013275146484375, "kernel": 0.0390625}}, "final_layer_norm": {"bias": 0.0084228515625, "scale": 0.00701904296875}, "layer_norm": {"bias": 0.01458740234375, "scale": 0.014892578125}}, "3": {"attention": {"k_proj": {"bias": 4.100799560546875e-05, "kernel": 0.0223388671875}, "out_proj": {"bias": 0.002960205078125, "kernel": 0.06298828125}, "q_proj": {"bias": 0.0022430419921875, "kernel": 0.0216064453125}, "v_proj": {"bias": 0.005828857421875, "kernel": 0.048828125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00567626953125, "kernel": 0.080078125}, "output_dense": {"bias": 0.002716064453125, "kernel": 0.05078125}}, "final_layer_norm": {"bias": 0.01312255859375, "scale": 0.010986328125}, "layer_norm": {"bias": 0.0157470703125, "scale": 0.02392578125}}, "30": {"attention": {"k_proj": {"bias": 4.4345855712890625e-05, "kernel": 0.0498046875}, "out_proj": {"bias": 0.00128173828125, "kernel": 0.05224609375}, "q_proj": {"bias": 0.0028228759765625, "kernel": 0.0517578125}, "v_proj": {"bias": 0.00347900390625, "kernel": 0.054931640625}}, "feed_forward": {"intermediate_dense": {"bias": 0.003204345703125, "kernel": 0.0634765625}, "output_dense": {"bias": 0.0012054443359375, "kernel": 0.034423828125}}, "final_layer_norm": {"bias": 0.0078125, "scale": 0.0064697265625}, "layer_norm": {"bias": 0.0115966796875, "scale": 0.01806640625}}, "31": {"attention": {"k_proj": {"bias": 4.6253204345703125e-05, "kernel": 0.049560546875}, "out_proj": {"bias": 0.00116729736328125, "kernel": 0.0439453125}, "q_proj": {"bias": 0.00311279296875, "kernel": 0.052734375}, "v_proj": {"bias": 0.003021240234375, "kernel": 0.04736328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00286865234375, "kernel": 0.05517578125}, "output_dense": {"bias": 0.0010986328125, "kernel": 0.031494140625}}, "final_layer_norm": {"bias": 0.006683349609375, "scale": 0.0072021484375}, "layer_norm": {"bias": 0.0113525390625, "scale": 0.0213623046875}}, "32": {"attention": {"k_proj": {"bias": 3.8623809814453125e-05, "kernel": 0.04150390625}, "out_proj": {"bias": 0.0010833740234375, "kernel": 0.0380859375}, "q_proj": {"bias": 0.002532958984375, "kernel": 0.04248046875}, "v_proj": {"bias": 0.0029144287109375, "kernel": 0.042236328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.002899169921875, "kernel": 0.055908203125}, "output_dense": {"bias": 0.00099945068359375, "kernel": 0.028564453125}}, "final_layer_norm": {"bias": 0.006927490234375, "scale": 0.00714111328125}, "layer_norm": {"bias": 0.01031494140625, "scale": 0.0130615234375}}, "33": {"attention": {"k_proj": {"bias": 4.291534423828125e-05, "kernel": 0.0498046875}, "out_proj": {"bias": 0.00096893310546875, "kernel": 0.036376953125}, "q_proj": {"bias": 0.0029296875, "kernel": 0.04931640625}, "v_proj": {"bias": 0.0026702880859375, "kernel": 0.040283203125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0025482177734375, "kernel": 0.049072265625}, "output_dense": {"bias": 0.000896453857421875, "kernel": 0.026611328125}}, "final_layer_norm": {"bias": 0.00653076171875, "scale": 0.0054931640625}, "layer_norm": {"bias": 0.01007080078125, "scale": 0.018310546875}}, "34": {"attention": {"k_proj": {"bias": 6.29425048828125e-05, "kernel": 0.05322265625}, "out_proj": {"bias": 0.0008392333984375, "kernel": 0.03515625}, "q_proj": {"bias": 0.003326416015625, "kernel": 0.055419921875}, "v_proj": {"bias": 0.0023193359375, "kernel": 0.04150390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.002197265625, "kernel": 0.04052734375}, "output_dense": {"bias": 0.0007781982421875, "kernel": 0.0234375}}, "final_layer_norm": {"bias": 0.006103515625, "scale": 0.005828857421875}, "layer_norm": {"bias": 0.0107421875, "scale": 0.0125732421875}}, "35": {"attention": {"k_proj": {"bias": 4.100799560546875e-05, "kernel": 0.036865234375}, "out_proj": {"bias": 0.000736236572265625, "kernel": 0.03662109375}, "q_proj": {"bias": 0.0021820068359375, "kernel": 0.0390625}, "v_proj": {"bias": 0.001922607421875, "kernel": 0.037109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0017852783203125, "kernel": 0.0322265625}, "output_dense": {"bias": 0.00069427490234375, "kernel": 0.020751953125}}, "final_layer_norm": {"bias": 0.00457763671875, "scale": 0.0048828125}, "layer_norm": {"bias": 0.009033203125, "scale": 0.0096435546875}}, "36": {"attention": {"k_proj": {"bias": 6.532669067382812e-05, "kernel": 0.02978515625}, "out_proj": {"bias": 0.00067138671875, "kernel": 0.028564453125}, "q_proj": {"bias": 0.00180816650390625, "kernel": 0.0306396484375}, "v_proj": {"bias": 0.0016021728515625, "kernel": 0.0289306640625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0015411376953125, "kernel": 0.028076171875}, "output_dense": {"bias": 0.000640869140625, "kernel": 0.0172119140625}}, "final_layer_norm": {"bias": 0.003753662109375, "scale": 0.0033111572265625}, "layer_norm": {"bias": 0.006988525390625, "scale": 0.00897216796875}}, "37": {"attention": {"k_proj": {"bias": 4.5299530029296875e-05, "kernel": 0.0322265625}, "out_proj": {"bias": 0.00061798095703125, "kernel": 0.025390625}, "q_proj": {"bias": 0.0020904541015625, "kernel": 0.03466796875}, "v_proj": {"bias": 0.00153350830078125, "kernel": 0.0279541015625}}, "feed_forward": {"intermediate_dense": {"bias": 0.001495361328125, "kernel": 0.02783203125}, "output_dense": {"bias": 0.000583648681640625, "kernel": 0.0166015625}}, "final_layer_norm": {"bias": 0.003692626953125, "scale": 0.003143310546875}, "layer_norm": {"bias": 0.007080078125, "scale": 0.007781982421875}}, "38": {"attention": {"k_proj": {"bias": 5.173683166503906e-05, "kernel": 0.03564453125}, "out_proj": {"bias": 0.0005645751953125, "kernel": 0.024169921875}, "q_proj": {"bias": 0.0020751953125, "kernel": 0.03564453125}, "v_proj": {"bias": 0.0013427734375, "kernel": 0.0262451171875}}, "feed_forward": {"intermediate_dense": {"bias": 0.001373291015625, "kernel": 0.026123046875}, "output_dense": {"bias": 0.0005340576171875, "kernel": 0.01611328125}}, "final_layer_norm": {"bias": 0.0034332275390625, "scale": 0.003936767578125}, "layer_norm": {"bias": 0.0064697265625, "scale": 0.00726318359375}}, "39": {"attention": {"k_proj": {"bias": 3.314018249511719e-05, "kernel": 0.033203125}, "out_proj": {"bias": 0.000507354736328125, "kernel": 0.0225830078125}, "q_proj": {"bias": 0.002044677734375, "kernel": 0.037353515625}, "v_proj": {"bias": 0.00128173828125, "kernel": 0.026611328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00127410888671875, "kernel": 0.02490234375}, "output_dense": {"bias": 0.0004749298095703125, "kernel": 0.01531982421875}}, "final_layer_norm": {"bias": 0.0033111572265625, "scale": 0.0036468505859375}, "layer_norm": {"bias": 0.006591796875, "scale": 0.00799560546875}}, "4": {"attention": {"k_proj": {"bias": 5.8650970458984375e-05, "kernel": 0.026611328125}, "out_proj": {"bias": 0.002838134765625, "kernel": 0.0810546875}, "q_proj": {"bias": 0.0025634765625, "kernel": 0.0272216796875}, "v_proj": {"bias": 0.0054931640625, "kernel": 0.061767578125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00531005859375, "kernel": 0.0712890625}, "output_dense": {"bias": 0.002655029296875, "kernel": 0.052734375}}, "final_layer_norm": {"bias": 0.0115966796875, "scale": 0.01043701171875}, "layer_norm": {"bias": 0.01513671875, "scale": 0.0185546875}}, "40": {"attention": {"k_proj": {"bias": 2.5153160095214844e-05, "kernel": 0.0234375}, "out_proj": {"bias": 0.0004730224609375, "kernel": 0.017333984375}, "q_proj": {"bias": 0.0013427734375, "kernel": 0.024658203125}, "v_proj": {"bias": 0.00102996826171875, "kernel": 0.01904296875}}, "feed_forward": {"intermediate_dense": {"bias": 0.00119781494140625, "kernel": 0.0224609375}, "output_dense": {"bias": 0.000446319580078125, "kernel": 0.01324462890625}}, "final_layer_norm": {"bias": 0.0031890869140625, "scale": 0.005340576171875}, "layer_norm": {"bias": 0.004730224609375, "scale": 0.0078125}}, "41": {"attention": {"k_proj": {"bias": 4.076957702636719e-05, "kernel": 0.03515625}, "out_proj": {"bias": 0.000423431396484375, "kernel": 0.02197265625}, "q_proj": {"bias": 0.0017242431640625, "kernel": 0.036376953125}, "v_proj": {"bias": 0.00121307373046875, "kernel": 0.0279541015625}}, "feed_forward": {"intermediate_dense": {"bias": 0.001068115234375, "kernel": 0.02294921875}, "output_dense": {"bias": 0.0003871917724609375, "kernel": 0.01470947265625}}, "final_layer_norm": {"bias": 0.0031280517578125, "scale": 0.005126953125}, "layer_norm": {"bias": 0.006256103515625, "scale": 0.0076904296875}}, "42": {"attention": {"k_proj": {"bias": 1.0192394256591797e-05, "kernel": 0.011474609375}, "out_proj": {"bias": 0.000392913818359375, "kernel": 0.0128173828125}, "q_proj": {"bias": 0.00067138671875, "kernel": 0.01287841796875}, "v_proj": {"bias": 0.0008087158203125, "kernel": 0.0146484375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00106048583984375, "kernel": 0.023681640625}, "output_dense": {"bias": 0.0003452301025390625, "kernel": 0.01416015625}}, "final_layer_norm": {"bias": 0.00341796875, "scale": 0.0064697265625}, "layer_norm": {"bias": 0.002899169921875, "scale": 0.007354736328125}}, "43": {"attention": {"k_proj": {"bias": 5.27501106262207e-06, "kernel": 0.0062255859375}, "out_proj": {"bias": 0.0003528594970703125, "kernel": 0.00946044921875}, "q_proj": {"bias": 0.00042724609375, "kernel": 0.00726318359375}, "v_proj": {"bias": 0.000629425048828125, "kernel": 0.01031494140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00096893310546875, "kernel": 0.023193359375}, "output_dense": {"bias": 0.000308990478515625, "kernel": 0.0137939453125}}, "final_layer_norm": {"bias": 0.00311279296875, "scale": 0.00439453125}, "layer_norm": {"bias": 0.001953125, "scale": 0.003448486328125}}, "44": {"attention": {"k_proj": {"bias": 8.702278137207031e-06, "kernel": 0.00665283203125}, "out_proj": {"bias": 0.0003147125244140625, "kernel": 0.0091552734375}, "q_proj": {"bias": 0.000476837158203125, "kernel": 0.0081787109375}, "v_proj": {"bias": 0.00058746337890625, "kernel": 0.01025390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.000896453857421875, "kernel": 0.02392578125}, "output_dense": {"bias": 0.000274658203125, "kernel": 0.01470947265625}}, "final_layer_norm": {"bias": 0.0032501220703125, "scale": 0.003631591796875}, "layer_norm": {"bias": 0.00213623046875, "scale": 0.0026092529296875}}, "45": {"attention": {"k_proj": {"bias": 7.271766662597656e-06, "kernel": 0.00537109375}, "out_proj": {"bias": 0.000286102294921875, "kernel": 0.0087890625}, "q_proj": {"bias": 0.0003948211669921875, "kernel": 0.00677490234375}, "v_proj": {"bias": 0.00052642822265625, "kernel": 0.0091552734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00080108642578125, "kernel": 0.02099609375}, "output_dense": {"bias": 0.000255584716796875, "kernel": 0.0155029296875}}, "final_layer_norm": {"bias": 0.00286865234375, "scale": 0.0033416748046875}, "layer_norm": {"bias": 0.001922607421875, "scale": 0.00238037109375}}, "46": {"attention": {"k_proj": {"bias": 1.4066696166992188e-05, "kernel": 0.005126953125}, "out_proj": {"bias": 0.00026702880859375, "kernel": 0.0096435546875}, "q_proj": {"bias": 0.000362396240234375, "kernel": 0.00604248046875}, "v_proj": {"bias": 0.0005340576171875, "kernel": 0.00970458984375}}, "feed_forward": {"intermediate_dense": {"bias": 0.000732421875, "kernel": 0.018310546875}, "output_dense": {"bias": 0.00023651123046875, "kernel": 0.0250244140625}}, "final_layer_norm": {"bias": 0.003326416015625, "scale": 0.00408935546875}, "layer_norm": {"bias": 0.0022125244140625, "scale": 0.00274658203125}}, "47": {"attention": {"k_proj": {"bias": 0.000148773193359375, "kernel": 0.00494384765625}, "out_proj": {"bias": 0.000240325927734375, "kernel": 0.02099609375}, "q_proj": {"bias": 0.00021076202392578125, "kernel": 0.0036163330078125}, "v_proj": {"bias": 0.0004291534423828125, "kernel": 0.00830078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00049591064453125, "kernel": 0.011474609375}, "output_dense": {"bias": 0.00022792816162109375, "kernel": 0.05615234375}}, "final_layer_norm": {"bias": 0.004180908203125, "scale": 0.005828857421875}, "layer_norm": {"bias": 0.002593994140625, "scale": 0.0037994384765625}}, "5": {"attention": {"k_proj": {"bias": 1.3470649719238281e-05, "kernel": 0.022705078125}, "out_proj": {"bias": 0.002899169921875, "kernel": 0.03564453125}, "q_proj": {"bias": 0.002044677734375, "kernel": 0.0225830078125}, "v_proj": {"bias": 0.0057373046875, "kernel": 0.033203125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00555419921875, "kernel": 0.068359375}, "output_dense": {"bias": 0.002685546875, "kernel": 0.052001953125}}, "final_layer_norm": {"bias": 0.013427734375, "scale": 0.012451171875}, "layer_norm": {"bias": 0.016357421875, "scale": 0.0191650390625}}, "6": {"attention": {"k_proj": {"bias": 4.6253204345703125e-05, "kernel": 0.0380859375}, "out_proj": {"bias": 0.002716064453125, "kernel": 0.0751953125}, "q_proj": {"bias": 0.003326416015625, "kernel": 0.03515625}, "v_proj": {"bias": 0.0059814453125, "kernel": 0.064453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00537109375, "kernel": 0.0703125}, "output_dense": {"bias": 0.002532958984375, "kernel": 0.05078125}}, "final_layer_norm": {"bias": 0.011474609375, "scale": 0.01104736328125}, "layer_norm": {"bias": 0.0166015625, "scale": 0.016845703125}}, "7": {"attention": {"k_proj": {"bias": 5.4836273193359375e-05, "kernel": 0.0380859375}, "out_proj": {"bias": 0.00262451171875, "kernel": 0.080078125}, "q_proj": {"bias": 0.0031585693359375, "kernel": 0.0361328125}, "v_proj": {"bias": 0.0054931640625, "kernel": 0.0654296875}}, "feed_forward": {"intermediate_dense": {"bias": 0.005035400390625, "kernel": 0.0693359375}, "output_dense": {"bias": 0.002471923828125, "kernel": 0.050537109375}}, "final_layer_norm": {"bias": 0.010986328125, "scale": 0.0087890625}, "layer_norm": {"bias": 0.016357421875, "scale": 0.022216796875}}, "8": {"attention": {"k_proj": {"bias": 4.744529724121094e-05, "kernel": 0.035400390625}, "out_proj": {"bias": 0.0025177001953125, "kernel": 0.064453125}, "q_proj": {"bias": 0.002960205078125, "kernel": 0.03369140625}, "v_proj": {"bias": 0.0052490234375, "kernel": 0.0546875}}, "feed_forward": {"intermediate_dense": {"bias": 0.005035400390625, "kernel": 0.0693359375}, "output_dense": {"bias": 0.002410888671875, "kernel": 0.0498046875}}, "final_layer_norm": {"bias": 0.01165771484375, "scale": 0.0103759765625}, "layer_norm": {"bias": 0.0152587890625, "scale": 0.020751953125}}, "9": {"attention": {"k_proj": {"bias": 7.295608520507812e-05, "kernel": 0.043212890625}, "out_proj": {"bias": 0.0023193359375, "kernel": 0.09375}, "q_proj": {"bias": 0.003143310546875, "kernel": 0.04150390625}, "v_proj": {"bias": 0.0048828125, "kernel": 0.083984375}}, "feed_forward": {"intermediate_dense": {"bias": 0.004119873046875, "kernel": 0.0615234375}, "output_dense": {"bias": 0.00225830078125, "kernel": 0.044677734375}}, "final_layer_norm": {"bias": 0.0089111328125, "scale": 0.00897216796875}, "layer_norm": {"bias": 0.0142822265625, "scale": 0.017822265625}}}, "pos_conv_embed": {"conv": {"bias": 0.00823974609375, "weight_g": 0.0260009765625, "weight_v": 0.08984375}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "1": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "2": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "3": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "4": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "5": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "6": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}}}, "feature_projection": {"layer_norm": {"bias": 0.037109375, "scale": 0.06787109375}, "projection": {"bias": 0.0126953125, "kernel": 0.2734375}}, "masked_spec_embed": 0.0}}, "layer_param_norm/": {"lm_head": {"bias": 0.5428199768066406, "kernel": 39.53935241699219}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 44.34605026245117, "scale": 73.3198471069336}, "layers": {"0": {"attention": {"k_proj": {"bias": 0.8326878547668457, "kernel": 33.13389587402344}, "out_proj": {"bias": 1.9269423484802246, "kernel": 29.364213943481445}, "q_proj": {"bias": 2.1786375045776367, "kernel": 32.84344482421875}, "v_proj": {"bias": 0.5624852776527405, "kernel": 29.58292579650879}}, "feed_forward": {"intermediate_dense": {"bias": 2.675503730773926, "kernel": 116.23121643066406}, "output_dense": {"bias": 1.303141474723816, "kernel": 112.41329956054688}}, "final_layer_norm": {"bias": 1.6423965692520142, "scale": 19.779216766357422}, "layer_norm": {"bias": 2.610494375228882, "scale": 13.602749824523926}}, "1": {"attention": {"k_proj": {"bias": 0.9677044749259949, "kernel": 47.55170440673828}, "out_proj": {"bias": 1.4074771404266357, "kernel": 48.71104431152344}, "q_proj": {"bias": 3.487532615661621, "kernel": 47.43404769897461}, "v_proj": {"bias": 0.40163999795913696, "kernel": 46.684722900390625}}, "feed_forward": {"intermediate_dense": {"bias": 2.579317092895508, "kernel": 109.14014434814453}, "output_dense": {"bias": 0.9453935623168945, "kernel": 95.64762878417969}}, "final_layer_norm": {"bias": 1.4216214418411255, "scale": 19.20606803894043}, "layer_norm": {"bias": 1.7669947147369385, "scale": 15.537803649902344}}, "10": {"attention": {"k_proj": {"bias": 1.0214643478393555, "kernel": 55.45654296875}, "out_proj": {"bias": 1.3683927059173584, "kernel": 57.103294372558594}, "q_proj": {"bias": 2.743344306945801, "kernel": 55.54225158691406}, "v_proj": {"bias": 0.43790698051452637, "kernel": 56.592411041259766}}, "feed_forward": {"intermediate_dense": {"bias": 2.456008195877075, "kernel": 112.09131622314453}, "output_dense": {"bias": 0.7022168636322021, "kernel": 104.06380462646484}}, "final_layer_norm": {"bias": 2.2165937423706055, "scale": 21.315162658691406}, "layer_norm": {"bias": 1.719329595565796, "scale": 19.529136657714844}}, "11": {"attention": {"k_proj": {"bias": 1.132107138633728, "kernel": 56.78343963623047}, "out_proj": {"bias": 1.2154752016067505, "kernel": 58.082176208496094}, "q_proj": {"bias": 2.707162618637085, "kernel": 56.87295150756836}, "v_proj": {"bias": 0.5470556020736694, "kernel": 57.515804290771484}}, "feed_forward": {"intermediate_dense": {"bias": 2.4912850856781006, "kernel": 113.05030822753906}, "output_dense": {"bias": 0.6732709407806396, "kernel": 105.67353820800781}}, "final_layer_norm": {"bias": 2.2047781944274902, "scale": 21.52802848815918}, "layer_norm": {"bias": 1.7760242223739624, "scale": 20.865413665771484}}, "12": {"attention": {"k_proj": {"bias": 1.0565543174743652, "kernel": 57.1015510559082}, "out_proj": {"bias": 1.1852445602416992, "kernel": 57.51365280151367}, "q_proj": {"bias": 2.5627965927124023, "kernel": 57.20713806152344}, "v_proj": {"bias": 0.5132226943969727, "kernel": 57.104278564453125}}, "feed_forward": {"intermediate_dense": {"bias": 2.5303268432617188, "kernel": 114.10084533691406}, "output_dense": {"bias": 0.6569243669509888, "kernel": 107.61695861816406}}, "final_layer_norm": {"bias": 2.1192846298217773, "scale": 21.54072380065918}, "layer_norm": {"bias": 1.8206603527069092, "scale": 20.66368865966797}}, "13": {"attention": {"k_proj": {"bias": 1.109609842300415, "kernel": 59.139137268066406}, "out_proj": {"bias": 1.204558253288269, "kernel": 57.836463928222656}, "q_proj": {"bias": 2.508016586303711, "kernel": 59.25297546386719}, "v_proj": {"bias": 0.49843183159828186, "kernel": 57.290313720703125}}, "feed_forward": {"intermediate_dense": {"bias": 2.553337574005127, "kernel": 115.32876586914062}, "output_dense": {"bias": 0.6907990574836731, "kernel": 108.72101593017578}}, "final_layer_norm": {"bias": 2.0585994720458984, "scale": 21.673809051513672}, "layer_norm": {"bias": 1.9956371784210205, "scale": 21.9807186126709}}, "14": {"attention": {"k_proj": {"bias": 1.0712106227874756, "kernel": 58.71227264404297}, "out_proj": {"bias": 1.3670698404312134, "kernel": 55.26051330566406}, "q_proj": {"bias": 2.6314339637756348, "kernel": 58.89701843261719}, "v_proj": {"bias": 0.46164584159851074, "kernel": 54.09172821044922}}, "feed_forward": {"intermediate_dense": {"bias": 2.5965352058410645, "kernel": 116.16297912597656}, "output_dense": {"bias": 0.746627926826477, "kernel": 110.53941345214844}}, "final_layer_norm": {"bias": 2.130291223526001, "scale": 21.923263549804688}, "layer_norm": {"bias": 2.070974826812744, "scale": 21.296051025390625}}, "15": {"attention": {"k_proj": {"bias": 1.1247915029525757, "kernel": 59.72203826904297}, "out_proj": {"bias": 1.4958913326263428, "kernel": 58.015098571777344}, "q_proj": {"bias": 2.660292625427246, "kernel": 60.07173156738281}, "v_proj": {"bias": 0.598690390586853, "kernel": 57.066253662109375}}, "feed_forward": {"intermediate_dense": {"bias": 2.634087085723877, "kernel": 116.24002075195312}, "output_dense": {"bias": 0.9244471192359924, "kernel": 112.09469604492188}}, "final_layer_norm": {"bias": 2.292771339416504, "scale": 21.723724365234375}, "layer_norm": {"bias": 2.3929615020751953, "scale": 23.508939743041992}}, "16": {"attention": {"k_proj": {"bias": 1.0540928840637207, "kernel": 59.11586380004883}, "out_proj": {"bias": 1.4032483100891113, "kernel": 56.60514831542969}, "q_proj": {"bias": 2.7260608673095703, "kernel": 59.254310607910156}, "v_proj": {"bias": 0.4804057478904724, "kernel": 55.42608642578125}}, "feed_forward": {"intermediate_dense": {"bias": 2.6082377433776855, "kernel": 116.49190521240234}, "output_dense": {"bias": 0.9517278671264648, "kernel": 112.44628143310547}}, "final_layer_norm": {"bias": 2.338442802429199, "scale": 22.12885284423828}, "layer_norm": {"bias": 2.233637571334839, "scale": 21.41533851623535}}, "17": {"attention": {"k_proj": {"bias": 0.9837132692337036, "kernel": 58.42851638793945}, "out_proj": {"bias": 1.3364170789718628, "kernel": 55.23937225341797}, "q_proj": {"bias": 2.8030295372009277, "kernel": 58.6165771484375}, "v_proj": {"bias": 0.5120510458946228, "kernel": 54.12096405029297}}, "feed_forward": {"intermediate_dense": {"bias": 2.5946855545043945, "kernel": 117.4000244140625}, "output_dense": {"bias": 0.9760332107543945, "kernel": 112.84657287597656}}, "final_layer_norm": {"bias": 2.3310747146606445, "scale": 22.577356338500977}, "layer_norm": {"bias": 2.1474575996398926, "scale": 19.647247314453125}}, "18": {"attention": {"k_proj": {"bias": 1.1168575286865234, "kernel": 60.995887756347656}, "out_proj": {"bias": 1.4672904014587402, "kernel": 57.6978874206543}, "q_proj": {"bias": 2.7197160720825195, "kernel": 61.35331726074219}, "v_proj": {"bias": 0.5715624690055847, "kernel": 56.36853790283203}}, "feed_forward": {"intermediate_dense": {"bias": 2.6232266426086426, "kernel": 117.55758666992188}, "output_dense": {"bias": 1.1174592971801758, "kernel": 114.36212158203125}}, "final_layer_norm": {"bias": 2.5425472259521484, "scale": 22.800260543823242}, "layer_norm": {"bias": 2.3893299102783203, "scale": 22.76236915588379}}, "19": {"attention": {"k_proj": {"bias": 1.0208070278167725, "kernel": 59.06885528564453}, "out_proj": {"bias": 1.4778454303741455, "kernel": 56.21014404296875}, "q_proj": {"bias": 2.8918728828430176, "kernel": 59.33289337158203}, "v_proj": {"bias": 0.5297054052352905, "kernel": 54.7474365234375}}, "feed_forward": {"intermediate_dense": {"bias": 2.669632911682129, "kernel": 117.91584777832031}, "output_dense": {"bias": 1.2106082439422607, "kernel": 115.20211791992188}}, "final_layer_norm": {"bias": 2.573245048522949, "scale": 23.14816665649414}, "layer_norm": {"bias": 2.252002477645874, "scale": 21.209671020507812}}, "2": {"attention": {"k_proj": {"bias": 1.0454318523406982, "kernel": 54.686676025390625}, "out_proj": {"bias": 1.2734014987945557, "kernel": 51.490577697753906}, "q_proj": {"bias": 3.575977325439453, "kernel": 54.497100830078125}, "v_proj": {"bias": 0.3959110677242279, "kernel": 50.49995422363281}}, "feed_forward": {"intermediate_dense": {"bias": 2.594606399536133, "kernel": 113.68925476074219}, "output_dense": {"bias": 0.8178951740264893, "kernel": 99.30268859863281}}, "final_layer_norm": {"bias": 1.4472203254699707, "scale": 21.849597930908203}, "layer_norm": {"bias": 1.584557056427002, "scale": 18.51214599609375}}, "20": {"attention": {"k_proj": {"bias": 0.9627430438995361, "kernel": 58.09296798706055}, "out_proj": {"bias": 1.5276570320129395, "kernel": 54.73149871826172}, "q_proj": {"bias": 2.841951847076416, "kernel": 58.31769561767578}, "v_proj": {"bias": 0.501281201839447, "kernel": 53.01441955566406}}, "feed_forward": {"intermediate_dense": {"bias": 2.6359379291534424, "kernel": 118.42901611328125}, "output_dense": {"bias": 1.3258147239685059, "kernel": 115.80030059814453}}, "final_layer_norm": {"bias": 2.5501017570495605, "scale": 23.993675231933594}, "layer_norm": {"bias": 2.172488212585449, "scale": 20.22334098815918}}, "21": {"attention": {"k_proj": {"bias": 1.0651466846466064, "kernel": 59.5225830078125}, "out_proj": {"bias": 1.5949819087982178, "kernel": 55.423763275146484}, "q_proj": {"bias": 2.7778890132904053, "kernel": 59.900787353515625}, "v_proj": {"bias": 0.6249045133590698, "kernel": 53.91331100463867}}, "feed_forward": {"intermediate_dense": {"bias": 2.67600154876709, "kernel": 118.85958862304688}, "output_dense": {"bias": 1.4759166240692139, "kernel": 116.21441650390625}}, "final_layer_norm": {"bias": 2.586477518081665, "scale": 23.85940933227539}, "layer_norm": {"bias": 2.245365619659424, "scale": 20.780105590820312}}, "22": {"attention": {"k_proj": {"bias": 1.108874797821045, "kernel": 60.636451721191406}, "out_proj": {"bias": 1.5651140213012695, "kernel": 55.89690017700195}, "q_proj": {"bias": 2.7927298545837402, "kernel": 60.867923736572266}, "v_proj": {"bias": 0.6132253408432007, "kernel": 55.12964630126953}}, "feed_forward": {"intermediate_dense": {"bias": 2.5806026458740234, "kernel": 118.8563003540039}, "output_dense": {"bias": 1.5526583194732666, "kernel": 115.60234069824219}}, "final_layer_norm": {"bias": 2.440950632095337, "scale": 23.38101577758789}, "layer_norm": {"bias": 2.2541518211364746, "scale": 19.850492477416992}}, "23": {"attention": {"k_proj": {"bias": 1.1803855895996094, "kernel": 63.56022262573242}, "out_proj": {"bias": 1.8045732975006104, "kernel": 58.528839111328125}, "q_proj": {"bias": 2.795893669128418, "kernel": 63.81403350830078}, "v_proj": {"bias": 0.760036289691925, "kernel": 58.50494384765625}}, "feed_forward": {"intermediate_dense": {"bias": 2.5427958965301514, "kernel": 119.45539855957031}, "output_dense": {"bias": 1.5261703729629517, "kernel": 117.31689453125}}, "final_layer_norm": {"bias": 2.8512122631073, "scale": 23.814239501953125}, "layer_norm": {"bias": 2.7235336303710938, "scale": 23.143463134765625}}, "24": {"attention": {"k_proj": {"bias": 1.2392760515213013, "kernel": 63.35820388793945}, "out_proj": {"bias": 1.8753303289413452, "kernel": 60.957786560058594}, "q_proj": {"bias": 2.969503402709961, "kernel": 63.347145080566406}, "v_proj": {"bias": 0.7420969009399414, "kernel": 60.3218994140625}}, "feed_forward": {"intermediate_dense": {"bias": 2.6453800201416016, "kernel": 118.71673583984375}, "output_dense": {"bias": 1.5667115449905396, "kernel": 119.97584533691406}}, "final_layer_norm": {"bias": 2.882058620452881, "scale": 23.88472557067871}, "layer_norm": {"bias": 2.4996657371520996, "scale": 22.191129684448242}}, "25": {"attention": {"k_proj": {"bias": 1.1723105907440186, "kernel": 62.71778106689453}, "out_proj": {"bias": 1.5817502737045288, "kernel": 57.569122314453125}, "q_proj": {"bias": 2.969485282897949, "kernel": 62.733360290527344}, "v_proj": {"bias": 0.7101114988327026, "kernel": 57.584442138671875}}, "feed_forward": {"intermediate_dense": {"bias": 2.5500712394714355, "kernel": 119.05323791503906}, "output_dense": {"bias": 1.3005712032318115, "kernel": 119.923583984375}}, "final_layer_norm": {"bias": 2.696660280227661, "scale": 24.387502670288086}, "layer_norm": {"bias": 2.4880142211914062, "scale": 20.928028106689453}}, "26": {"attention": {"k_proj": {"bias": 1.127084493637085, "kernel": 61.64911651611328}, "out_proj": {"bias": 1.4543545246124268, "kernel": 57.30519485473633}, "q_proj": {"bias": 3.0726661682128906, "kernel": 61.54557800292969}, "v_proj": {"bias": 0.5901228189468384, "kernel": 57.70402908325195}}, "feed_forward": {"intermediate_dense": {"bias": 2.644829750061035, "kernel": 118.23487091064453}, "output_dense": {"bias": 1.250557541847229, "kernel": 116.10880279541016}}, "final_layer_norm": {"bias": 2.3120343685150146, "scale": 23.21187973022461}, "layer_norm": {"bias": 2.371980905532837, "scale": 20.954845428466797}}, "27": {"attention": {"k_proj": {"bias": 1.317265272140503, "kernel": 64.05865478515625}, "out_proj": {"bias": 1.6727519035339355, "kernel": 60.77666473388672}, "q_proj": {"bias": 2.917194128036499, "kernel": 64.23175048828125}, "v_proj": {"bias": 0.7484986782073975, "kernel": 60.96894073486328}}, "feed_forward": {"intermediate_dense": {"bias": 2.7722606658935547, "kernel": 117.14579772949219}, "output_dense": {"bias": 1.0733555555343628, "kernel": 116.0633773803711}}, "final_layer_norm": {"bias": 2.5812458992004395, "scale": 22.717479705810547}, "layer_norm": {"bias": 2.550534725189209, "scale": 22.94876480102539}}, "28": {"attention": {"k_proj": {"bias": 1.2572720050811768, "kernel": 64.16455078125}, "out_proj": {"bias": 1.6966627836227417, "kernel": 60.11882019042969}, "q_proj": {"bias": 3.1611275672912598, "kernel": 64.0079574584961}, "v_proj": {"bias": 0.664945125579834, "kernel": 60.57808303833008}}, "feed_forward": {"intermediate_dense": {"bias": 2.746213436126709, "kernel": 117.54061889648438}, "output_dense": {"bias": 0.9000833034515381, "kernel": 117.95083618164062}}, "final_layer_norm": {"bias": 2.5712246894836426, "scale": 23.1610050201416}, "layer_norm": {"bias": 2.155653953552246, "scale": 23.39959716796875}}, "29": {"attention": {"k_proj": {"bias": 1.2412364482879639, "kernel": 63.07917022705078}, "out_proj": {"bias": 1.6412043571472168, "kernel": 63.08409881591797}, "q_proj": {"bias": 3.1162118911743164, "kernel": 63.035369873046875}, "v_proj": {"bias": 0.6149334907531738, "kernel": 63.20928955078125}}, "feed_forward": {"intermediate_dense": {"bias": 2.7683238983154297, "kernel": 118.87762451171875}, "output_dense": {"bias": 1.1248301267623901, "kernel": 123.00424194335938}}, "final_layer_norm": {"bias": 2.8067712783813477, "scale": 24.53290367126465}, "layer_norm": {"bias": 2.2788681983947754, "scale": 24.371368408203125}}, "3": {"attention": {"k_proj": {"bias": 1.0682117938995361, "kernel": 58.82686996459961}, "out_proj": {"bias": 1.4460158348083496, "kernel": 53.96826171875}, "q_proj": {"bias": 3.1143360137939453, "kernel": 59.086524963378906}, "v_proj": {"bias": 0.38192105293273926, "kernel": 53.452857971191406}}, "feed_forward": {"intermediate_dense": {"bias": 2.5585412979125977, "kernel": 115.36900329589844}, "output_dense": {"bias": 0.7798702716827393, "kernel": 102.53439331054688}}, "final_layer_norm": {"bias": 1.6873037815093994, "scale": 22.31991195678711}, "layer_norm": {"bias": 1.8086010217666626, "scale": 21.418964385986328}}, "30": {"attention": {"k_proj": {"bias": 1.2559503316879272, "kernel": 63.75602722167969}, "out_proj": {"bias": 1.4681390523910522, "kernel": 59.37921905517578}, "q_proj": {"bias": 3.175278425216675, "kernel": 63.94244384765625}, "v_proj": {"bias": 0.6481107473373413, "kernel": 60.02722930908203}}, "feed_forward": {"intermediate_dense": {"bias": 2.740499973297119, "kernel": 119.35234069824219}, "output_dense": {"bias": 1.0934734344482422, "kernel": 122.9593734741211}}, "final_layer_norm": {"bias": 2.7396388053894043, "scale": 25.626995086669922}, "layer_norm": {"bias": 2.2996180057525635, "scale": 24.201406478881836}}, "31": {"attention": {"k_proj": {"bias": 1.2782238721847534, "kernel": 62.301849365234375}, "out_proj": {"bias": 1.4012553691864014, "kernel": 58.65263366699219}, "q_proj": {"bias": 2.939175605773926, "kernel": 62.559688568115234}, "v_proj": {"bias": 0.6553064584732056, "kernel": 59.218605041503906}}, "feed_forward": {"intermediate_dense": {"bias": 2.817246437072754, "kernel": 117.91593933105469}, "output_dense": {"bias": 1.2606602907180786, "kernel": 119.70211791992188}}, "final_layer_norm": {"bias": 2.5327694416046143, "scale": 25.40445327758789}, "layer_norm": {"bias": 2.305802583694458, "scale": 23.823705673217773}}, "32": {"attention": {"k_proj": {"bias": 1.3094589710235596, "kernel": 63.533992767333984}, "out_proj": {"bias": 1.3854526281356812, "kernel": 58.824134826660156}, "q_proj": {"bias": 3.114076614379883, "kernel": 63.627105712890625}, "v_proj": {"bias": 0.6093175411224365, "kernel": 59.80768585205078}}, "feed_forward": {"intermediate_dense": {"bias": 2.7801480293273926, "kernel": 117.20947265625}, "output_dense": {"bias": 1.324599027633667, "kernel": 119.54325866699219}}, "final_layer_norm": {"bias": 2.5435423851013184, "scale": 25.68441390991211}, "layer_norm": {"bias": 2.4235501289367676, "scale": 23.902536392211914}}, "33": {"attention": {"k_proj": {"bias": 1.3366351127624512, "kernel": 63.296287536621094}, "out_proj": {"bias": 1.430068850517273, "kernel": 58.4971809387207}, "q_proj": {"bias": 3.2261195182800293, "kernel": 63.553199768066406}, "v_proj": {"bias": 0.647566020488739, "kernel": 59.47398376464844}}, "feed_forward": {"intermediate_dense": {"bias": 2.7881383895874023, "kernel": 115.89324188232422}, "output_dense": {"bias": 1.3512067794799805, "kernel": 117.43673706054688}}, "final_layer_norm": {"bias": 2.453516721725464, "scale": 25.343490600585938}, "layer_norm": {"bias": 2.507319450378418, "scale": 23.96185302734375}}, "34": {"attention": {"k_proj": {"bias": 1.2941782474517822, "kernel": 62.228370666503906}, "out_proj": {"bias": 1.6845568418502808, "kernel": 58.70769119262695}, "q_proj": {"bias": 3.1820318698883057, "kernel": 62.562862396240234}, "v_proj": {"bias": 0.5970059633255005, "kernel": 59.69546127319336}}, "feed_forward": {"intermediate_dense": {"bias": 2.897738456726074, "kernel": 114.57638549804688}, "output_dense": {"bias": 1.2965514659881592, "kernel": 115.59634399414062}}, "final_layer_norm": {"bias": 2.396876573562622, "scale": 24.340225219726562}, "layer_norm": {"bias": 2.5630249977111816, "scale": 24.65895652770996}}, "35": {"attention": {"k_proj": {"bias": 1.4279742240905762, "kernel": 66.40350341796875}, "out_proj": {"bias": 1.591472864151001, "kernel": 57.53840637207031}, "q_proj": {"bias": 2.823566436767578, "kernel": 67.20199584960938}, "v_proj": {"bias": 0.5889960527420044, "kernel": 58.97966003417969}}, "feed_forward": {"intermediate_dense": {"bias": 3.0181026458740234, "kernel": 112.65335083007812}, "output_dense": {"bias": 1.1505603790283203, "kernel": 112.97894287109375}}, "final_layer_norm": {"bias": 2.441446304321289, "scale": 23.994781494140625}, "layer_norm": {"bias": 2.542743682861328, "scale": 25.000572204589844}}, "36": {"attention": {"k_proj": {"bias": 1.4029040336608887, "kernel": 63.1389045715332}, "out_proj": {"bias": 1.5916166305541992, "kernel": 57.22056198120117}, "q_proj": {"bias": 2.8235912322998047, "kernel": 63.60721969604492}, "v_proj": {"bias": 0.5017426013946533, "kernel": 59.000877380371094}}, "feed_forward": {"intermediate_dense": {"bias": 2.8778231143951416, "kernel": 110.61883544921875}, "output_dense": {"bias": 1.1144599914550781, "kernel": 111.57343292236328}}, "final_layer_norm": {"bias": 2.349557399749756, "scale": 24.3450927734375}, "layer_norm": {"bias": 2.4408788681030273, "scale": 24.49927520751953}}, "37": {"attention": {"k_proj": {"bias": 1.3937087059020996, "kernel": 60.52811813354492}, "out_proj": {"bias": 1.8286242485046387, "kernel": 56.00824737548828}, "q_proj": {"bias": 2.5970423221588135, "kernel": 60.78678894042969}, "v_proj": {"bias": 0.4639900326728821, "kernel": 57.38051223754883}}, "feed_forward": {"intermediate_dense": {"bias": 2.771303653717041, "kernel": 109.14910888671875}, "output_dense": {"bias": 1.1065362691879272, "kernel": 110.55940246582031}}, "final_layer_norm": {"bias": 2.10972261428833, "scale": 24.696584701538086}, "layer_norm": {"bias": 2.427600383758545, "scale": 24.660919189453125}}, "38": {"attention": {"k_proj": {"bias": 1.3824093341827393, "kernel": 58.65778350830078}, "out_proj": {"bias": 1.515190601348877, "kernel": 54.93376159667969}, "q_proj": {"bias": 2.484334945678711, "kernel": 58.929527282714844}, "v_proj": {"bias": 0.4655402898788452, "kernel": 56.142242431640625}}, "feed_forward": {"intermediate_dense": {"bias": 2.735954999923706, "kernel": 106.66136169433594}, "output_dense": {"bias": 1.0608540773391724, "kernel": 108.26023864746094}}, "final_layer_norm": {"bias": 2.145535945892334, "scale": 25.38375473022461}, "layer_norm": {"bias": 2.560453414916992, "scale": 25.668073654174805}}, "39": {"attention": {"k_proj": {"bias": 1.3524219989776611, "kernel": 57.31468200683594}, "out_proj": {"bias": 1.774656891822815, "kernel": 54.37468719482422}, "q_proj": {"bias": 2.257288932800293, "kernel": 57.76541519165039}, "v_proj": {"bias": 0.4829084873199463, "kernel": 55.80891418457031}}, "feed_forward": {"intermediate_dense": {"bias": 2.6967740058898926, "kernel": 103.42811584472656}, "output_dense": {"bias": 1.1364665031433105, "kernel": 107.01261901855469}}, "final_layer_norm": {"bias": 2.0256223678588867, "scale": 26.028522491455078}, "layer_norm": {"bias": 2.5359292030334473, "scale": 26.22036361694336}}, "4": {"attention": {"k_proj": {"bias": 1.055710792541504, "kernel": 61.09189987182617}, "out_proj": {"bias": 1.6698455810546875, "kernel": 55.757781982421875}, "q_proj": {"bias": 2.8421502113342285, "kernel": 61.419281005859375}, "v_proj": {"bias": 0.4117211103439331, "kernel": 55.36876678466797}}, "feed_forward": {"intermediate_dense": {"bias": 2.505772590637207, "kernel": 114.95476531982422}, "output_dense": {"bias": 0.9291812181472778, "kernel": 104.35224914550781}}, "final_layer_norm": {"bias": 1.8547258377075195, "scale": 21.94887924194336}, "layer_norm": {"bias": 1.9757022857666016, "scale": 22.878047943115234}}, "40": {"attention": {"k_proj": {"bias": 1.3348209857940674, "kernel": 55.37572479248047}, "out_proj": {"bias": 1.7106268405914307, "kernel": 51.78199768066406}, "q_proj": {"bias": 2.1897947788238525, "kernel": 56.08899688720703}, "v_proj": {"bias": 0.51689612865448, "kernel": 52.4681396484375}}, "feed_forward": {"intermediate_dense": {"bias": 2.542057514190674, "kernel": 100.45166015625}, "output_dense": {"bias": 1.158740758895874, "kernel": 103.93336486816406}}, "final_layer_norm": {"bias": 1.9524149894714355, "scale": 25.311817169189453}, "layer_norm": {"bias": 2.4308900833129883, "scale": 25.665002822875977}}, "41": {"attention": {"k_proj": {"bias": 2.1292548179626465, "kernel": 54.88892364501953}, "out_proj": {"bias": 1.4778457880020142, "kernel": 53.82907485961914}, "q_proj": {"bias": 1.8950426578521729, "kernel": 55.096229553222656}, "v_proj": {"bias": 0.5350444316864014, "kernel": 54.30274200439453}}, "feed_forward": {"intermediate_dense": {"bias": 2.6337058544158936, "kernel": 96.57034301757812}, "output_dense": {"bias": 1.195610761642456, "kernel": 101.90911865234375}}, "final_layer_norm": {"bias": 2.2974343299865723, "scale": 28.653759002685547}, "layer_norm": {"bias": 2.4208178520202637, "scale": 27.652393341064453}}, "42": {"attention": {"k_proj": {"bias": 1.4556612968444824, "kernel": 49.96910095214844}, "out_proj": {"bias": 1.5277200937271118, "kernel": 46.47171401977539}, "q_proj": {"bias": 1.772294521331787, "kernel": 50.624691009521484}, "v_proj": {"bias": 0.7890019416809082, "kernel": 45.6768798828125}}, "feed_forward": {"intermediate_dense": {"bias": 2.533141613006592, "kernel": 96.34883880615234}, "output_dense": {"bias": 1.2392287254333496, "kernel": 100.5073013305664}}, "final_layer_norm": {"bias": 2.0547590255737305, "scale": 29.897579193115234}, "layer_norm": {"bias": 1.867573618888855, "scale": 26.137269973754883}}, "43": {"attention": {"k_proj": {"bias": 1.6400136947631836, "kernel": 43.884910583496094}, "out_proj": {"bias": 1.4834134578704834, "kernel": 42.488441467285156}, "q_proj": {"bias": 1.5609513521194458, "kernel": 44.40974426269531}, "v_proj": {"bias": 0.6456707715988159, "kernel": 40.85655975341797}}, "feed_forward": {"intermediate_dense": {"bias": 2.4410853385925293, "kernel": 93.97550964355469}, "output_dense": {"bias": 0.971923828125, "kernel": 97.76197052001953}}, "final_layer_norm": {"bias": 2.2073206901550293, "scale": 32.08831024169922}, "layer_norm": {"bias": 1.9457297325134277, "scale": 24.571487426757812}}, "44": {"attention": {"k_proj": {"bias": 2.736769914627075, "kernel": 42.59820556640625}, "out_proj": {"bias": 1.2384448051452637, "kernel": 45.90458679199219}, "q_proj": {"bias": 1.5544309616088867, "kernel": 43.10401916503906}, "v_proj": {"bias": 0.4339163303375244, "kernel": 45.27587890625}}, "feed_forward": {"intermediate_dense": {"bias": 2.3976941108703613, "kernel": 92.50536346435547}, "output_dense": {"bias": 0.8718729019165039, "kernel": 95.40474700927734}}, "final_layer_norm": {"bias": 2.2857229709625244, "scale": 34.305206298828125}, "layer_norm": {"bias": 1.8322646617889404, "scale": 24.769798278808594}}, "45": {"attention": {"k_proj": {"bias": 2.3045191764831543, "kernel": 41.025367736816406}, "out_proj": {"bias": 1.1059207916259766, "kernel": 49.433837890625}, "q_proj": {"bias": 1.5799487829208374, "kernel": 41.23019790649414}, "v_proj": {"bias": 0.4706912934780121, "kernel": 49.692665100097656}}, "feed_forward": {"intermediate_dense": {"bias": 2.3830413818359375, "kernel": 89.87250518798828}, "output_dense": {"bias": 0.9723072052001953, "kernel": 91.52435302734375}}, "final_layer_norm": {"bias": 1.8866634368896484, "scale": 33.39029312133789}, "layer_norm": {"bias": 1.680398941040039, "scale": 23.549400329589844}}, "46": {"attention": {"k_proj": {"bias": 1.8261332511901855, "kernel": 40.715065002441406}, "out_proj": {"bias": 0.8854400515556335, "kernel": 51.898719787597656}, "q_proj": {"bias": 1.6973485946655273, "kernel": 41.74032211303711}, "v_proj": {"bias": 0.47354304790496826, "kernel": 52.74561309814453}}, "feed_forward": {"intermediate_dense": {"bias": 2.2698826789855957, "kernel": 85.94120025634766}, "output_dense": {"bias": 1.1624335050582886, "kernel": 83.30751037597656}}, "final_layer_norm": {"bias": 1.7028648853302002, "scale": 29.320556640625}, "layer_norm": {"bias": 1.506533145904541, "scale": 22.56432342529297}}, "47": {"attention": {"k_proj": {"bias": 1.2841383218765259, "kernel": 43.89749526977539}, "out_proj": {"bias": 0.7760035991668701, "kernel": 48.40948486328125}, "q_proj": {"bias": 1.8595157861709595, "kernel": 45.508384704589844}, "v_proj": {"bias": 0.7395519018173218, "kernel": 49.07368469238281}}, "feed_forward": {"intermediate_dense": {"bias": 2.2525475025177, "kernel": 82.39459228515625}, "output_dense": {"bias": 0.6880208849906921, "kernel": 77.71249389648438}}, "final_layer_norm": {"bias": 1.4860002994537354, "scale": 23.946605682373047}, "layer_norm": {"bias": 1.5497658252716064, "scale": 20.52420997619629}}, "5": {"attention": {"k_proj": {"bias": 1.0119032859802246, "kernel": 56.31935119628906}, "out_proj": {"bias": 1.6057957410812378, "kernel": 55.774261474609375}, "q_proj": {"bias": 3.061042308807373, "kernel": 56.34562683105469}, "v_proj": {"bias": 0.37706372141838074, "kernel": 55.92870330810547}}, "feed_forward": {"intermediate_dense": {"bias": 2.438316822052002, "kernel": 114.76457214355469}, "output_dense": {"bias": 0.9636346101760864, "kernel": 103.40338134765625}}, "final_layer_norm": {"bias": 1.989949345588684, "scale": 22.142147064208984}, "layer_norm": {"bias": 1.836284875869751, "scale": 20.958904266357422}}, "6": {"attention": {"k_proj": {"bias": 1.1155825853347778, "kernel": 59.055572509765625}, "out_proj": {"bias": 1.6090573072433472, "kernel": 56.853553771972656}, "q_proj": {"bias": 3.048717975616455, "kernel": 59.478973388671875}, "v_proj": {"bias": 0.410391241312027, "kernel": 56.41505813598633}}, "feed_forward": {"intermediate_dense": {"bias": 2.4068522453308105, "kernel": 114.2471694946289}, "output_dense": {"bias": 0.8525809049606323, "kernel": 103.95584106445312}}, "final_layer_norm": {"bias": 2.2757976055145264, "scale": 21.61276626586914}, "layer_norm": {"bias": 1.9055910110473633, "scale": 22.35369110107422}}, "7": {"attention": {"k_proj": {"bias": 0.9782332181930542, "kernel": 57.479583740234375}, "out_proj": {"bias": 1.4624639749526978, "kernel": 56.76895523071289}, "q_proj": {"bias": 2.7724175453186035, "kernel": 57.976966857910156}, "v_proj": {"bias": 0.46141791343688965, "kernel": 55.92349624633789}}, "feed_forward": {"intermediate_dense": {"bias": 2.4027278423309326, "kernel": 113.92277526855469}, "output_dense": {"bias": 0.6965302228927612, "kernel": 103.7275619506836}}, "final_layer_norm": {"bias": 2.259871244430542, "scale": 21.703615188598633}, "layer_norm": {"bias": 1.866142988204956, "scale": 21.641742706298828}}, "8": {"attention": {"k_proj": {"bias": 1.0691603422164917, "kernel": 57.60441970825195}, "out_proj": {"bias": 1.2814300060272217, "kernel": 57.156776428222656}, "q_proj": {"bias": 2.776516914367676, "kernel": 57.734466552734375}, "v_proj": {"bias": 0.434237003326416, "kernel": 56.514854431152344}}, "feed_forward": {"intermediate_dense": {"bias": 2.438565254211426, "kernel": 113.38650512695312}, "output_dense": {"bias": 0.6708989143371582, "kernel": 103.30772399902344}}, "final_layer_norm": {"bias": 2.1530230045318604, "scale": 21.4518985748291}, "layer_norm": {"bias": 1.8316454887390137, "scale": 21.361042022705078}}, "9": {"attention": {"k_proj": {"bias": 1.1614017486572266, "kernel": 59.272682189941406}, "out_proj": {"bias": 1.5716955661773682, "kernel": 59.24287414550781}, "q_proj": {"bias": 2.592710018157959, "kernel": 59.63836669921875}, "v_proj": {"bias": 0.5274032354354858, "kernel": 58.61853790283203}}, "feed_forward": {"intermediate_dense": {"bias": 2.498476982116699, "kernel": 111.90713500976562}, "output_dense": {"bias": 0.7959620952606201, "kernel": 103.15461730957031}}, "final_layer_norm": {"bias": 2.1939809322357178, "scale": 20.66909408569336}, "layer_norm": {"bias": 2.0584192276000977, "scale": 23.691593170166016}}}, "pos_conv_embed": {"conv": {"bias": 6.139623641967773, "weight_g": 9.604846954345703, "weight_v": 124.31410217285156}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 2.0290679931640625, "kernel": 20.55536460876465}, "layer_norm": {"bias": 4.550922393798828, "scale": 16.167570114135742}}, "1": {"conv": {"bias": 1.7790228128433228, "kernel": 51.24136734008789}, "layer_norm": {"bias": 5.962646961212158, "scale": 23.268157958984375}}, "2": {"conv": {"bias": 1.140576720237732, "kernel": 46.50312042236328}, "layer_norm": {"bias": 4.176670551300049, "scale": 20.370853424072266}}, "3": {"conv": {"bias": 0.6725863218307495, "kernel": 44.397525787353516}, "layer_norm": {"bias": 3.888174533843994, "scale": 17.53795051574707}}, "4": {"conv": {"bias": 0.6373162269592285, "kernel": 41.314056396484375}, "layer_norm": {"bias": 2.385471820831299, "scale": 16.34571647644043}}, "5": {"conv": {"bias": 0.5147221684455872, "kernel": 37.479759216308594}, "layer_norm": {"bias": 2.020900011062622, "scale": 17.064470291137695}}, "6": {"conv": {"bias": 0.4947893023490906, "kernel": 40.64780044555664}, "layer_norm": {"bias": 0.5876954793930054, "scale": 19.058603286743164}}}}, "feature_projection": {"layer_norm": {"bias": 6.456684589385986, "scale": 16.53719139099121}, "projection": {"bias": 2.147587776184082, "kernel": 43.111968994140625}}, "masked_spec_embed": 11.914372444152832}}, "train/learning_rate": 0.0003397299733478576, "train/loss": 0.27065393328666687, "train/param_norm": 1370.627197265625, "_runtime": 90372, "_timestamp": 1659383116, "_step": 31900, "eval/loss": 0.5839321613311768, "eval/wer": 0.42758277671575956, "eval/cer": 0.12089792986748961, "eval/step_4k": {"_type": "table-file", "path": "media/table/eval/step_4k_4000_af4cafd73c286841ef2f.table.json", "sha256": "af4cafd73c286841ef2fce257a64583667ab5412cd6837e4b951b2f851540450", "size": 24260, "artifact_path": "wandb-client-artifact://8wsujunwuradmkiy9teyal5atub99m9zxf1gks0x84p3y9bcbrdwjhad89ar5fxnyqtn8bopmk4501qsp1nuyvfeafw7p4spyxkt5zczfl5bxv88dzgxi32ukaj3dp4j:latest/eval/step_4k.table.json", "_latest_artifact_path": "wandb-client-artifact://8wsujunwuradmkiy9teyal5atub99m9zxf1gks0x84p3y9bcbrdwjhad89ar5fxnyqtn8bopmk4501qsp1nuyvfeafw7p4spyxkt5zczfl5bxv88dzgxi32ukaj3dp4j:latest/eval/step_4k.table.json", "ncols": 2, "nrows": 50}, "eval/step_8k": {"_type": "table-file", "path": "media/table/eval/step_8k_8000_c8ddc6e8e3a9e52ebbba.table.json", "sha256": "c8ddc6e8e3a9e52ebbbae9ac6ec8bb7ae6684781548fb4ea5c57a4b03a72d655", "size": 25822, "artifact_path": "wandb-client-artifact://och35iwsgdf5e7r6ebwp8bo4p13eowiyessoabvv4cf6keb0gk4e1577q2io23l2jh4jrzauz2qyodfw4w6u4eyf8llym88t1brov3snl0vrwcrq3dalvmazc40labfc:latest/eval/step_8k.table.json", "_latest_artifact_path": "wandb-client-artifact://och35iwsgdf5e7r6ebwp8bo4p13eowiyessoabvv4cf6keb0gk4e1577q2io23l2jh4jrzauz2qyodfw4w6u4eyf8llym88t1brov3snl0vrwcrq3dalvmazc40labfc:latest/eval/step_8k.table.json", "ncols": 2, "nrows": 50}, "eval/step_12k": {"_type": "table-file", "path": "media/table/eval/step_12k_12000_697630eb77c56222f807.table.json", "sha256": "697630eb77c56222f80728b3497df5ebfe62fb1dd060725ab84ec28fcf8448a3", "size": 25625, "artifact_path": "wandb-client-artifact://ndqwcshgeo7e3e4lndtrefve494sq9zmx9n9lrqdirtgy63uloydfb95oh1cytys0xi7ugpxbq1rub03y9scmcm41ocpbk826sbeejkgr3aubqet78b4jx4d8fb3z14k:latest/eval/step_12k.table.json", "_latest_artifact_path": "wandb-client-artifact://ndqwcshgeo7e3e4lndtrefve494sq9zmx9n9lrqdirtgy63uloydfb95oh1cytys0xi7ugpxbq1rub03y9scmcm41ocpbk826sbeejkgr3aubqet78b4jx4d8fb3z14k:latest/eval/step_12k.table.json", "ncols": 2, "nrows": 50}, "eval/step_16k": {"_type": "table-file", "path": "media/table/eval/step_16k_16000_a8af015baca8352e331a.table.json", "sha256": "a8af015baca8352e331a32965ddaa7fe22e2119a1c1256e539aedfd2cb876b87", "size": 25878, "artifact_path": "wandb-client-artifact://5lnl6ihmavrhu81b4ehn69ru2j2zbl0h27qv20im44v9928o4s69g9nvnb8oni6t0b921jwgbo24pz870kgos572o5h3vkwm71kr1brda3f3ooretb0u164vptzypekc:latest/eval/step_16k.table.json", "_latest_artifact_path": "wandb-client-artifact://5lnl6ihmavrhu81b4ehn69ru2j2zbl0h27qv20im44v9928o4s69g9nvnb8oni6t0b921jwgbo24pz870kgos572o5h3vkwm71kr1brda3f3ooretb0u164vptzypekc:latest/eval/step_16k.table.json", "ncols": 2, "nrows": 50}, "eval/step_20k": {"_type": "table-file", "path": "media/table/eval/step_20k_20000_37ce73b5cf7c7934cf62.table.json", "sha256": "37ce73b5cf7c7934cf62628174d0b0af065414de25e751b9e98983313d2b352e", "size": 25998, "artifact_path": "wandb-client-artifact://dhxussc15vwxa3x78tfa9270hif7fap7tu63y3eysgx549vfxso9k96434vhylx2uacjad6ldxuh7iavo5ogtn6fqtsv0u6d60zpvy5g99bbhwo8mmzkveldckon4ngf:latest/eval/step_20k.table.json", "_latest_artifact_path": "wandb-client-artifact://dhxussc15vwxa3x78tfa9270hif7fap7tu63y3eysgx549vfxso9k96434vhylx2uacjad6ldxuh7iavo5ogtn6fqtsv0u6d60zpvy5g99bbhwo8mmzkveldckon4ngf:latest/eval/step_20k.table.json", "ncols": 2, "nrows": 50}, "eval/step_24k": {"_type": "table-file", "path": "media/table/eval/step_24k_24000_6d0ed7e79108396fc292.table.json", "sha256": "6d0ed7e79108396fc292429957faec4cfead67d1cb5df1dfb6e4064ac1b8efd8", "size": 26106, "artifact_path": "wandb-client-artifact://vc70e9r1bqlbymt7rqzpvpurafx4zsrts0ic4qbqjmtia7a4qefxrwovecenb19dn3y65kbrjbmz63f1534kl7xoa0xst4f09yfus7kr0h0wn4i1hua9wqcf4hmyqgi6:latest/eval/step_24k.table.json", "_latest_artifact_path": "wandb-client-artifact://vc70e9r1bqlbymt7rqzpvpurafx4zsrts0ic4qbqjmtia7a4qefxrwovecenb19dn3y65kbrjbmz63f1534kl7xoa0xst4f09yfus7kr0h0wn4i1hua9wqcf4hmyqgi6:latest/eval/step_24k.table.json", "ncols": 2, "nrows": 50}, "eval/step_28k": {"_type": "table-file", "path": "media/table/eval/step_28k_28000_7186c63d506b9c841f41.table.json", "sha256": "7186c63d506b9c841f410c33dd5d77206b2d413f991c3b48e3b2b1265afbc518", "size": 26279, "artifact_path": "wandb-client-artifact://fdox4htembz1otwyfgt305vpupai0gi1hwwh6nh5zybkkex9v54fvigfcehdhyuj2kffoxiwqocy0n6gno40mw2grt1mc34m2kbnvn5z6b60nr5wh5uh1w8f9wmgicgs:latest/eval/step_28k.table.json", "_latest_artifact_path": "wandb-client-artifact://fdox4htembz1otwyfgt305vpupai0gi1hwwh6nh5zybkkex9v54fvigfcehdhyuj2kffoxiwqocy0n6gno40mw2grt1mc34m2kbnvn5z6b60nr5wh5uh1w8f9wmgicgs:latest/eval/step_28k.table.json", "ncols": 2, "nrows": 50}} \ No newline at end of file +{"train/grad_norm": 10.125, "layer_grad_norm/": {"lm_head": {"bias": 0.049560546875, "kernel": 9.4375}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 0.0673828125, "scale": 0.08154296875}, "layers": {"0": {"attention": {"k_proj": {"bias": 8.58306884765625e-05, "kernel": 0.03759765625}, "out_proj": {"bias": 0.0478515625, "kernel": 0.26171875}, "q_proj": {"bias": 0.0078125, "kernel": 0.056396484375}, "v_proj": {"bias": 0.060546875, "kernel": 0.30078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.06591796875, "kernel": 0.765625}, "output_dense": {"bias": 0.0128173828125, "kernel": 0.578125}}, "final_layer_norm": {"bias": 0.2890625, "scale": 0.5234375}, "layer_norm": {"bias": 0.15625, "scale": 0.2578125}}, "1": {"attention": {"k_proj": {"bias": 7.295608520507812e-05, "kernel": 0.0439453125}, "out_proj": {"bias": 0.013671875, "kernel": 0.14453125}, "q_proj": {"bias": 0.00494384765625, "kernel": 0.0458984375}, "v_proj": {"bias": 0.0263671875, "kernel": 0.1044921875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0244140625, "kernel": 0.361328125}, "output_dense": {"bias": 0.012939453125, "kernel": 0.2421875}}, "final_layer_norm": {"bias": 0.0546875, "scale": 0.0693359375}, "layer_norm": {"bias": 0.078125, "scale": 0.0703125}}, "10": {"attention": {"k_proj": {"bias": 3.7670135498046875e-05, "kernel": 0.0966796875}, "out_proj": {"bias": 0.0096435546875, "kernel": 0.13671875}, "q_proj": {"bias": 0.0089111328125, "kernel": 0.095703125}, "v_proj": {"bias": 0.020263671875, "kernel": 0.1416015625}}, "feed_forward": {"intermediate_dense": {"bias": 0.019287109375, "kernel": 0.2431640625}, "output_dense": {"bias": 0.00897216796875, "kernel": 0.1630859375}}, "final_layer_norm": {"bias": 0.0439453125, "scale": 0.040771484375}, "layer_norm": {"bias": 0.06298828125, "scale": 0.058837890625}}, "11": {"attention": {"k_proj": {"bias": 0.0001010894775390625, "kernel": 0.16015625}, "out_proj": {"bias": 0.00927734375, "kernel": 0.21875}, "q_proj": {"bias": 0.01318359375, "kernel": 0.1396484375}, "v_proj": {"bias": 0.0228271484375, "kernel": 0.236328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0181884765625, "kernel": 0.2412109375}, "output_dense": {"bias": 0.008544921875, "kernel": 0.1513671875}}, "final_layer_norm": {"bias": 0.0458984375, "scale": 0.054443359375}, "layer_norm": {"bias": 0.0703125, "scale": 0.06494140625}}, "12": {"attention": {"k_proj": {"bias": 5.7220458984375e-05, "kernel": 0.1416015625}, "out_proj": {"bias": 0.00872802734375, "kernel": 0.15234375}, "q_proj": {"bias": 0.0120849609375, "kernel": 0.1298828125}, "v_proj": {"bias": 0.020263671875, "kernel": 0.1630859375}}, "feed_forward": {"intermediate_dense": {"bias": 0.016357421875, "kernel": 0.208984375}, "output_dense": {"bias": 0.00830078125, "kernel": 0.140625}}, "final_layer_norm": {"bias": 0.037841796875, "scale": 0.03466796875}, "layer_norm": {"bias": 0.06298828125, "scale": 0.08154296875}}, "13": {"attention": {"k_proj": {"bias": 0.0001239776611328125, "kernel": 0.142578125}, "out_proj": {"bias": 0.008544921875, "kernel": 0.2099609375}, "q_proj": {"bias": 0.01129150390625, "kernel": 0.130859375}, "v_proj": {"bias": 0.021484375, "kernel": 0.2177734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0164794921875, "kernel": 0.2119140625}, "output_dense": {"bias": 0.008056640625, "kernel": 0.150390625}}, "final_layer_norm": {"bias": 0.0400390625, "scale": 0.02880859375}, "layer_norm": {"bias": 0.06298828125, "scale": 0.0771484375}}, "14": {"attention": {"k_proj": {"bias": 0.0001163482666015625, "kernel": 0.08203125}, "out_proj": {"bias": 0.0081787109375, "kernel": 0.1572265625}, "q_proj": {"bias": 0.00628662109375, "kernel": 0.078125}, "v_proj": {"bias": 0.016357421875, "kernel": 0.1552734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.017333984375, "kernel": 0.224609375}, "output_dense": {"bias": 0.00775146484375, "kernel": 0.16015625}}, "final_layer_norm": {"bias": 0.04443359375, "scale": 0.038818359375}, "layer_norm": {"bias": 0.03955078125, "scale": 0.03564453125}}, "15": {"attention": {"k_proj": {"bias": 0.00032806396484375, "kernel": 0.15625}, "out_proj": {"bias": 0.007659912109375, "kernel": 0.29296875}, "q_proj": {"bias": 0.0113525390625, "kernel": 0.142578125}, "v_proj": {"bias": 0.01806640625, "kernel": 0.2265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.014892578125, "kernel": 0.197265625}, "output_dense": {"bias": 0.007476806640625, "kernel": 0.162109375}}, "final_layer_norm": {"bias": 0.031982421875, "scale": 0.0302734375}, "layer_norm": {"bias": 0.04931640625, "scale": 0.05908203125}}, "16": {"attention": {"k_proj": {"bias": 0.0001220703125, "kernel": 0.109375}, "out_proj": {"bias": 0.0076904296875, "kernel": 0.1865234375}, "q_proj": {"bias": 0.0078125, "kernel": 0.09912109375}, "v_proj": {"bias": 0.01611328125, "kernel": 0.154296875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0145263671875, "kernel": 0.193359375}, "output_dense": {"bias": 0.00738525390625, "kernel": 0.1572265625}}, "final_layer_norm": {"bias": 0.0302734375, "scale": 0.02783203125}, "layer_norm": {"bias": 0.044921875, "scale": 0.040771484375}}, "17": {"attention": {"k_proj": {"bias": 4.291534423828125e-05, "kernel": 0.083984375}, "out_proj": {"bias": 0.007720947265625, "kernel": 0.11328125}, "q_proj": {"bias": 0.00665283203125, "kernel": 0.07958984375}, "v_proj": {"bias": 0.01611328125, "kernel": 0.1171875}}, "feed_forward": {"intermediate_dense": {"bias": 0.01458740234375, "kernel": 0.197265625}, "output_dense": {"bias": 0.0074462890625, "kernel": 0.158203125}}, "final_layer_norm": {"bias": 0.033203125, "scale": 0.02880859375}, "layer_norm": {"bias": 0.0478515625, "scale": 0.044189453125}}, "18": {"attention": {"k_proj": {"bias": 0.000141143798828125, "kernel": 0.1142578125}, "out_proj": {"bias": 0.00750732421875, "kernel": 0.2138671875}, "q_proj": {"bias": 0.007537841796875, "kernel": 0.107421875}, "v_proj": {"bias": 0.0164794921875, "kernel": 0.173828125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0140380859375, "kernel": 0.197265625}, "output_dense": {"bias": 0.007171630859375, "kernel": 0.16015625}}, "final_layer_norm": {"bias": 0.031005859375, "scale": 0.0230712890625}, "layer_norm": {"bias": 0.04296875, "scale": 0.060546875}}, "19": {"attention": {"k_proj": {"bias": 6.771087646484375e-05, "kernel": 0.078125}, "out_proj": {"bias": 0.00726318359375, "kernel": 0.138671875}, "q_proj": {"bias": 0.00567626953125, "kernel": 0.078125}, "v_proj": {"bias": 0.01470947265625, "kernel": 0.130859375}}, "feed_forward": {"intermediate_dense": {"bias": 0.012939453125, "kernel": 0.193359375}, "output_dense": {"bias": 0.007080078125, "kernel": 0.1591796875}}, "final_layer_norm": {"bias": 0.026123046875, "scale": 0.0235595703125}, "layer_norm": {"bias": 0.03515625, "scale": 0.031982421875}}, "2": {"attention": {"k_proj": {"bias": 7.05718994140625e-05, "kernel": 0.06005859375}, "out_proj": {"bias": 0.0140380859375, "kernel": 0.171875}, "q_proj": {"bias": 0.006591796875, "kernel": 0.0634765625}, "v_proj": {"bias": 0.030029296875, "kernel": 0.1484375}}, "feed_forward": {"intermediate_dense": {"bias": 0.026611328125, "kernel": 0.408203125}, "output_dense": {"bias": 0.0130615234375, "kernel": 0.234375}}, "final_layer_norm": {"bias": 0.06103515625, "scale": 0.056640625}, "layer_norm": {"bias": 0.07958984375, "scale": 0.0615234375}}, "20": {"attention": {"k_proj": {"bias": 2.8848648071289062e-05, "kernel": 0.06005859375}, "out_proj": {"bias": 0.007415771484375, "kernel": 0.07958984375}, "q_proj": {"bias": 0.004425048828125, "kernel": 0.05517578125}, "v_proj": {"bias": 0.013916015625, "kernel": 0.0791015625}}, "feed_forward": {"intermediate_dense": {"bias": 0.01318359375, "kernel": 0.2021484375}, "output_dense": {"bias": 0.00714111328125, "kernel": 0.1552734375}}, "final_layer_norm": {"bias": 0.0284423828125, "scale": 0.022216796875}, "layer_norm": {"bias": 0.035400390625, "scale": 0.022216796875}}, "21": {"attention": {"k_proj": {"bias": 5.602836608886719e-05, "kernel": 0.07763671875}, "out_proj": {"bias": 0.00726318359375, "kernel": 0.12109375}, "q_proj": {"bias": 0.005615234375, "kernel": 0.07470703125}, "v_proj": {"bias": 0.0140380859375, "kernel": 0.1162109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.012939453125, "kernel": 0.205078125}, "output_dense": {"bias": 0.00714111328125, "kernel": 0.16015625}}, "final_layer_norm": {"bias": 0.02783203125, "scale": 0.020751953125}, "layer_norm": {"bias": 0.03369140625, "scale": 0.037109375}}, "22": {"attention": {"k_proj": {"bias": 4.673004150390625e-05, "kernel": 0.09765625}, "out_proj": {"bias": 0.00738525390625, "kernel": 0.119140625}, "q_proj": {"bias": 0.007659912109375, "kernel": 0.09521484375}, "v_proj": {"bias": 0.0150146484375, "kernel": 0.115234375}}, "feed_forward": {"intermediate_dense": {"bias": 0.01318359375, "kernel": 0.2099609375}, "output_dense": {"bias": 0.0072021484375, "kernel": 0.1533203125}}, "final_layer_norm": {"bias": 0.028564453125, "scale": 0.0262451171875}, "layer_norm": {"bias": 0.040283203125, "scale": 0.042236328125}}, "23": {"attention": {"k_proj": {"bias": 0.0001773834228515625, "kernel": 0.18359375}, "out_proj": {"bias": 0.007080078125, "kernel": 0.2333984375}, "q_proj": {"bias": 0.0111083984375, "kernel": 0.17578125}, "v_proj": {"bias": 0.016357421875, "kernel": 0.203125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0126953125, "kernel": 0.216796875}, "output_dense": {"bias": 0.00689697265625, "kernel": 0.154296875}}, "final_layer_norm": {"bias": 0.029052734375, "scale": 0.03955078125}, "layer_norm": {"bias": 0.048583984375, "scale": 0.0615234375}}, "24": {"attention": {"k_proj": {"bias": 0.0001544952392578125, "kernel": 0.1748046875}, "out_proj": {"bias": 0.006561279296875, "kernel": 0.208984375}, "q_proj": {"bias": 0.010498046875, "kernel": 0.15625}, "v_proj": {"bias": 0.016845703125, "kernel": 0.2041015625}}, "feed_forward": {"intermediate_dense": {"bias": 0.01263427734375, "kernel": 0.21875}, "output_dense": {"bias": 0.0062255859375, "kernel": 0.1357421875}}, "final_layer_norm": {"bias": 0.02880859375, "scale": 0.0283203125}, "layer_norm": {"bias": 0.05078125, "scale": 0.0537109375}}, "25": {"attention": {"k_proj": {"bias": 0.0001201629638671875, "kernel": 0.11328125}, "out_proj": {"bias": 0.00628662109375, "kernel": 0.16015625}, "q_proj": {"bias": 0.007568359375, "kernel": 0.10791015625}, "v_proj": {"bias": 0.01507568359375, "kernel": 0.162109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.012451171875, "kernel": 0.22265625}, "output_dense": {"bias": 0.006011962890625, "kernel": 0.1328125}}, "final_layer_norm": {"bias": 0.029052734375, "scale": 0.032470703125}, "layer_norm": {"bias": 0.04296875, "scale": 0.0400390625}}, "26": {"attention": {"k_proj": {"bias": 8.20159912109375e-05, "kernel": 0.1259765625}, "out_proj": {"bias": 0.006011962890625, "kernel": 0.1376953125}, "q_proj": {"bias": 0.008056640625, "kernel": 0.1171875}, "v_proj": {"bias": 0.01409912109375, "kernel": 0.138671875}}, "feed_forward": {"intermediate_dense": {"bias": 0.01123046875, "kernel": 0.1923828125}, "output_dense": {"bias": 0.00579833984375, "kernel": 0.1220703125}}, "final_layer_norm": {"bias": 0.026123046875, "scale": 0.0302734375}, "layer_norm": {"bias": 0.0390625, "scale": 0.06689453125}}, "27": {"attention": {"k_proj": {"bias": 0.00015544891357421875, "kernel": 0.1357421875}, "out_proj": {"bias": 0.005523681640625, "kernel": 0.2060546875}, "q_proj": {"bias": 0.0086669921875, "kernel": 0.1357421875}, "v_proj": {"bias": 0.01446533203125, "kernel": 0.1953125}}, "feed_forward": {"intermediate_dense": {"bias": 0.01104736328125, "kernel": 0.1875}, "output_dense": {"bias": 0.00537109375, "kernel": 0.126953125}}, "final_layer_norm": {"bias": 0.0252685546875, "scale": 0.0272216796875}, "layer_norm": {"bias": 0.045166015625, "scale": 0.05078125}}, "28": {"attention": {"k_proj": {"bias": 0.00013065338134765625, "kernel": 0.125}, "out_proj": {"bias": 0.00506591796875, "kernel": 0.181640625}, "q_proj": {"bias": 0.0076904296875, "kernel": 0.12890625}, "v_proj": {"bias": 0.01324462890625, "kernel": 0.1875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0101318359375, "kernel": 0.17578125}, "output_dense": {"bias": 0.004974365234375, "kernel": 0.1201171875}}, "final_layer_norm": {"bias": 0.022705078125, "scale": 0.0257568359375}, "layer_norm": {"bias": 0.03955078125, "scale": 0.04052734375}}, "29": {"attention": {"k_proj": {"bias": 0.0001392364501953125, "kernel": 0.1533203125}, "out_proj": {"bias": 0.004791259765625, "kernel": 0.162109375}, "q_proj": {"bias": 0.0087890625, "kernel": 0.15234375}, "v_proj": {"bias": 0.013671875, "kernel": 0.189453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0103759765625, "kernel": 0.19140625}, "output_dense": {"bias": 0.004608154296875, "kernel": 0.1142578125}}, "final_layer_norm": {"bias": 0.0234375, "scale": 0.0234375}, "layer_norm": {"bias": 0.045166015625, "scale": 0.04931640625}}, "3": {"attention": {"k_proj": {"bias": 0.0001316070556640625, "kernel": 0.09765625}, "out_proj": {"bias": 0.0137939453125, "kernel": 0.2734375}, "q_proj": {"bias": 0.01025390625, "kernel": 0.09423828125}, "v_proj": {"bias": 0.0281982421875, "kernel": 0.22265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.02685546875, "kernel": 0.37109375}, "output_dense": {"bias": 0.0126953125, "kernel": 0.2216796875}}, "final_layer_norm": {"bias": 0.061767578125, "scale": 0.052490234375}, "layer_norm": {"bias": 0.0732421875, "scale": 0.09765625}}, "30": {"attention": {"k_proj": {"bias": 0.0001468658447265625, "kernel": 0.11328125}, "out_proj": {"bias": 0.004547119140625, "kernel": 0.1611328125}, "q_proj": {"bias": 0.00689697265625, "kernel": 0.1162109375}, "v_proj": {"bias": 0.0126953125, "kernel": 0.1796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.01025390625, "kernel": 0.1943359375}, "output_dense": {"bias": 0.004302978515625, "kernel": 0.1044921875}}, "final_layer_norm": {"bias": 0.02392578125, "scale": 0.0205078125}, "layer_norm": {"bias": 0.03662109375, "scale": 0.050537109375}}, "31": {"attention": {"k_proj": {"bias": 0.0001010894775390625, "kernel": 0.142578125}, "out_proj": {"bias": 0.004119873046875, "kernel": 0.1357421875}, "q_proj": {"bias": 0.00927734375, "kernel": 0.1455078125}, "v_proj": {"bias": 0.0111083984375, "kernel": 0.1591796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.00927734375, "kernel": 0.171875}, "output_dense": {"bias": 0.003875732421875, "kernel": 0.091796875}}, "final_layer_norm": {"bias": 0.021240234375, "scale": 0.0172119140625}, "layer_norm": {"bias": 0.037353515625, "scale": 0.0654296875}}, "32": {"attention": {"k_proj": {"bias": 0.00015354156494140625, "kernel": 0.123046875}, "out_proj": {"bias": 0.003814697265625, "kernel": 0.1220703125}, "q_proj": {"bias": 0.0072021484375, "kernel": 0.12060546875}, "v_proj": {"bias": 0.0103759765625, "kernel": 0.146484375}}, "feed_forward": {"intermediate_dense": {"bias": 0.010009765625, "kernel": 0.18359375}, "output_dense": {"bias": 0.003387451171875, "kernel": 0.0849609375}}, "final_layer_norm": {"bias": 0.026611328125, "scale": 0.0234375}, "layer_norm": {"bias": 0.03173828125, "scale": 0.033935546875}}, "33": {"attention": {"k_proj": {"bias": 0.00012493133544921875, "kernel": 0.12109375}, "out_proj": {"bias": 0.003326416015625, "kernel": 0.123046875}, "q_proj": {"bias": 0.00750732421875, "kernel": 0.123046875}, "v_proj": {"bias": 0.009521484375, "kernel": 0.138671875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0089111328125, "kernel": 0.162109375}, "output_dense": {"bias": 0.002960205078125, "kernel": 0.078125}}, "final_layer_norm": {"bias": 0.0262451171875, "scale": 0.0228271484375}, "layer_norm": {"bias": 0.03125, "scale": 0.056640625}}, "34": {"attention": {"k_proj": {"bias": 0.00018310546875, "kernel": 0.11328125}, "out_proj": {"bias": 0.00286865234375, "kernel": 0.11474609375}, "q_proj": {"bias": 0.0064697265625, "kernel": 0.1123046875}, "v_proj": {"bias": 0.0084228515625, "kernel": 0.140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00775146484375, "kernel": 0.1357421875}, "output_dense": {"bias": 0.0025787353515625, "kernel": 0.07470703125}}, "final_layer_norm": {"bias": 0.022216796875, "scale": 0.0169677734375}, "layer_norm": {"bias": 0.02880859375, "scale": 0.0400390625}}, "35": {"attention": {"k_proj": {"bias": 0.00012111663818359375, "kernel": 0.1103515625}, "out_proj": {"bias": 0.002410888671875, "kernel": 0.1103515625}, "q_proj": {"bias": 0.0067138671875, "kernel": 0.11474609375}, "v_proj": {"bias": 0.006378173828125, "kernel": 0.11328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00592041015625, "kernel": 0.103515625}, "output_dense": {"bias": 0.0022430419921875, "kernel": 0.0634765625}}, "final_layer_norm": {"bias": 0.01611328125, "scale": 0.0128173828125}, "layer_norm": {"bias": 0.02880859375, "scale": 0.036865234375}}, "36": {"attention": {"k_proj": {"bias": 0.0001087188720703125, "kernel": 0.0927734375}, "out_proj": {"bias": 0.002166748046875, "kernel": 0.0849609375}, "q_proj": {"bias": 0.005706787109375, "kernel": 0.095703125}, "v_proj": {"bias": 0.00543212890625, "kernel": 0.0859375}}, "feed_forward": {"intermediate_dense": {"bias": 0.005279541015625, "kernel": 0.09130859375}, "output_dense": {"bias": 0.00201416015625, "kernel": 0.052734375}}, "final_layer_norm": {"bias": 0.014404296875, "scale": 0.0140380859375}, "layer_norm": {"bias": 0.0224609375, "scale": 0.0260009765625}}, "37": {"attention": {"k_proj": {"bias": 0.00011873245239257812, "kernel": 0.0830078125}, "out_proj": {"bias": 0.00201416015625, "kernel": 0.0771484375}, "q_proj": {"bias": 0.0048828125, "kernel": 0.08349609375}, "v_proj": {"bias": 0.00518798828125, "kernel": 0.08447265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.005157470703125, "kernel": 0.091796875}, "output_dense": {"bias": 0.00183868408203125, "kernel": 0.051025390625}}, "final_layer_norm": {"bias": 0.0140380859375, "scale": 0.0159912109375}, "layer_norm": {"bias": 0.020263671875, "scale": 0.0225830078125}}, "38": {"attention": {"k_proj": {"bias": 0.00016307830810546875, "kernel": 0.08447265625}, "out_proj": {"bias": 0.0018310546875, "kernel": 0.07421875}, "q_proj": {"bias": 0.004638671875, "kernel": 0.08203125}, "v_proj": {"bias": 0.004547119140625, "kernel": 0.08203125}}, "feed_forward": {"intermediate_dense": {"bias": 0.004638671875, "kernel": 0.08544921875}, "output_dense": {"bias": 0.00167083740234375, "kernel": 0.047119140625}}, "final_layer_norm": {"bias": 0.01434326171875, "scale": 0.012451171875}, "layer_norm": {"bias": 0.017822265625, "scale": 0.023193359375}}, "39": {"attention": {"k_proj": {"bias": 0.0001316070556640625, "kernel": 0.07275390625}, "out_proj": {"bias": 0.00160980224609375, "kernel": 0.068359375}, "q_proj": {"bias": 0.00396728515625, "kernel": 0.0732421875}, "v_proj": {"bias": 0.00439453125, "kernel": 0.08251953125}}, "feed_forward": {"intermediate_dense": {"bias": 0.004150390625, "kernel": 0.078125}, "output_dense": {"bias": 0.00146484375, "kernel": 0.04736328125}}, "final_layer_norm": {"bias": 0.01171875, "scale": 0.0101318359375}, "layer_norm": {"bias": 0.0177001953125, "scale": 0.0228271484375}}, "4": {"attention": {"k_proj": {"bias": 0.0002918243408203125, "kernel": 0.126953125}, "out_proj": {"bias": 0.01336669921875, "kernel": 0.33203125}, "q_proj": {"bias": 0.01123046875, "kernel": 0.12109375}, "v_proj": {"bias": 0.02734375, "kernel": 0.28515625}}, "feed_forward": {"intermediate_dense": {"bias": 0.02587890625, "kernel": 0.333984375}, "output_dense": {"bias": 0.0120849609375, "kernel": 0.216796875}}, "final_layer_norm": {"bias": 0.0654296875, "scale": 0.0625}, "layer_norm": {"bias": 0.0712890625, "scale": 0.0771484375}}, "40": {"attention": {"k_proj": {"bias": 7.724761962890625e-05, "kernel": 0.0654296875}, "out_proj": {"bias": 0.0014801025390625, "kernel": 0.05517578125}, "q_proj": {"bias": 0.003753662109375, "kernel": 0.06689453125}, "v_proj": {"bias": 0.003173828125, "kernel": 0.054931640625}}, "feed_forward": {"intermediate_dense": {"bias": 0.003509521484375, "kernel": 0.06396484375}, "output_dense": {"bias": 0.00136566162109375, "kernel": 0.0380859375}}, "final_layer_norm": {"bias": 0.009765625, "scale": 0.0108642578125}, "layer_norm": {"bias": 0.01348876953125, "scale": 0.018798828125}}, "41": {"attention": {"k_proj": {"bias": 0.0001583099365234375, "kernel": 0.056396484375}, "out_proj": {"bias": 0.0013275146484375, "kernel": 0.06884765625}, "q_proj": {"bias": 0.003173828125, "kernel": 0.06298828125}, "v_proj": {"bias": 0.00390625, "kernel": 0.0888671875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0032501220703125, "kernel": 0.0673828125}, "output_dense": {"bias": 0.00119781494140625, "kernel": 0.04443359375}}, "final_layer_norm": {"bias": 0.010498046875, "scale": 0.011474609375}, "layer_norm": {"bias": 0.017578125, "scale": 0.021240234375}}, "42": {"attention": {"k_proj": {"bias": 2.491474151611328e-05, "kernel": 0.026611328125}, "out_proj": {"bias": 0.001220703125, "kernel": 0.04052734375}, "q_proj": {"bias": 0.00147247314453125, "kernel": 0.0274658203125}, "v_proj": {"bias": 0.002410888671875, "kernel": 0.04296875}}, "feed_forward": {"intermediate_dense": {"bias": 0.003204345703125, "kernel": 0.0693359375}, "output_dense": {"bias": 0.0010833740234375, "kernel": 0.04052734375}}, "final_layer_norm": {"bias": 0.0096435546875, "scale": 0.0135498046875}, "layer_norm": {"bias": 0.00738525390625, "scale": 0.01318359375}}, "43": {"attention": {"k_proj": {"bias": 1.823902130126953e-05, "kernel": 0.0184326171875}, "out_proj": {"bias": 0.0011138916015625, "kernel": 0.029296875}, "q_proj": {"bias": 0.0010833740234375, "kernel": 0.0194091796875}, "v_proj": {"bias": 0.0019683837890625, "kernel": 0.031494140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00286865234375, "kernel": 0.0673828125}, "output_dense": {"bias": 0.0009918212890625, "kernel": 0.0400390625}}, "final_layer_norm": {"bias": 0.009033203125, "scale": 0.0157470703125}, "layer_norm": {"bias": 0.005615234375, "scale": 0.008544921875}}, "44": {"attention": {"k_proj": {"bias": 1.9550323486328125e-05, "kernel": 0.0220947265625}, "out_proj": {"bias": 0.0010223388671875, "kernel": 0.030029296875}, "q_proj": {"bias": 0.001312255859375, "kernel": 0.0233154296875}, "v_proj": {"bias": 0.00189208984375, "kernel": 0.032470703125}}, "feed_forward": {"intermediate_dense": {"bias": 0.002716064453125, "kernel": 0.0693359375}, "output_dense": {"bias": 0.000919342041015625, "kernel": 0.0419921875}}, "final_layer_norm": {"bias": 0.00946044921875, "scale": 0.0125732421875}, "layer_norm": {"bias": 0.00616455078125, "scale": 0.0096435546875}}, "45": {"attention": {"k_proj": {"bias": 1.5974044799804688e-05, "kernel": 0.0169677734375}, "out_proj": {"bias": 0.00095367431640625, "kernel": 0.02880859375}, "q_proj": {"bias": 0.0009918212890625, "kernel": 0.0174560546875}, "v_proj": {"bias": 0.00177764892578125, "kernel": 0.0306396484375}}, "feed_forward": {"intermediate_dense": {"bias": 0.002410888671875, "kernel": 0.0615234375}, "output_dense": {"bias": 0.00084686279296875, "kernel": 0.044921875}}, "final_layer_norm": {"bias": 0.008544921875, "scale": 0.0115966796875}, "layer_norm": {"bias": 0.0057373046875, "scale": 0.007293701171875}}, "46": {"attention": {"k_proj": {"bias": 5.650520324707031e-05, "kernel": 0.012939453125}, "out_proj": {"bias": 0.00087738037109375, "kernel": 0.033935546875}, "q_proj": {"bias": 0.000759124755859375, "kernel": 0.01416015625}, "v_proj": {"bias": 0.00164794921875, "kernel": 0.02978515625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00213623046875, "kernel": 0.0517578125}, "output_dense": {"bias": 0.0007781982421875, "kernel": 0.06640625}}, "final_layer_norm": {"bias": 0.01043701171875, "scale": 0.01171875}, "layer_norm": {"bias": 0.0059814453125, "scale": 0.00665283203125}}, "47": {"attention": {"k_proj": {"bias": 0.000461578369140625, "kernel": 0.01953125}, "out_proj": {"bias": 0.000797271728515625, "kernel": 0.0625}, "q_proj": {"bias": 0.000690460205078125, "kernel": 0.01361083984375}, "v_proj": {"bias": 0.00140380859375, "kernel": 0.02734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0014801025390625, "kernel": 0.03369140625}, "output_dense": {"bias": 0.000751495361328125, "kernel": 0.18359375}}, "final_layer_norm": {"bias": 0.010498046875, "scale": 0.0135498046875}, "layer_norm": {"bias": 0.0084228515625, "scale": 0.01251220703125}}, "5": {"attention": {"k_proj": {"bias": 5.125999450683594e-05, "kernel": 0.107421875}, "out_proj": {"bias": 0.0133056640625, "kernel": 0.1669921875}, "q_proj": {"bias": 0.0091552734375, "kernel": 0.10400390625}, "v_proj": {"bias": 0.027099609375, "kernel": 0.162109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.026611328125, "kernel": 0.318359375}, "output_dense": {"bias": 0.011962890625, "kernel": 0.2109375}}, "final_layer_norm": {"bias": 0.064453125, "scale": 0.051025390625}, "layer_norm": {"bias": 0.076171875, "scale": 0.11083984375}}, "6": {"attention": {"k_proj": {"bias": 0.0001659393310546875, "kernel": 0.1552734375}, "out_proj": {"bias": 0.01220703125, "kernel": 0.314453125}, "q_proj": {"bias": 0.013671875, "kernel": 0.146484375}, "v_proj": {"bias": 0.02783203125, "kernel": 0.30078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0257568359375, "kernel": 0.3203125}, "output_dense": {"bias": 0.0111083984375, "kernel": 0.2001953125}}, "final_layer_norm": {"bias": 0.0595703125, "scale": 0.056640625}, "layer_norm": {"bias": 0.0751953125, "scale": 0.126953125}}, "7": {"attention": {"k_proj": {"bias": 0.00016498565673828125, "kernel": 0.1337890625}, "out_proj": {"bias": 0.01177978515625, "kernel": 0.31640625}, "q_proj": {"bias": 0.010986328125, "kernel": 0.125}, "v_proj": {"bias": 0.027587890625, "kernel": 0.275390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.025634765625, "kernel": 0.33203125}, "output_dense": {"bias": 0.01055908203125, "kernel": 0.2060546875}}, "final_layer_norm": {"bias": 0.061767578125, "scale": 0.04736328125}, "layer_norm": {"bias": 0.07568359375, "scale": 0.08984375}}, "8": {"attention": {"k_proj": {"bias": 0.000141143798828125, "kernel": 0.1328125}, "out_proj": {"bias": 0.0108642578125, "kernel": 0.255859375}, "q_proj": {"bias": 0.0115966796875, "kernel": 0.12353515625}, "v_proj": {"bias": 0.0242919921875, "kernel": 0.23828125}}, "feed_forward": {"intermediate_dense": {"bias": 0.023193359375, "kernel": 0.3046875}, "output_dense": {"bias": 0.00994873046875, "kernel": 0.1962890625}}, "final_layer_norm": {"bias": 0.056884765625, "scale": 0.05224609375}, "layer_norm": {"bias": 0.0673828125, "scale": 0.08642578125}}, "9": {"attention": {"k_proj": {"bias": 0.000316619873046875, "kernel": 0.1689453125}, "out_proj": {"bias": 0.0096435546875, "kernel": 0.34765625}, "q_proj": {"bias": 0.0133056640625, "kernel": 0.1640625}, "v_proj": {"bias": 0.021484375, "kernel": 0.341796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.01806640625, "kernel": 0.259765625}, "output_dense": {"bias": 0.00921630859375, "kernel": 0.173828125}}, "final_layer_norm": {"bias": 0.0390625, "scale": 0.032958984375}, "layer_norm": {"bias": 0.0625, "scale": 0.09228515625}}}, "pos_conv_embed": {"conv": {"bias": 0.04150390625, "weight_g": 0.08203125, "weight_v": 0.421875}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "1": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "2": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "3": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "4": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "5": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "6": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}}}, "feature_projection": {"layer_norm": {"bias": 0.21484375, "scale": 0.390625}, "projection": {"bias": 0.06640625, "kernel": 1.4921875}}, "masked_spec_embed": 0.0}}, "layer_param_norm/": {"lm_head": {"bias": 0.5461622476577759, "kernel": 39.839385986328125}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 44.90663146972656, "scale": 73.95352172851562}, "layers": {"0": {"attention": {"k_proj": {"bias": 0.8405735492706299, "kernel": 33.188358306884766}, "out_proj": {"bias": 1.9296445846557617, "kernel": 29.395679473876953}, "q_proj": {"bias": 2.1864075660705566, "kernel": 32.891292572021484}, "v_proj": {"bias": 0.5654440522193909, "kernel": 29.608684539794922}}, "feed_forward": {"intermediate_dense": {"bias": 2.684551477432251, "kernel": 116.43584442138672}, "output_dense": {"bias": 1.3077489137649536, "kernel": 112.6067886352539}}, "final_layer_norm": {"bias": 1.6429417133331299, "scale": 19.779016494750977}, "layer_norm": {"bias": 2.6000375747680664, "scale": 13.585526466369629}}, "1": {"attention": {"k_proj": {"bias": 0.9666125774383545, "kernel": 47.637168884277344}, "out_proj": {"bias": 1.4105943441390991, "kernel": 48.786338806152344}, "q_proj": {"bias": 3.501117706298828, "kernel": 47.51670455932617}, "v_proj": {"bias": 0.40444618463516235, "kernel": 46.75560760498047}}, "feed_forward": {"intermediate_dense": {"bias": 2.5892038345336914, "kernel": 109.27725219726562}, "output_dense": {"bias": 0.9482756853103638, "kernel": 95.75152587890625}}, "final_layer_norm": {"bias": 1.4253612756729126, "scale": 19.206933975219727}, "layer_norm": {"bias": 1.7637306451797485, "scale": 15.505125045776367}}, "10": {"attention": {"k_proj": {"bias": 1.0242327451705933, "kernel": 55.53233337402344}, "out_proj": {"bias": 1.367168664932251, "kernel": 57.173431396484375}, "q_proj": {"bias": 2.752026081085205, "kernel": 55.61391830444336}, "v_proj": {"bias": 0.4381069540977478, "kernel": 56.66081619262695}}, "feed_forward": {"intermediate_dense": {"bias": 2.463383913040161, "kernel": 112.25586700439453}, "output_dense": {"bias": 0.7043792009353638, "kernel": 104.20065307617188}}, "final_layer_norm": {"bias": 2.201185941696167, "scale": 21.335468292236328}, "layer_norm": {"bias": 1.712852954864502, "scale": 19.497753143310547}}, "11": {"attention": {"k_proj": {"bias": 1.1429743766784668, "kernel": 56.88335037231445}, "out_proj": {"bias": 1.2165141105651855, "kernel": 58.17745590209961}, "q_proj": {"bias": 2.7136616706848145, "kernel": 56.972164154052734}, "v_proj": {"bias": 0.5456156730651855, "kernel": 57.60665512084961}}, "feed_forward": {"intermediate_dense": {"bias": 2.5000743865966797, "kernel": 113.21629333496094}, "output_dense": {"bias": 0.6750454902648926, "kernel": 105.81027221679688}}, "final_layer_norm": {"bias": 2.2055253982543945, "scale": 21.54452133178711}, "layer_norm": {"bias": 1.7747254371643066, "scale": 20.849197387695312}}, "12": {"attention": {"k_proj": {"bias": 1.0585795640945435, "kernel": 57.1939697265625}, "out_proj": {"bias": 1.1860017776489258, "kernel": 57.59892272949219}, "q_proj": {"bias": 2.563828468322754, "kernel": 57.300819396972656}, "v_proj": {"bias": 0.5150290727615356, "kernel": 57.185508728027344}}, "feed_forward": {"intermediate_dense": {"bias": 2.539796829223633, "kernel": 114.2740249633789}, "output_dense": {"bias": 0.6583839058876038, "kernel": 107.7599105834961}}, "final_layer_norm": {"bias": 2.1181397438049316, "scale": 21.561912536621094}, "layer_norm": {"bias": 1.8195334672927856, "scale": 20.63427734375}}, "13": {"attention": {"k_proj": {"bias": 1.115009069442749, "kernel": 59.23179626464844}, "out_proj": {"bias": 1.2053178548812866, "kernel": 57.92833709716797}, "q_proj": {"bias": 2.5155858993530273, "kernel": 59.344017028808594}, "v_proj": {"bias": 0.4998053312301636, "kernel": 57.38002014160156}}, "feed_forward": {"intermediate_dense": {"bias": 2.5615246295928955, "kernel": 115.5089111328125}, "output_dense": {"bias": 0.6926842331886292, "kernel": 108.87437438964844}}, "final_layer_norm": {"bias": 2.0524191856384277, "scale": 21.694881439208984}, "layer_norm": {"bias": 1.9935195446014404, "scale": 21.96600341796875}}, "14": {"attention": {"k_proj": {"bias": 1.073731780052185, "kernel": 58.80509948730469}, "out_proj": {"bias": 1.3677575588226318, "kernel": 55.34349822998047}, "q_proj": {"bias": 2.6377768516540527, "kernel": 58.98688507080078}, "v_proj": {"bias": 0.46090763807296753, "kernel": 54.16999816894531}}, "feed_forward": {"intermediate_dense": {"bias": 2.6042683124542236, "kernel": 116.34944152832031}, "output_dense": {"bias": 0.7485758066177368, "kernel": 110.69842529296875}}, "final_layer_norm": {"bias": 2.1236798763275146, "scale": 21.947933197021484}, "layer_norm": {"bias": 2.066810369491577, "scale": 21.274852752685547}}, "15": {"attention": {"k_proj": {"bias": 1.1299641132354736, "kernel": 59.81207275390625}, "out_proj": {"bias": 1.4975049495697021, "kernel": 58.11487579345703}, "q_proj": {"bias": 2.6617274284362793, "kernel": 60.16179656982422}, "v_proj": {"bias": 0.5998342037200928, "kernel": 57.15870666503906}}, "feed_forward": {"intermediate_dense": {"bias": 2.643594264984131, "kernel": 116.42779541015625}, "output_dense": {"bias": 0.9256726503372192, "kernel": 112.26786804199219}}, "final_layer_norm": {"bias": 2.297308921813965, "scale": 21.73810386657715}, "layer_norm": {"bias": 2.3929758071899414, "scale": 23.49250030517578}}, "16": {"attention": {"k_proj": {"bias": 1.0598461627960205, "kernel": 59.214599609375}, "out_proj": {"bias": 1.4043794870376587, "kernel": 56.70579528808594}, "q_proj": {"bias": 2.724116325378418, "kernel": 59.354209899902344}, "v_proj": {"bias": 0.4809194803237915, "kernel": 55.51671600341797}}, "feed_forward": {"intermediate_dense": {"bias": 2.6150827407836914, "kernel": 116.68328857421875}, "output_dense": {"bias": 0.9540103673934937, "kernel": 112.6202392578125}}, "final_layer_norm": {"bias": 2.325410842895508, "scale": 22.15210723876953}, "layer_norm": {"bias": 2.2350716590881348, "scale": 21.39291763305664}}, "17": {"attention": {"k_proj": {"bias": 0.9898021221160889, "kernel": 58.52842712402344}, "out_proj": {"bias": 1.338525414466858, "kernel": 55.334320068359375}, "q_proj": {"bias": 2.813828229904175, "kernel": 58.714149475097656}, "v_proj": {"bias": 0.5121897459030151, "kernel": 54.20729446411133}}, "feed_forward": {"intermediate_dense": {"bias": 2.6031410694122314, "kernel": 117.58949279785156}, "output_dense": {"bias": 0.9783872365951538, "kernel": 113.01774597167969}}, "final_layer_norm": {"bias": 2.3302841186523438, "scale": 22.59705352783203}, "layer_norm": {"bias": 2.144998550415039, "scale": 19.615447998046875}}, "18": {"attention": {"k_proj": {"bias": 1.121544599533081, "kernel": 61.104530334472656}, "out_proj": {"bias": 1.469778060913086, "kernel": 57.81272506713867}, "q_proj": {"bias": 2.7208876609802246, "kernel": 61.46155548095703}, "v_proj": {"bias": 0.5729200839996338, "kernel": 56.47471237182617}}, "feed_forward": {"intermediate_dense": {"bias": 2.6296653747558594, "kernel": 117.752197265625}, "output_dense": {"bias": 1.1205723285675049, "kernel": 114.53910827636719}}, "final_layer_norm": {"bias": 2.530479907989502, "scale": 22.824764251708984}, "layer_norm": {"bias": 2.3926596641540527, "scale": 22.74325942993164}}, "19": {"attention": {"k_proj": {"bias": 1.0290982723236084, "kernel": 59.17970657348633}, "out_proj": {"bias": 1.4808452129364014, "kernel": 56.314361572265625}, "q_proj": {"bias": 2.9019203186035156, "kernel": 59.4432373046875}, "v_proj": {"bias": 0.5309575796127319, "kernel": 54.84058380126953}}, "feed_forward": {"intermediate_dense": {"bias": 2.678683042526245, "kernel": 118.11012268066406}, "output_dense": {"bias": 1.2137682437896729, "kernel": 115.38187408447266}}, "final_layer_norm": {"bias": 2.5742831230163574, "scale": 23.17055892944336}, "layer_norm": {"bias": 2.2464845180511475, "scale": 21.181095123291016}}, "2": {"attention": {"k_proj": {"bias": 1.05588960647583, "kernel": 54.78446578979492}, "out_proj": {"bias": 1.2735207080841064, "kernel": 51.57759475708008}, "q_proj": {"bias": 3.5925040245056152, "kernel": 54.590633392333984}, "v_proj": {"bias": 0.39714381098747253, "kernel": 50.5794677734375}}, "feed_forward": {"intermediate_dense": {"bias": 2.6043550968170166, "kernel": 113.84341430664062}, "output_dense": {"bias": 0.8194270133972168, "kernel": 99.42637634277344}}, "final_layer_norm": {"bias": 1.444088339805603, "scale": 21.86184310913086}, "layer_norm": {"bias": 1.5789973735809326, "scale": 18.492341995239258}}, "20": {"attention": {"k_proj": {"bias": 0.965216875076294, "kernel": 58.20840072631836}, "out_proj": {"bias": 1.5312200784683228, "kernel": 54.83057403564453}, "q_proj": {"bias": 2.8467516899108887, "kernel": 58.42839050292969}, "v_proj": {"bias": 0.502078115940094, "kernel": 53.10401916503906}}, "feed_forward": {"intermediate_dense": {"bias": 2.6429688930511475, "kernel": 118.61834716796875}, "output_dense": {"bias": 1.329444408416748, "kernel": 115.97885131835938}}, "final_layer_norm": {"bias": 2.5429582595825195, "scale": 24.01703453063965}, "layer_norm": {"bias": 2.1694822311401367, "scale": 20.190940856933594}}, "21": {"attention": {"k_proj": {"bias": 1.0720064640045166, "kernel": 59.644710540771484}, "out_proj": {"bias": 1.5987458229064941, "kernel": 55.53227615356445}, "q_proj": {"bias": 2.784161329269409, "kernel": 60.01811218261719}, "v_proj": {"bias": 0.6275163292884827, "kernel": 54.01409912109375}}, "feed_forward": {"intermediate_dense": {"bias": 2.6848812103271484, "kernel": 119.05130004882812}, "output_dense": {"bias": 1.4793968200683594, "kernel": 116.39517974853516}}, "final_layer_norm": {"bias": 2.593503713607788, "scale": 23.880634307861328}, "layer_norm": {"bias": 2.2402238845825195, "scale": 20.75188446044922}}, "22": {"attention": {"k_proj": {"bias": 1.1139270067214966, "kernel": 60.752567291259766}, "out_proj": {"bias": 1.5686640739440918, "kernel": 56.0057373046875}, "q_proj": {"bias": 2.8004791736602783, "kernel": 60.983642578125}, "v_proj": {"bias": 0.6161484718322754, "kernel": 55.23126983642578}}, "feed_forward": {"intermediate_dense": {"bias": 2.587413787841797, "kernel": 119.051025390625}, "output_dense": {"bias": 1.5563151836395264, "kernel": 115.78994750976562}}, "final_layer_norm": {"bias": 2.436882972717285, "scale": 23.40850830078125}, "layer_norm": {"bias": 2.2453255653381348, "scale": 19.8262882232666}}, "23": {"attention": {"k_proj": {"bias": 1.1907281875610352, "kernel": 63.68222427368164}, "out_proj": {"bias": 1.809654712677002, "kernel": 58.65605545043945}, "q_proj": {"bias": 2.800086259841919, "kernel": 63.93647003173828}, "v_proj": {"bias": 0.7624514102935791, "kernel": 58.62438201904297}}, "feed_forward": {"intermediate_dense": {"bias": 2.5535194873809814, "kernel": 119.66822052001953}, "output_dense": {"bias": 1.5295751094818115, "kernel": 117.52777099609375}}, "final_layer_norm": {"bias": 2.8684325218200684, "scale": 23.841670989990234}, "layer_norm": {"bias": 2.7219667434692383, "scale": 23.12306785583496}}, "24": {"attention": {"k_proj": {"bias": 1.242169737815857, "kernel": 63.470767974853516}, "out_proj": {"bias": 1.8797014951705933, "kernel": 61.08135986328125}, "q_proj": {"bias": 2.9713377952575684, "kernel": 63.461265563964844}, "v_proj": {"bias": 0.743897557258606, "kernel": 60.439483642578125}}, "feed_forward": {"intermediate_dense": {"bias": 2.652211904525757, "kernel": 118.92523193359375}, "output_dense": {"bias": 1.5710970163345337, "kernel": 120.18524169921875}}, "final_layer_norm": {"bias": 2.881837844848633, "scale": 23.916698455810547}, "layer_norm": {"bias": 2.498955011367798, "scale": 22.160133361816406}}, "25": {"attention": {"k_proj": {"bias": 1.180206537246704, "kernel": 62.844261169433594}, "out_proj": {"bias": 1.5854909420013428, "kernel": 57.69280242919922}, "q_proj": {"bias": 2.9712910652160645, "kernel": 62.8609504699707}, "v_proj": {"bias": 0.7109427452087402, "kernel": 57.70327377319336}}, "feed_forward": {"intermediate_dense": {"bias": 2.556978225708008, "kernel": 119.26038360595703}, "output_dense": {"bias": 1.3030881881713867, "kernel": 120.13079071044922}}, "final_layer_norm": {"bias": 2.697941780090332, "scale": 24.4180965423584}, "layer_norm": {"bias": 2.488910675048828, "scale": 20.902450561523438}}, "26": {"attention": {"k_proj": {"bias": 1.1336801052093506, "kernel": 61.764259338378906}, "out_proj": {"bias": 1.4570807218551636, "kernel": 57.414268493652344}, "q_proj": {"bias": 3.073143482208252, "kernel": 61.663429260253906}, "v_proj": {"bias": 0.5911090970039368, "kernel": 57.81249237060547}}, "feed_forward": {"intermediate_dense": {"bias": 2.6508240699768066, "kernel": 118.44386291503906}, "output_dense": {"bias": 1.252658724784851, "kernel": 116.30419921875}}, "final_layer_norm": {"bias": 2.3045406341552734, "scale": 23.247268676757812}, "layer_norm": {"bias": 2.3693594932556152, "scale": 20.930042266845703}}, "27": {"attention": {"k_proj": {"bias": 1.3271775245666504, "kernel": 64.17154693603516}, "out_proj": {"bias": 1.6753045320510864, "kernel": 60.90010070800781}, "q_proj": {"bias": 2.921905040740967, "kernel": 64.34422302246094}, "v_proj": {"bias": 0.752945065498352, "kernel": 61.092735290527344}}, "feed_forward": {"intermediate_dense": {"bias": 2.7781379222869873, "kernel": 117.36526489257812}, "output_dense": {"bias": 1.0757441520690918, "kernel": 116.27136993408203}}, "final_layer_norm": {"bias": 2.576234817504883, "scale": 22.753360748291016}, "layer_norm": {"bias": 2.5472233295440674, "scale": 22.923778533935547}}, "28": {"attention": {"k_proj": {"bias": 1.2666256427764893, "kernel": 64.27379608154297}, "out_proj": {"bias": 1.6995872259140015, "kernel": 60.235652923583984}, "q_proj": {"bias": 3.1661453247070312, "kernel": 64.11647033691406}, "v_proj": {"bias": 0.6675081253051758, "kernel": 60.69976043701172}}, "feed_forward": {"intermediate_dense": {"bias": 2.7549471855163574, "kernel": 117.76258850097656}, "output_dense": {"bias": 0.9017891883850098, "kernel": 118.14676666259766}}, "final_layer_norm": {"bias": 2.5793368816375732, "scale": 23.193828582763672}, "layer_norm": {"bias": 2.1511776447296143, "scale": 23.38344955444336}}, "29": {"attention": {"k_proj": {"bias": 1.2495554685592651, "kernel": 63.185630798339844}, "out_proj": {"bias": 1.6429697275161743, "kernel": 63.193111419677734}, "q_proj": {"bias": 3.124325752258301, "kernel": 63.141929626464844}, "v_proj": {"bias": 0.6155401468276978, "kernel": 63.324485778808594}}, "feed_forward": {"intermediate_dense": {"bias": 2.7750461101531982, "kernel": 119.10281372070312}, "output_dense": {"bias": 1.1279209852218628, "kernel": 123.20352172851562}}, "final_layer_norm": {"bias": 2.805612325668335, "scale": 24.572498321533203}, "layer_norm": {"bias": 2.2777209281921387, "scale": 24.348262786865234}}, "3": {"attention": {"k_proj": {"bias": 1.078221082687378, "kernel": 58.898963928222656}, "out_proj": {"bias": 1.445481300354004, "kernel": 54.040443420410156}, "q_proj": {"bias": 3.119992971420288, "kernel": 59.15928268432617}, "v_proj": {"bias": 0.38343948125839233, "kernel": 53.51526641845703}}, "feed_forward": {"intermediate_dense": {"bias": 2.567943572998047, "kernel": 115.53511047363281}, "output_dense": {"bias": 0.7814530730247498, "kernel": 102.66831970214844}}, "final_layer_norm": {"bias": 1.6788685321807861, "scale": 22.33795166015625}, "layer_norm": {"bias": 1.8063347339630127, "scale": 21.390769958496094}}, "30": {"attention": {"k_proj": {"bias": 1.2616255283355713, "kernel": 63.85869216918945}, "out_proj": {"bias": 1.471329927444458, "kernel": 59.489501953125}, "q_proj": {"bias": 3.18064022064209, "kernel": 64.04553985595703}, "v_proj": {"bias": 0.6499152779579163, "kernel": 60.14347839355469}}, "feed_forward": {"intermediate_dense": {"bias": 2.7467663288116455, "kernel": 119.56776428222656}, "output_dense": {"bias": 1.0970830917358398, "kernel": 123.16006469726562}}, "final_layer_norm": {"bias": 2.7424988746643066, "scale": 25.66531753540039}, "layer_norm": {"bias": 2.304725170135498, "scale": 24.183040618896484}}, "31": {"attention": {"k_proj": {"bias": 1.2824972867965698, "kernel": 62.397823333740234}, "out_proj": {"bias": 1.404205322265625, "kernel": 58.74039840698242}, "q_proj": {"bias": 2.95278263092041, "kernel": 62.655418395996094}, "v_proj": {"bias": 0.6554225087165833, "kernel": 59.31267166137695}}, "feed_forward": {"intermediate_dense": {"bias": 2.82468318939209, "kernel": 118.12161254882812}, "output_dense": {"bias": 1.2635955810546875, "kernel": 119.88945007324219}}, "final_layer_norm": {"bias": 2.5344667434692383, "scale": 25.44107437133789}, "layer_norm": {"bias": 2.3017654418945312, "scale": 23.798931121826172}}, "32": {"attention": {"k_proj": {"bias": 1.3159348964691162, "kernel": 63.632083892822266}, "out_proj": {"bias": 1.388063669204712, "kernel": 58.91508483886719}, "q_proj": {"bias": 3.120231866836548, "kernel": 63.72563934326172}, "v_proj": {"bias": 0.6113967895507812, "kernel": 59.90412902832031}}, "feed_forward": {"intermediate_dense": {"bias": 2.7874908447265625, "kernel": 117.41062927246094}, "output_dense": {"bias": 1.32664954662323, "kernel": 119.7316665649414}}, "final_layer_norm": {"bias": 2.5520706176757812, "scale": 25.718929290771484}, "layer_norm": {"bias": 2.4218976497650146, "scale": 23.878816604614258}}, "33": {"attention": {"k_proj": {"bias": 1.3422125577926636, "kernel": 63.386512756347656}, "out_proj": {"bias": 1.43208646774292, "kernel": 58.583740234375}, "q_proj": {"bias": 3.2318835258483887, "kernel": 63.64509963989258}, "v_proj": {"bias": 0.6489442586898804, "kernel": 59.56697082519531}}, "feed_forward": {"intermediate_dense": {"bias": 2.7939913272857666, "kernel": 116.08616638183594}, "output_dense": {"bias": 1.3533964157104492, "kernel": 117.61251831054688}}, "final_layer_norm": {"bias": 2.456862688064575, "scale": 25.378297805786133}, "layer_norm": {"bias": 2.506263256072998, "scale": 23.93863868713379}}, "34": {"attention": {"k_proj": {"bias": 1.3009648323059082, "kernel": 62.313804626464844}, "out_proj": {"bias": 1.686741590499878, "kernel": 58.781829833984375}, "q_proj": {"bias": 3.1833298206329346, "kernel": 62.64984893798828}, "v_proj": {"bias": 0.5979348421096802, "kernel": 59.77677917480469}}, "feed_forward": {"intermediate_dense": {"bias": 2.9047768115997314, "kernel": 114.74907684326172}, "output_dense": {"bias": 1.2986286878585815, "kernel": 115.74496459960938}}, "final_layer_norm": {"bias": 2.4029273986816406, "scale": 24.362728118896484}, "layer_norm": {"bias": 2.5617756843566895, "scale": 24.644542694091797}}, "35": {"attention": {"k_proj": {"bias": 1.4322527647018433, "kernel": 66.4811782836914}, "out_proj": {"bias": 1.5937174558639526, "kernel": 57.604217529296875}, "q_proj": {"bias": 2.8287200927734375, "kernel": 67.28070068359375}, "v_proj": {"bias": 0.5922476649284363, "kernel": 59.05278778076172}}, "feed_forward": {"intermediate_dense": {"bias": 3.0254740715026855, "kernel": 112.81585693359375}, "output_dense": {"bias": 1.1526522636413574, "kernel": 113.11029052734375}}, "final_layer_norm": {"bias": 2.4485955238342285, "scale": 24.011043548583984}, "layer_norm": {"bias": 2.547135353088379, "scale": 24.982009887695312}}, "36": {"attention": {"k_proj": {"bias": 1.41251540184021, "kernel": 63.22689437866211}, "out_proj": {"bias": 1.59316086769104, "kernel": 57.276268005371094}, "q_proj": {"bias": 2.829296112060547, "kernel": 63.69819641113281}, "v_proj": {"bias": 0.5033815503120422, "kernel": 59.06623458862305}}, "feed_forward": {"intermediate_dense": {"bias": 2.883385181427002, "kernel": 110.77723693847656}, "output_dense": {"bias": 1.1163235902786255, "kernel": 111.69965362548828}}, "final_layer_norm": {"bias": 2.3496971130371094, "scale": 24.36453628540039}, "layer_norm": {"bias": 2.4404845237731934, "scale": 24.4769287109375}}, "37": {"attention": {"k_proj": {"bias": 1.3988757133483887, "kernel": 60.617889404296875}, "out_proj": {"bias": 1.8305881023406982, "kernel": 56.0605354309082}, "q_proj": {"bias": 2.59777569770813, "kernel": 60.87814712524414}, "v_proj": {"bias": 0.4659739136695862, "kernel": 57.44471740722656}}, "feed_forward": {"intermediate_dense": {"bias": 2.7766342163085938, "kernel": 109.30412292480469}, "output_dense": {"bias": 1.1074990034103394, "kernel": 110.6824951171875}}, "final_layer_norm": {"bias": 2.116551399230957, "scale": 24.71451187133789}, "layer_norm": {"bias": 2.4344444274902344, "scale": 24.64586067199707}}, "38": {"attention": {"k_proj": {"bias": 1.3856570720672607, "kernel": 58.7510986328125}, "out_proj": {"bias": 1.5166680812835693, "kernel": 54.980918884277344}, "q_proj": {"bias": 2.485156774520874, "kernel": 59.025543212890625}, "v_proj": {"bias": 0.4676969647407532, "kernel": 56.200782775878906}}, "feed_forward": {"intermediate_dense": {"bias": 2.7416515350341797, "kernel": 106.81379699707031}, "output_dense": {"bias": 1.062075138092041, "kernel": 108.38115692138672}}, "final_layer_norm": {"bias": 2.151566743850708, "scale": 25.398880004882812}, "layer_norm": {"bias": 2.563223361968994, "scale": 25.652034759521484}}, "39": {"attention": {"k_proj": {"bias": 1.354166865348816, "kernel": 57.40496826171875}, "out_proj": {"bias": 1.7764017581939697, "kernel": 54.42431640625}, "q_proj": {"bias": 2.26041316986084, "kernel": 57.854835510253906}, "v_proj": {"bias": 0.48617270588874817, "kernel": 55.874244689941406}}, "feed_forward": {"intermediate_dense": {"bias": 2.7035012245178223, "kernel": 103.57746887207031}, "output_dense": {"bias": 1.1379706859588623, "kernel": 107.12802124023438}}, "final_layer_norm": {"bias": 2.033827304840088, "scale": 26.043302536010742}, "layer_norm": {"bias": 2.538135051727295, "scale": 26.209089279174805}}, "4": {"attention": {"k_proj": {"bias": 1.0600544214248657, "kernel": 61.16484451293945}, "out_proj": {"bias": 1.6709744930267334, "kernel": 55.83249282836914}, "q_proj": {"bias": 2.849893569946289, "kernel": 61.49186706542969}, "v_proj": {"bias": 0.4110661745071411, "kernel": 55.43656921386719}}, "feed_forward": {"intermediate_dense": {"bias": 2.5149760246276855, "kernel": 115.13277435302734}, "output_dense": {"bias": 0.9309735298156738, "kernel": 104.49896240234375}}, "final_layer_norm": {"bias": 1.8523075580596924, "scale": 21.96944808959961}, "layer_norm": {"bias": 1.9729788303375244, "scale": 22.860248565673828}}, "40": {"attention": {"k_proj": {"bias": 1.3378098011016846, "kernel": 55.470542907714844}, "out_proj": {"bias": 1.7115695476531982, "kernel": 51.82088088989258}, "q_proj": {"bias": 2.1932735443115234, "kernel": 56.186309814453125}, "v_proj": {"bias": 0.5169155597686768, "kernel": 52.51631546020508}}, "feed_forward": {"intermediate_dense": {"bias": 2.5498483180999756, "kernel": 100.59152221679688}, "output_dense": {"bias": 1.159660816192627, "kernel": 104.05167388916016}}, "final_layer_norm": {"bias": 1.9587419033050537, "scale": 25.324588775634766}, "layer_norm": {"bias": 2.432610034942627, "scale": 25.65159034729004}}, "41": {"attention": {"k_proj": {"bias": 2.130091667175293, "kernel": 54.96054458618164}, "out_proj": {"bias": 1.4786579608917236, "kernel": 53.876068115234375}, "q_proj": {"bias": 1.8969475030899048, "kernel": 55.1676139831543}, "v_proj": {"bias": 0.5349438786506653, "kernel": 54.362918853759766}}, "feed_forward": {"intermediate_dense": {"bias": 2.6417007446289062, "kernel": 96.72621154785156}, "output_dense": {"bias": 1.1966527700424194, "kernel": 102.03675842285156}}, "final_layer_norm": {"bias": 2.2963690757751465, "scale": 28.664409637451172}, "layer_norm": {"bias": 2.4220476150512695, "scale": 27.64695167541504}}, "42": {"attention": {"k_proj": {"bias": 1.4577641487121582, "kernel": 50.08251190185547}, "out_proj": {"bias": 1.5287971496582031, "kernel": 46.49677276611328}, "q_proj": {"bias": 1.7737632989883423, "kernel": 50.73402404785156}, "v_proj": {"bias": 0.7890726327896118, "kernel": 45.71551513671875}}, "feed_forward": {"intermediate_dense": {"bias": 2.542179584503174, "kernel": 96.53634643554688}, "output_dense": {"bias": 1.2405011653900146, "kernel": 100.6738052368164}}, "final_layer_norm": {"bias": 2.0554826259613037, "scale": 29.911632537841797}, "layer_norm": {"bias": 1.8721435070037842, "scale": 26.126949310302734}}, "43": {"attention": {"k_proj": {"bias": 1.6378315687179565, "kernel": 44.044639587402344}, "out_proj": {"bias": 1.484229564666748, "kernel": 42.509376525878906}, "q_proj": {"bias": 1.5668015480041504, "kernel": 44.567378997802734}, "v_proj": {"bias": 0.6460988521575928, "kernel": 40.88817596435547}}, "feed_forward": {"intermediate_dense": {"bias": 2.4490790367126465, "kernel": 94.14591217041016}, "output_dense": {"bias": 0.9738006591796875, "kernel": 97.92840576171875}}, "final_layer_norm": {"bias": 2.2158241271972656, "scale": 32.10318374633789}, "layer_norm": {"bias": 1.9498822689056396, "scale": 24.55547523498535}}, "44": {"attention": {"k_proj": {"bias": 2.7415380477905273, "kernel": 42.770957946777344}, "out_proj": {"bias": 1.2400130033493042, "kernel": 45.921775817871094}, "q_proj": {"bias": 1.5585159063339233, "kernel": 43.28406524658203}, "v_proj": {"bias": 0.4356144070625305, "kernel": 45.300472259521484}}, "feed_forward": {"intermediate_dense": {"bias": 2.4085373878479004, "kernel": 92.65846252441406}, "output_dense": {"bias": 0.8733353614807129, "kernel": 95.567138671875}}, "final_layer_norm": {"bias": 2.30061674118042, "scale": 34.31207275390625}, "layer_norm": {"bias": 1.8362083435058594, "scale": 24.757728576660156}}, "45": {"attention": {"k_proj": {"bias": 2.3060803413391113, "kernel": 41.182865142822266}, "out_proj": {"bias": 1.1073520183563232, "kernel": 49.445579528808594}, "q_proj": {"bias": 1.57840895652771, "kernel": 41.39216995239258}, "v_proj": {"bias": 0.4719974994659424, "kernel": 49.70841979980469}}, "feed_forward": {"intermediate_dense": {"bias": 2.391817092895508, "kernel": 90.02035522460938}, "output_dense": {"bias": 0.9732615351676941, "kernel": 91.68502044677734}}, "final_layer_norm": {"bias": 1.8946698904037476, "scale": 33.40074157714844}, "layer_norm": {"bias": 1.686659812927246, "scale": 23.539291381835938}}, "46": {"attention": {"k_proj": {"bias": 1.8325300216674805, "kernel": 40.81360626220703}, "out_proj": {"bias": 0.8864041566848755, "kernel": 51.90528869628906}, "q_proj": {"bias": 1.6992783546447754, "kernel": 41.834598541259766}, "v_proj": {"bias": 0.4735630750656128, "kernel": 52.756187438964844}}, "feed_forward": {"intermediate_dense": {"bias": 2.2736101150512695, "kernel": 86.09542846679688}, "output_dense": {"bias": 1.1628128290176392, "kernel": 83.48152923583984}}, "final_layer_norm": {"bias": 1.71113920211792, "scale": 29.332897186279297}, "layer_norm": {"bias": 1.5088059902191162, "scale": 22.557247161865234}}, "47": {"attention": {"k_proj": {"bias": 1.2887405157089233, "kernel": 43.97747039794922}, "out_proj": {"bias": 0.7762619256973267, "kernel": 48.44514083862305}, "q_proj": {"bias": 1.8569612503051758, "kernel": 45.550689697265625}, "v_proj": {"bias": 0.7446385025978088, "kernel": 49.13134765625}}, "feed_forward": {"intermediate_dense": {"bias": 2.2502808570861816, "kernel": 82.49830627441406}, "output_dense": {"bias": 0.688392162322998, "kernel": 77.83268737792969}}, "final_layer_norm": {"bias": 1.481636643409729, "scale": 23.951671600341797}, "layer_norm": {"bias": 1.561356782913208, "scale": 20.545452117919922}}, "5": {"attention": {"k_proj": {"bias": 1.017003059387207, "kernel": 56.419891357421875}, "out_proj": {"bias": 1.606635570526123, "kernel": 55.851890563964844}, "q_proj": {"bias": 3.0716123580932617, "kernel": 56.44148254394531}, "v_proj": {"bias": 0.37802064418792725, "kernel": 56.004173278808594}}, "feed_forward": {"intermediate_dense": {"bias": 2.447377920150757, "kernel": 114.9388427734375}, "output_dense": {"bias": 0.9671586751937866, "kernel": 103.54225158691406}}, "final_layer_norm": {"bias": 1.9784588813781738, "scale": 22.167034149169922}, "layer_norm": {"bias": 1.8326001167297363, "scale": 20.94150161743164}}, "6": {"attention": {"k_proj": {"bias": 1.11629056930542, "kernel": 59.132747650146484}, "out_proj": {"bias": 1.611763834953308, "kernel": 56.93556213378906}, "q_proj": {"bias": 3.0534703731536865, "kernel": 59.554317474365234}, "v_proj": {"bias": 0.4122595489025116, "kernel": 56.48927688598633}}, "feed_forward": {"intermediate_dense": {"bias": 2.416886806488037, "kernel": 114.41998291015625}, "output_dense": {"bias": 0.8545989990234375, "kernel": 104.102294921875}}, "final_layer_norm": {"bias": 2.2814440727233887, "scale": 21.629745483398438}, "layer_norm": {"bias": 1.9043277502059937, "scale": 22.332149505615234}}, "7": {"attention": {"k_proj": {"bias": 0.9863499999046326, "kernel": 57.55821228027344}, "out_proj": {"bias": 1.4638142585754395, "kernel": 56.85308074951172}, "q_proj": {"bias": 2.776463031768799, "kernel": 58.05531311035156}, "v_proj": {"bias": 0.46414023637771606, "kernel": 55.99853515625}}, "feed_forward": {"intermediate_dense": {"bias": 2.4128031730651855, "kernel": 114.094970703125}, "output_dense": {"bias": 0.6985006928443909, "kernel": 103.8736572265625}}, "final_layer_norm": {"bias": 2.260223865509033, "scale": 21.72259521484375}, "layer_norm": {"bias": 1.8673933744430542, "scale": 21.632614135742188}}, "8": {"attention": {"k_proj": {"bias": 1.0781749486923218, "kernel": 57.6885986328125}, "out_proj": {"bias": 1.2823827266693115, "kernel": 57.238609313964844}, "q_proj": {"bias": 2.785170316696167, "kernel": 57.816680908203125}, "v_proj": {"bias": 0.4353232681751251, "kernel": 56.5908203125}}, "feed_forward": {"intermediate_dense": {"bias": 2.44878888130188, "kernel": 113.56178283691406}, "output_dense": {"bias": 0.6725527048110962, "kernel": 103.45906066894531}}, "final_layer_norm": {"bias": 2.159350633621216, "scale": 21.46961784362793}, "layer_norm": {"bias": 1.830787181854248, "scale": 21.343868255615234}}, "9": {"attention": {"k_proj": {"bias": 1.1662862300872803, "kernel": 59.34821319580078}, "out_proj": {"bias": 1.5716854333877563, "kernel": 59.33329772949219}, "q_proj": {"bias": 2.597203254699707, "kernel": 59.712547302246094}, "v_proj": {"bias": 0.5281751751899719, "kernel": 58.699546813964844}}, "feed_forward": {"intermediate_dense": {"bias": 2.5084381103515625, "kernel": 112.08121490478516}, "output_dense": {"bias": 0.7975071668624878, "kernel": 103.30007934570312}}, "final_layer_norm": {"bias": 2.193687915802002, "scale": 20.689008712768555}, "layer_norm": {"bias": 2.059560775756836, "scale": 23.670970916748047}}}, "pos_conv_embed": {"conv": {"bias": 6.141986846923828, "weight_g": 9.608701705932617, "weight_v": 124.78767395019531}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 2.0290679931640625, "kernel": 20.55536460876465}, "layer_norm": {"bias": 4.550922393798828, "scale": 16.167570114135742}}, "1": {"conv": {"bias": 1.7790228128433228, "kernel": 51.24136734008789}, "layer_norm": {"bias": 5.962646961212158, "scale": 23.268157958984375}}, "2": {"conv": {"bias": 1.140576720237732, "kernel": 46.50312042236328}, "layer_norm": {"bias": 4.176670551300049, "scale": 20.370853424072266}}, "3": {"conv": {"bias": 0.6725863218307495, "kernel": 44.397525787353516}, "layer_norm": {"bias": 3.888174533843994, "scale": 17.53795051574707}}, "4": {"conv": {"bias": 0.6373162269592285, "kernel": 41.314056396484375}, "layer_norm": {"bias": 2.385471820831299, "scale": 16.34571647644043}}, "5": {"conv": {"bias": 0.5147221684455872, "kernel": 37.479759216308594}, "layer_norm": {"bias": 2.020900011062622, "scale": 17.064470291137695}}, "6": {"conv": {"bias": 0.4947893023490906, "kernel": 40.64780044555664}, "layer_norm": {"bias": 0.5876954793930054, "scale": 19.058603286743164}}}}, "feature_projection": {"layer_norm": {"bias": 6.459627151489258, "scale": 16.536361694335938}, "projection": {"bias": 2.151271343231201, "kernel": 43.19651794433594}}, "masked_spec_embed": 11.914372444152832}}, "train/learning_rate": 0.00033713760785758495, "train/loss": 0.4021756649017334, "train/param_norm": 1372.766845703125, "_runtime": 94582, "_timestamp": 1659387326, "_step": 33100, "eval/loss": 0.6790516376495361, "eval/wer": 0.46884239453991444, "eval/cer": 0.12758939024118068, "eval/step_4k": {"_type": "table-file", "path": "media/table/eval/step_4k_4000_af4cafd73c286841ef2f.table.json", "sha256": "af4cafd73c286841ef2fce257a64583667ab5412cd6837e4b951b2f851540450", "size": 24260, "artifact_path": "wandb-client-artifact://8wsujunwuradmkiy9teyal5atub99m9zxf1gks0x84p3y9bcbrdwjhad89ar5fxnyqtn8bopmk4501qsp1nuyvfeafw7p4spyxkt5zczfl5bxv88dzgxi32ukaj3dp4j:latest/eval/step_4k.table.json", "_latest_artifact_path": "wandb-client-artifact://8wsujunwuradmkiy9teyal5atub99m9zxf1gks0x84p3y9bcbrdwjhad89ar5fxnyqtn8bopmk4501qsp1nuyvfeafw7p4spyxkt5zczfl5bxv88dzgxi32ukaj3dp4j:latest/eval/step_4k.table.json", "ncols": 2, "nrows": 50}, "eval/step_8k": {"_type": "table-file", "path": "media/table/eval/step_8k_8000_c8ddc6e8e3a9e52ebbba.table.json", "sha256": "c8ddc6e8e3a9e52ebbbae9ac6ec8bb7ae6684781548fb4ea5c57a4b03a72d655", "size": 25822, "artifact_path": "wandb-client-artifact://och35iwsgdf5e7r6ebwp8bo4p13eowiyessoabvv4cf6keb0gk4e1577q2io23l2jh4jrzauz2qyodfw4w6u4eyf8llym88t1brov3snl0vrwcrq3dalvmazc40labfc:latest/eval/step_8k.table.json", "_latest_artifact_path": "wandb-client-artifact://och35iwsgdf5e7r6ebwp8bo4p13eowiyessoabvv4cf6keb0gk4e1577q2io23l2jh4jrzauz2qyodfw4w6u4eyf8llym88t1brov3snl0vrwcrq3dalvmazc40labfc:latest/eval/step_8k.table.json", "ncols": 2, "nrows": 50}, "eval/step_12k": {"_type": "table-file", "path": "media/table/eval/step_12k_12000_697630eb77c56222f807.table.json", "sha256": "697630eb77c56222f80728b3497df5ebfe62fb1dd060725ab84ec28fcf8448a3", "size": 25625, "artifact_path": "wandb-client-artifact://ndqwcshgeo7e3e4lndtrefve494sq9zmx9n9lrqdirtgy63uloydfb95oh1cytys0xi7ugpxbq1rub03y9scmcm41ocpbk826sbeejkgr3aubqet78b4jx4d8fb3z14k:latest/eval/step_12k.table.json", "_latest_artifact_path": "wandb-client-artifact://ndqwcshgeo7e3e4lndtrefve494sq9zmx9n9lrqdirtgy63uloydfb95oh1cytys0xi7ugpxbq1rub03y9scmcm41ocpbk826sbeejkgr3aubqet78b4jx4d8fb3z14k:latest/eval/step_12k.table.json", "ncols": 2, "nrows": 50}, "eval/step_16k": {"_type": "table-file", "path": "media/table/eval/step_16k_16000_a8af015baca8352e331a.table.json", "sha256": "a8af015baca8352e331a32965ddaa7fe22e2119a1c1256e539aedfd2cb876b87", "size": 25878, "artifact_path": "wandb-client-artifact://5lnl6ihmavrhu81b4ehn69ru2j2zbl0h27qv20im44v9928o4s69g9nvnb8oni6t0b921jwgbo24pz870kgos572o5h3vkwm71kr1brda3f3ooretb0u164vptzypekc:latest/eval/step_16k.table.json", "_latest_artifact_path": "wandb-client-artifact://5lnl6ihmavrhu81b4ehn69ru2j2zbl0h27qv20im44v9928o4s69g9nvnb8oni6t0b921jwgbo24pz870kgos572o5h3vkwm71kr1brda3f3ooretb0u164vptzypekc:latest/eval/step_16k.table.json", "ncols": 2, "nrows": 50}, "eval/step_20k": {"_type": "table-file", "path": "media/table/eval/step_20k_20000_37ce73b5cf7c7934cf62.table.json", "sha256": "37ce73b5cf7c7934cf62628174d0b0af065414de25e751b9e98983313d2b352e", "size": 25998, "artifact_path": "wandb-client-artifact://dhxussc15vwxa3x78tfa9270hif7fap7tu63y3eysgx549vfxso9k96434vhylx2uacjad6ldxuh7iavo5ogtn6fqtsv0u6d60zpvy5g99bbhwo8mmzkveldckon4ngf:latest/eval/step_20k.table.json", "_latest_artifact_path": "wandb-client-artifact://dhxussc15vwxa3x78tfa9270hif7fap7tu63y3eysgx549vfxso9k96434vhylx2uacjad6ldxuh7iavo5ogtn6fqtsv0u6d60zpvy5g99bbhwo8mmzkveldckon4ngf:latest/eval/step_20k.table.json", "ncols": 2, "nrows": 50}, "eval/step_24k": {"_type": "table-file", "path": "media/table/eval/step_24k_24000_6d0ed7e79108396fc292.table.json", "sha256": "6d0ed7e79108396fc292429957faec4cfead67d1cb5df1dfb6e4064ac1b8efd8", "size": 26106, "artifact_path": "wandb-client-artifact://vc70e9r1bqlbymt7rqzpvpurafx4zsrts0ic4qbqjmtia7a4qefxrwovecenb19dn3y65kbrjbmz63f1534kl7xoa0xst4f09yfus7kr0h0wn4i1hua9wqcf4hmyqgi6:latest/eval/step_24k.table.json", "_latest_artifact_path": "wandb-client-artifact://vc70e9r1bqlbymt7rqzpvpurafx4zsrts0ic4qbqjmtia7a4qefxrwovecenb19dn3y65kbrjbmz63f1534kl7xoa0xst4f09yfus7kr0h0wn4i1hua9wqcf4hmyqgi6:latest/eval/step_24k.table.json", "ncols": 2, "nrows": 50}, "eval/step_28k": {"_type": "table-file", "path": "media/table/eval/step_28k_28000_7186c63d506b9c841f41.table.json", "sha256": "7186c63d506b9c841f410c33dd5d77206b2d413f991c3b48e3b2b1265afbc518", "size": 26279, "artifact_path": "wandb-client-artifact://fdox4htembz1otwyfgt305vpupai0gi1hwwh6nh5zybkkex9v54fvigfcehdhyuj2kffoxiwqocy0n6gno40mw2grt1mc34m2kbnvn5z6b60nr5wh5uh1w8f9wmgicgs:latest/eval/step_28k.table.json", "_latest_artifact_path": "wandb-client-artifact://fdox4htembz1otwyfgt305vpupai0gi1hwwh6nh5zybkkex9v54fvigfcehdhyuj2kffoxiwqocy0n6gno40mw2grt1mc34m2kbnvn5z6b60nr5wh5uh1w8f9wmgicgs:latest/eval/step_28k.table.json", "ncols": 2, "nrows": 50}, "eval/step_32k": {"_type": "table-file", "path": "media/table/eval/step_32k_32000_c06fd8316235f01c9293.table.json", "sha256": "c06fd8316235f01c92939d85dbef4cfa3a9923249758ae21540020f5964e86e0", "size": 26339, "artifact_path": "wandb-client-artifact://nj3mc42ln908nf07bxueyratxdbzy955z8xe0w6unrzeupjeyzy24mks3az81d3d75a5fuyez97vi260kq2ylvdwkwgb1qafnnfo69bfxn3xwgwj3rcmvr4uvmj47iho:latest/eval/step_32k.table.json", "_latest_artifact_path": "wandb-client-artifact://nj3mc42ln908nf07bxueyratxdbzy955z8xe0w6unrzeupjeyzy24mks3az81d3d75a5fuyez97vi260kq2ylvdwkwgb1qafnnfo69bfxn3xwgwj3rcmvr4uvmj47iho:latest/eval/step_32k.table.json", "ncols": 2, "nrows": 50}, "_wandb": {"runtime": 94583}} \ No newline at end of file