diff --git "a/wandb/run-20220730_174606-j2u4n7h4/files/wandb-summary.json" "b/wandb/run-20220730_174606-j2u4n7h4/files/wandb-summary.json" --- "a/wandb/run-20220730_174606-j2u4n7h4/files/wandb-summary.json" +++ "b/wandb/run-20220730_174606-j2u4n7h4/files/wandb-summary.json" @@ -1 +1 @@ -{"train/grad_norm": 8.625, "layer_grad_norm/": {"lm_head": {"bias": 0.078125, "kernel": 1.53125}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 0.0615234375, "scale": 0.05517578125}, "layers": {"0": {"attention": {"k_proj": {"bias": 0.000286102294921875, "kernel": 0.162109375}, "out_proj": {"bias": 0.10693359375, "kernel": 0.9296875}, "q_proj": {"bias": 0.012451171875, "kernel": 0.2119140625}, "v_proj": {"bias": 0.07861328125, "kernel": 0.67578125}}, "feed_forward": {"intermediate_dense": {"bias": 0.10595703125, "kernel": 1.46875}, "output_dense": {"bias": 0.04541015625, "kernel": 1.25}}, "final_layer_norm": {"bias": 0.263671875, "scale": 0.3046875}, "layer_norm": {"bias": 0.138671875, "scale": 0.177734375}}, "1": {"attention": {"k_proj": {"bias": 0.0001163482666015625, "kernel": 0.12109375}, "out_proj": {"bias": 0.052490234375, "kernel": 0.6171875}, "q_proj": {"bias": 0.01068115234375, "kernel": 0.12255859375}, "v_proj": {"bias": 0.07470703125, "kernel": 0.4921875}}, "feed_forward": {"intermediate_dense": {"bias": 0.06640625, "kernel": 0.9609375}, "output_dense": {"bias": 0.04833984375, "kernel": 0.84375}}, "final_layer_norm": {"bias": 0.109375, "scale": 0.1083984375}, "layer_norm": {"bias": 0.11669921875, "scale": 0.15234375}}, "10": {"attention": {"k_proj": {"bias": 6.961822509765625e-05, "kernel": 0.21875}, "out_proj": {"bias": 0.033447265625, "kernel": 0.42578125}, "q_proj": {"bias": 0.01336669921875, "kernel": 0.228515625}, "v_proj": {"bias": 0.051025390625, "kernel": 0.48046875}}, "feed_forward": {"intermediate_dense": {"bias": 0.047607421875, "kernel": 0.75}, "output_dense": {"bias": 0.032470703125, "kernel": 0.578125}}, "final_layer_norm": {"bias": 0.08251953125, "scale": 0.0712890625}, "layer_norm": {"bias": 0.0947265625, "scale": 0.0908203125}}, "11": {"attention": {"k_proj": {"bias": 0.0001201629638671875, "kernel": 0.234375}, "out_proj": {"bias": 0.0322265625, "kernel": 0.5078125}, "q_proj": {"bias": 0.0135498046875, "kernel": 0.216796875}, "v_proj": {"bias": 0.053466796875, "kernel": 0.5703125}}, "feed_forward": {"intermediate_dense": {"bias": 0.04443359375, "kernel": 0.703125}, "output_dense": {"bias": 0.03173828125, "kernel": 0.5234375}}, "final_layer_norm": {"bias": 0.0712890625, "scale": 0.05859375}, "layer_norm": {"bias": 0.09130859375, "scale": 0.1259765625}}, "12": {"attention": {"k_proj": {"bias": 7.772445678710938e-05, "kernel": 0.203125}, "out_proj": {"bias": 0.03125, "kernel": 0.44140625}, "q_proj": {"bias": 0.0107421875, "kernel": 0.189453125}, "v_proj": {"bias": 0.051025390625, "kernel": 0.51953125}}, "feed_forward": {"intermediate_dense": {"bias": 0.046875, "kernel": 0.71875}, "output_dense": {"bias": 0.031005859375, "kernel": 0.5234375}}, "final_layer_norm": {"bias": 0.07666015625, "scale": 0.0615234375}, "layer_norm": {"bias": 0.0791015625, "scale": 0.08203125}}, "13": {"attention": {"k_proj": {"bias": 0.00011777877807617188, "kernel": 0.240234375}, "out_proj": {"bias": 0.03271484375, "kernel": 0.53125}, "q_proj": {"bias": 0.013916015625, "kernel": 0.2265625}, "v_proj": {"bias": 0.0556640625, "kernel": 0.6171875}}, "feed_forward": {"intermediate_dense": {"bias": 0.04736328125, "kernel": 0.703125}, "output_dense": {"bias": 0.032958984375, "kernel": 0.546875}}, "final_layer_norm": {"bias": 0.076171875, "scale": 0.07861328125}, "layer_norm": {"bias": 0.08154296875, "scale": 0.07958984375}}, "14": {"attention": {"k_proj": {"bias": 0.00014495849609375, "kernel": 0.212890625}, "out_proj": {"bias": 0.03125, "kernel": 0.47265625}, "q_proj": {"bias": 0.0150146484375, "kernel": 0.23046875}, "v_proj": {"bias": 0.052734375, "kernel": 0.5625}}, "feed_forward": {"intermediate_dense": {"bias": 0.04736328125, "kernel": 0.703125}, "output_dense": {"bias": 0.03173828125, "kernel": 0.5390625}}, "final_layer_norm": {"bias": 0.08203125, "scale": 0.0791015625}, "layer_norm": {"bias": 0.07861328125, "scale": 0.1083984375}}, "15": {"attention": {"k_proj": {"bias": 0.000141143798828125, "kernel": 0.244140625}, "out_proj": {"bias": 0.032470703125, "kernel": 0.60546875}, "q_proj": {"bias": 0.01446533203125, "kernel": 0.224609375}, "v_proj": {"bias": 0.056396484375, "kernel": 0.62109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.04345703125, "kernel": 0.625}, "output_dense": {"bias": 0.0322265625, "kernel": 0.5078125}}, "final_layer_norm": {"bias": 0.06884765625, "scale": 0.06201171875}, "layer_norm": {"bias": 0.07958984375, "scale": 0.08251953125}}, "16": {"attention": {"k_proj": {"bias": 8.869171142578125e-05, "kernel": 0.2333984375}, "out_proj": {"bias": 0.0306396484375, "kernel": 0.38671875}, "q_proj": {"bias": 0.012939453125, "kernel": 0.2197265625}, "v_proj": {"bias": 0.052490234375, "kernel": 0.474609375}}, "feed_forward": {"intermediate_dense": {"bias": 0.040771484375, "kernel": 0.60546875}, "output_dense": {"bias": 0.028076171875, "kernel": 0.4765625}}, "final_layer_norm": {"bias": 0.06640625, "scale": 0.05615234375}, "layer_norm": {"bias": 0.0791015625, "scale": 0.083984375}}, "17": {"attention": {"k_proj": {"bias": 7.200241088867188e-05, "kernel": 0.169921875}, "out_proj": {"bias": 0.0306396484375, "kernel": 0.34765625}, "q_proj": {"bias": 0.0096435546875, "kernel": 0.1630859375}, "v_proj": {"bias": 0.04736328125, "kernel": 0.421875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0419921875, "kernel": 0.6328125}, "output_dense": {"bias": 0.028564453125, "kernel": 0.486328125}}, "final_layer_norm": {"bias": 0.06787109375, "scale": 0.05810546875}, "layer_norm": {"bias": 0.0732421875, "scale": 0.0771484375}}, "18": {"attention": {"k_proj": {"bias": 0.00010442733764648438, "kernel": 0.234375}, "out_proj": {"bias": 0.025634765625, "kernel": 0.412109375}, "q_proj": {"bias": 0.0120849609375, "kernel": 0.20703125}, "v_proj": {"bias": 0.039794921875, "kernel": 0.41015625}}, "feed_forward": {"intermediate_dense": {"bias": 0.03466796875, "kernel": 0.546875}, "output_dense": {"bias": 0.0244140625, "kernel": 0.486328125}}, "final_layer_norm": {"bias": 0.055419921875, "scale": 0.046630859375}, "layer_norm": {"bias": 0.06298828125, "scale": 0.0673828125}}, "19": {"attention": {"k_proj": {"bias": 6.4849853515625e-05, "kernel": 0.162109375}, "out_proj": {"bias": 0.025146484375, "kernel": 0.30859375}, "q_proj": {"bias": 0.0091552734375, "kernel": 0.162109375}, "v_proj": {"bias": 0.037109375, "kernel": 0.34765625}}, "feed_forward": {"intermediate_dense": {"bias": 0.033447265625, "kernel": 0.55078125}, "output_dense": {"bias": 0.024169921875, "kernel": 0.48828125}}, "final_layer_norm": {"bias": 0.052734375, "scale": 0.052978515625}, "layer_norm": {"bias": 0.052734375, "scale": 0.05517578125}}, "2": {"attention": {"k_proj": {"bias": 0.0001277923583984375, "kernel": 0.173828125}, "out_proj": {"bias": 0.05712890625, "kernel": 0.66015625}, "q_proj": {"bias": 0.01519775390625, "kernel": 0.1689453125}, "v_proj": {"bias": 0.095703125, "kernel": 0.6875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0732421875, "kernel": 1.1953125}, "output_dense": {"bias": 0.051025390625, "kernel": 0.8671875}}, "final_layer_norm": {"bias": 0.12890625, "scale": 0.09619140625}, "layer_norm": {"bias": 0.130859375, "scale": 0.1484375}}, "20": {"attention": {"k_proj": {"bias": 3.0517578125e-05, "kernel": 0.09765625}, "out_proj": {"bias": 0.026611328125, "kernel": 0.205078125}, "q_proj": {"bias": 0.00604248046875, "kernel": 0.111328125}, "v_proj": {"bias": 0.03759765625, "kernel": 0.255859375}}, "feed_forward": {"intermediate_dense": {"bias": 0.03515625, "kernel": 0.59375}, "output_dense": {"bias": 0.0255126953125, "kernel": 0.4921875}}, "final_layer_norm": {"bias": 0.05712890625, "scale": 0.048095703125}, "layer_norm": {"bias": 0.0546875, "scale": 0.05908203125}}, "21": {"attention": {"k_proj": {"bias": 7.05718994140625e-05, "kernel": 0.1201171875}, "out_proj": {"bias": 0.026123046875, "kernel": 0.30078125}, "q_proj": {"bias": 0.00689697265625, "kernel": 0.12890625}, "v_proj": {"bias": 0.03662109375, "kernel": 0.333984375}}, "feed_forward": {"intermediate_dense": {"bias": 0.034423828125, "kernel": 0.59375}, "output_dense": {"bias": 0.0260009765625, "kernel": 0.490234375}}, "final_layer_norm": {"bias": 0.0556640625, "scale": 0.060791015625}, "layer_norm": {"bias": 0.049560546875, "scale": 0.04931640625}}, "22": {"attention": {"k_proj": {"bias": 4.7206878662109375e-05, "kernel": 0.1376953125}, "out_proj": {"bias": 0.0284423828125, "kernel": 0.2734375}, "q_proj": {"bias": 0.00830078125, "kernel": 0.154296875}, "v_proj": {"bias": 0.039306640625, "kernel": 0.30859375}}, "feed_forward": {"intermediate_dense": {"bias": 0.03662109375, "kernel": 0.6171875}, "output_dense": {"bias": 0.0283203125, "kernel": 0.48828125}}, "final_layer_norm": {"bias": 0.060791015625, "scale": 0.062255859375}, "layer_norm": {"bias": 0.05810546875, "scale": 0.083984375}}, "23": {"attention": {"k_proj": {"bias": 0.0001220703125, "kernel": 0.1904296875}, "out_proj": {"bias": 0.0302734375, "kernel": 0.4921875}, "q_proj": {"bias": 0.01025390625, "kernel": 0.1884765625}, "v_proj": {"bias": 0.044921875, "kernel": 0.49609375}}, "feed_forward": {"intermediate_dense": {"bias": 0.03662109375, "kernel": 0.625}, "output_dense": {"bias": 0.0296630859375, "kernel": 0.50390625}}, "final_layer_norm": {"bias": 0.06005859375, "scale": 0.06103515625}, "layer_norm": {"bias": 0.0654296875, "scale": 0.0732421875}}, "24": {"attention": {"k_proj": {"bias": 0.00010013580322265625, "kernel": 0.216796875}, "out_proj": {"bias": 0.0267333984375, "kernel": 0.388671875}, "q_proj": {"bias": 0.0115966796875, "kernel": 0.2109375}, "v_proj": {"bias": 0.04150390625, "kernel": 0.392578125}}, "feed_forward": {"intermediate_dense": {"bias": 0.03466796875, "kernel": 0.609375}, "output_dense": {"bias": 0.025634765625, "kernel": 0.4609375}}, "final_layer_norm": {"bias": 0.057373046875, "scale": 0.0595703125}, "layer_norm": {"bias": 0.0693359375, "scale": 0.0634765625}}, "25": {"attention": {"k_proj": {"bias": 0.00010061264038085938, "kernel": 0.1552734375}, "out_proj": {"bias": 0.027099609375, "kernel": 0.3984375}, "q_proj": {"bias": 0.0089111328125, "kernel": 0.154296875}, "v_proj": {"bias": 0.040771484375, "kernel": 0.4140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0341796875, "kernel": 0.59765625}, "output_dense": {"bias": 0.0263671875, "kernel": 0.4453125}}, "final_layer_norm": {"bias": 0.057861328125, "scale": 0.05859375}, "layer_norm": {"bias": 0.0615234375, "scale": 0.0703125}}, "26": {"attention": {"k_proj": {"bias": 8.535385131835938e-05, "kernel": 0.1669921875}, "out_proj": {"bias": 0.02734375, "kernel": 0.37890625}, "q_proj": {"bias": 0.009765625, "kernel": 0.17578125}, "v_proj": {"bias": 0.04296875, "kernel": 0.412109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.034912109375, "kernel": 0.5625}, "output_dense": {"bias": 0.02783203125, "kernel": 0.4453125}}, "final_layer_norm": {"bias": 0.0595703125, "scale": 0.0654296875}, "layer_norm": {"bias": 0.06103515625, "scale": 0.0625}}, "27": {"attention": {"k_proj": {"bias": 0.0001068115234375, "kernel": 0.205078125}, "out_proj": {"bias": 0.025390625, "kernel": 0.453125}, "q_proj": {"bias": 0.01171875, "kernel": 0.205078125}, "v_proj": {"bias": 0.040771484375, "kernel": 0.455078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0341796875, "kernel": 0.54296875}, "output_dense": {"bias": 0.026123046875, "kernel": 0.4296875}}, "final_layer_norm": {"bias": 0.058349609375, "scale": 0.05517578125}, "layer_norm": {"bias": 0.06591796875, "scale": 0.0634765625}}, "28": {"attention": {"k_proj": {"bias": 9.822845458984375e-05, "kernel": 0.17578125}, "out_proj": {"bias": 0.0240478515625, "kernel": 0.42578125}, "q_proj": {"bias": 0.009765625, "kernel": 0.1845703125}, "v_proj": {"bias": 0.0390625, "kernel": 0.431640625}}, "feed_forward": {"intermediate_dense": {"bias": 0.03173828125, "kernel": 0.52734375}, "output_dense": {"bias": 0.0247802734375, "kernel": 0.421875}}, "final_layer_norm": {"bias": 0.052734375, "scale": 0.0556640625}, "layer_norm": {"bias": 0.062255859375, "scale": 0.0498046875}}, "29": {"attention": {"k_proj": {"bias": 7.152557373046875e-05, "kernel": 0.1787109375}, "out_proj": {"bias": 0.022216796875, "kernel": 0.33984375}, "q_proj": {"bias": 0.009521484375, "kernel": 0.1875}, "v_proj": {"bias": 0.03515625, "kernel": 0.3671875}}, "feed_forward": {"intermediate_dense": {"bias": 0.030029296875, "kernel": 0.55078125}, "output_dense": {"bias": 0.0218505859375, "kernel": 0.40234375}}, "final_layer_norm": {"bias": 0.04736328125, "scale": 0.046875}, "layer_norm": {"bias": 0.06201171875, "scale": 0.048583984375}}, "3": {"attention": {"k_proj": {"bias": 0.000209808349609375, "kernel": 0.25390625}, "out_proj": {"bias": 0.0556640625, "kernel": 0.84375}, "q_proj": {"bias": 0.0181884765625, "kernel": 0.248046875}, "v_proj": {"bias": 0.09423828125, "kernel": 0.875}}, "feed_forward": {"intermediate_dense": {"bias": 0.07421875, "kernel": 1.203125}, "output_dense": {"bias": 0.048095703125, "kernel": 0.8359375}}, "final_layer_norm": {"bias": 0.12890625, "scale": 0.1171875}, "layer_norm": {"bias": 0.1298828125, "scale": 0.11865234375}}, "30": {"attention": {"k_proj": {"bias": 0.00010585784912109375, "kernel": 0.2119140625}, "out_proj": {"bias": 0.02099609375, "kernel": 0.37890625}, "q_proj": {"bias": 0.0113525390625, "kernel": 0.2255859375}, "v_proj": {"bias": 0.03271484375, "kernel": 0.392578125}}, "feed_forward": {"intermediate_dense": {"bias": 0.02783203125, "kernel": 0.51953125}, "output_dense": {"bias": 0.0205078125, "kernel": 0.365234375}}, "final_layer_norm": {"bias": 0.04345703125, "scale": 0.0478515625}, "layer_norm": {"bias": 0.04638671875, "scale": 0.06005859375}}, "31": {"attention": {"k_proj": {"bias": 0.00010395050048828125, "kernel": 0.181640625}, "out_proj": {"bias": 0.0194091796875, "kernel": 0.326171875}, "q_proj": {"bias": 0.0096435546875, "kernel": 0.18359375}, "v_proj": {"bias": 0.0296630859375, "kernel": 0.359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.02587890625, "kernel": 0.482421875}, "output_dense": {"bias": 0.0184326171875, "kernel": 0.337890625}}, "final_layer_norm": {"bias": 0.040771484375, "scale": 0.044921875}, "layer_norm": {"bias": 0.04541015625, "scale": 0.057373046875}}, "32": {"attention": {"k_proj": {"bias": 9.775161743164062e-05, "kernel": 0.15625}, "out_proj": {"bias": 0.017578125, "kernel": 0.271484375}, "q_proj": {"bias": 0.008056640625, "kernel": 0.169921875}, "v_proj": {"bias": 0.026123046875, "kernel": 0.310546875}}, "feed_forward": {"intermediate_dense": {"bias": 0.024169921875, "kernel": 0.462890625}, "output_dense": {"bias": 0.0159912109375, "kernel": 0.314453125}}, "final_layer_norm": {"bias": 0.03955078125, "scale": 0.037109375}, "layer_norm": {"bias": 0.03955078125, "scale": 0.05615234375}}, "33": {"attention": {"k_proj": {"bias": 0.0001125335693359375, "kernel": 0.2177734375}, "out_proj": {"bias": 0.0150146484375, "kernel": 0.28515625}, "q_proj": {"bias": 0.010498046875, "kernel": 0.2216796875}, "v_proj": {"bias": 0.023681640625, "kernel": 0.30859375}}, "feed_forward": {"intermediate_dense": {"bias": 0.021240234375, "kernel": 0.40234375}, "output_dense": {"bias": 0.0140380859375, "kernel": 0.2890625}}, "final_layer_norm": {"bias": 0.03564453125, "scale": 0.03662109375}, "layer_norm": {"bias": 0.0361328125, "scale": 0.06591796875}}, "34": {"attention": {"k_proj": {"bias": 0.000141143798828125, "kernel": 0.189453125}, "out_proj": {"bias": 0.0123291015625, "kernel": 0.275390625}, "q_proj": {"bias": 0.0089111328125, "kernel": 0.1875}, "v_proj": {"bias": 0.01953125, "kernel": 0.2890625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0179443359375, "kernel": 0.341796875}, "output_dense": {"bias": 0.011474609375, "kernel": 0.2578125}}, "final_layer_norm": {"bias": 0.0301513671875, "scale": 0.03076171875}, "layer_norm": {"bias": 0.0306396484375, "scale": 0.0390625}}, "35": {"attention": {"k_proj": {"bias": 0.00018405914306640625, "kernel": 0.12890625}, "out_proj": {"bias": 0.010498046875, "kernel": 0.26953125}, "q_proj": {"bias": 0.00579833984375, "kernel": 0.12158203125}, "v_proj": {"bias": 0.01434326171875, "kernel": 0.244140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0150146484375, "kernel": 0.28125}, "output_dense": {"bias": 0.0101318359375, "kernel": 0.228515625}}, "final_layer_norm": {"bias": 0.025390625, "scale": 0.0262451171875}, "layer_norm": {"bias": 0.02490234375, "scale": 0.025390625}}, "36": {"attention": {"k_proj": {"bias": 6.628036499023438e-05, "kernel": 0.130859375}, "out_proj": {"bias": 0.009521484375, "kernel": 0.20703125}, "q_proj": {"bias": 0.00592041015625, "kernel": 0.12353515625}, "v_proj": {"bias": 0.0128173828125, "kernel": 0.1875}}, "feed_forward": {"intermediate_dense": {"bias": 0.013427734375, "kernel": 0.24609375}, "output_dense": {"bias": 0.0093994140625, "kernel": 0.2021484375}}, "final_layer_norm": {"bias": 0.0223388671875, "scale": 0.021484375}, "layer_norm": {"bias": 0.022705078125, "scale": 0.025146484375}}, "37": {"attention": {"k_proj": {"bias": 6.67572021484375e-05, "kernel": 0.10302734375}, "out_proj": {"bias": 0.0089111328125, "kernel": 0.205078125}, "q_proj": {"bias": 0.00482177734375, "kernel": 0.10302734375}, "v_proj": {"bias": 0.0128173828125, "kernel": 0.189453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.0123291015625, "kernel": 0.2314453125}, "output_dense": {"bias": 0.0084228515625, "kernel": 0.19140625}}, "final_layer_norm": {"bias": 0.020751953125, "scale": 0.021484375}, "layer_norm": {"bias": 0.023193359375, "scale": 0.025634765625}}, "38": {"attention": {"k_proj": {"bias": 6.67572021484375e-05, "kernel": 0.11767578125}, "out_proj": {"bias": 0.00787353515625, "kernel": 0.185546875}, "q_proj": {"bias": 0.00445556640625, "kernel": 0.103515625}, "v_proj": {"bias": 0.01116943359375, "kernel": 0.173828125}}, "feed_forward": {"intermediate_dense": {"bias": 0.010986328125, "kernel": 0.212890625}, "output_dense": {"bias": 0.007476806640625, "kernel": 0.18359375}}, "final_layer_norm": {"bias": 0.018798828125, "scale": 0.018798828125}, "layer_norm": {"bias": 0.02001953125, "scale": 0.0234375}}, "39": {"attention": {"k_proj": {"bias": 4.8160552978515625e-05, "kernel": 0.07568359375}, "out_proj": {"bias": 0.0067138671875, "kernel": 0.17578125}, "q_proj": {"bias": 0.003173828125, "kernel": 0.07421875}, "v_proj": {"bias": 0.009521484375, "kernel": 0.1552734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.010009765625, "kernel": 0.2001953125}, "output_dense": {"bias": 0.006439208984375, "kernel": 0.19140625}}, "final_layer_norm": {"bias": 0.0177001953125, "scale": 0.01806640625}, "layer_norm": {"bias": 0.0166015625, "scale": 0.0167236328125}}, "4": {"attention": {"k_proj": {"bias": 0.00026702880859375, "kernel": 0.271484375}, "out_proj": {"bias": 0.05078125, "kernel": 0.90234375}, "q_proj": {"bias": 0.01708984375, "kernel": 0.26953125}, "v_proj": {"bias": 0.0791015625, "kernel": 0.91796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0712890625, "kernel": 1.09375}, "output_dense": {"bias": 0.048095703125, "kernel": 0.80078125}}, "final_layer_norm": {"bias": 0.1201171875, "scale": 0.125}, "layer_norm": {"bias": 0.111328125, "scale": 0.12109375}}, "40": {"attention": {"k_proj": {"bias": 2.7894973754882812e-05, "kernel": 0.061767578125}, "out_proj": {"bias": 0.00628662109375, "kernel": 0.14453125}, "q_proj": {"bias": 0.0025177001953125, "kernel": 0.060546875}, "v_proj": {"bias": 0.00848388671875, "kernel": 0.134765625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0091552734375, "kernel": 0.181640625}, "output_dense": {"bias": 0.00616455078125, "kernel": 0.1591796875}}, "final_layer_norm": {"bias": 0.0172119140625, "scale": 0.0264892578125}, "layer_norm": {"bias": 0.014404296875, "scale": 0.0152587890625}}, "41": {"attention": {"k_proj": {"bias": 2.765655517578125e-05, "kernel": 0.06689453125}, "out_proj": {"bias": 0.005706787109375, "kernel": 0.14453125}, "q_proj": {"bias": 0.0027923583984375, "kernel": 0.06494140625}, "v_proj": {"bias": 0.0084228515625, "kernel": 0.1552734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0079345703125, "kernel": 0.1708984375}, "output_dense": {"bias": 0.00567626953125, "kernel": 0.1484375}}, "final_layer_norm": {"bias": 0.0147705078125, "scale": 0.0211181640625}, "layer_norm": {"bias": 0.0146484375, "scale": 0.023681640625}}, "42": {"attention": {"k_proj": {"bias": 1.2695789337158203e-05, "kernel": 0.029296875}, "out_proj": {"bias": 0.0057373046875, "kernel": 0.1064453125}, "q_proj": {"bias": 0.00138092041015625, "kernel": 0.031005859375}, "v_proj": {"bias": 0.006866455078125, "kernel": 0.103515625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00714111328125, "kernel": 0.158203125}, "output_dense": {"bias": 0.005859375, "kernel": 0.126953125}}, "final_layer_norm": {"bias": 0.01220703125, "scale": 0.0130615234375}, "layer_norm": {"bias": 0.01123046875, "scale": 0.0159912109375}}, "43": {"attention": {"k_proj": {"bias": 6.705522537231445e-06, "kernel": 0.0203857421875}, "out_proj": {"bias": 0.00604248046875, "kernel": 0.0859375}, "q_proj": {"bias": 0.0010833740234375, "kernel": 0.02294921875}, "v_proj": {"bias": 0.00714111328125, "kernel": 0.09619140625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00775146484375, "kernel": 0.181640625}, "output_dense": {"bias": 0.0062255859375, "kernel": 0.1318359375}}, "final_layer_norm": {"bias": 0.013427734375, "scale": 0.012939453125}, "layer_norm": {"bias": 0.01202392578125, "scale": 0.0166015625}}, "44": {"attention": {"k_proj": {"bias": 5.781650543212891e-06, "kernel": 0.02197265625}, "out_proj": {"bias": 0.00640869140625, "kernel": 0.09765625}, "q_proj": {"bias": 0.00118255615234375, "kernel": 0.024658203125}, "v_proj": {"bias": 0.00726318359375, "kernel": 0.10107421875}}, "feed_forward": {"intermediate_dense": {"bias": 0.006927490234375, "kernel": 0.177734375}, "output_dense": {"bias": 0.006622314453125, "kernel": 0.1201171875}}, "final_layer_norm": {"bias": 0.010986328125, "scale": 0.01025390625}, "layer_norm": {"bias": 0.01361083984375, "scale": 0.0135498046875}}, "45": {"attention": {"k_proj": {"bias": 7.748603820800781e-06, "kernel": 0.0220947265625}, "out_proj": {"bias": 0.00665283203125, "kernel": 0.10107421875}, "q_proj": {"bias": 0.0012054443359375, "kernel": 0.0244140625}, "v_proj": {"bias": 0.0079345703125, "kernel": 0.1064453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00653076171875, "kernel": 0.166015625}, "output_dense": {"bias": 0.006744384765625, "kernel": 0.1201171875}}, "final_layer_norm": {"bias": 0.01055908203125, "scale": 0.0133056640625}, "layer_norm": {"bias": 0.0179443359375, "scale": 0.014892578125}}, "46": {"attention": {"k_proj": {"bias": 9.119510650634766e-06, "kernel": 0.0277099609375}, "out_proj": {"bias": 0.0067138671875, "kernel": 0.103515625}, "q_proj": {"bias": 0.001373291015625, "kernel": 0.027099609375}, "v_proj": {"bias": 0.00885009765625, "kernel": 0.1201171875}}, "feed_forward": {"intermediate_dense": {"bias": 0.00555419921875, "kernel": 0.1337890625}, "output_dense": {"bias": 0.00653076171875, "kernel": 0.11328125}}, "final_layer_norm": {"bias": 0.00946044921875, "scale": 0.01129150390625}, "layer_norm": {"bias": 0.0224609375, "scale": 0.0140380859375}}, "47": {"attention": {"k_proj": {"bias": 1.0609626770019531e-05, "kernel": 0.0286865234375}, "out_proj": {"bias": 0.006744384765625, "kernel": 0.08984375}, "q_proj": {"bias": 0.0015411376953125, "kernel": 0.025146484375}, "v_proj": {"bias": 0.01025390625, "kernel": 0.12890625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00482177734375, "kernel": 0.09521484375}, "output_dense": {"bias": 0.006591796875, "kernel": 0.10107421875}}, "final_layer_norm": {"bias": 0.0093994140625, "scale": 0.0101318359375}, "layer_norm": {"bias": 0.029296875, "scale": 0.022705078125}}, "5": {"attention": {"k_proj": {"bias": 0.00012493133544921875, "kernel": 0.25390625}, "out_proj": {"bias": 0.05078125, "kernel": 0.6171875}, "q_proj": {"bias": 0.01544189453125, "kernel": 0.255859375}, "v_proj": {"bias": 0.0810546875, "kernel": 0.68359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.06787109375, "kernel": 1.0078125}, "output_dense": {"bias": 0.048095703125, "kernel": 0.7421875}}, "final_layer_norm": {"bias": 0.1142578125, "scale": 0.103515625}, "layer_norm": {"bias": 0.1162109375, "scale": 0.09033203125}}, "6": {"attention": {"k_proj": {"bias": 0.0001678466796875, "kernel": 0.322265625}, "out_proj": {"bias": 0.044921875, "kernel": 0.75}, "q_proj": {"bias": 0.02099609375, "kernel": 0.31640625}, "v_proj": {"bias": 0.08154296875, "kernel": 0.84765625}}, "feed_forward": {"intermediate_dense": {"bias": 0.06201171875, "kernel": 0.98046875}, "output_dense": {"bias": 0.04248046875, "kernel": 0.671875}}, "final_layer_norm": {"bias": 0.10546875, "scale": 0.10546875}, "layer_norm": {"bias": 0.123046875, "scale": 0.1025390625}}, "7": {"attention": {"k_proj": {"bias": 0.00022792816162109375, "kernel": 0.2734375}, "out_proj": {"bias": 0.042724609375, "kernel": 0.71875}, "q_proj": {"bias": 0.016845703125, "kernel": 0.265625}, "v_proj": {"bias": 0.06982421875, "kernel": 0.765625}}, "feed_forward": {"intermediate_dense": {"bias": 0.06005859375, "kernel": 0.95703125}, "output_dense": {"bias": 0.040283203125, "kernel": 0.65234375}}, "final_layer_norm": {"bias": 0.10107421875, "scale": 0.10107421875}, "layer_norm": {"bias": 0.11279296875, "scale": 0.138671875}}, "8": {"attention": {"k_proj": {"bias": 0.00016880035400390625, "kernel": 0.25}, "out_proj": {"bias": 0.0390625, "kernel": 0.625}, "q_proj": {"bias": 0.01513671875, "kernel": 0.2421875}, "v_proj": {"bias": 0.06494140625, "kernel": 0.68359375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0556640625, "kernel": 0.890625}, "output_dense": {"bias": 0.037109375, "kernel": 0.625}}, "final_layer_norm": {"bias": 0.09765625, "scale": 0.095703125}, "layer_norm": {"bias": 0.1044921875, "scale": 0.07861328125}}, "9": {"attention": {"k_proj": {"bias": 0.0002117156982421875, "kernel": 0.2734375}, "out_proj": {"bias": 0.03369140625, "kernel": 0.734375}, "q_proj": {"bias": 0.015625, "kernel": 0.2734375}, "v_proj": {"bias": 0.0556640625, "kernel": 0.78125}}, "feed_forward": {"intermediate_dense": {"bias": 0.04736328125, "kernel": 0.7890625}, "output_dense": {"bias": 0.0341796875, "kernel": 0.609375}}, "final_layer_norm": {"bias": 0.080078125, "scale": 0.09228515625}, "layer_norm": {"bias": 0.091796875, "scale": 0.0869140625}}}, "pos_conv_embed": {"conv": {"bias": 0.099609375, "weight_g": 0.0771484375, "weight_v": 0.765625}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "1": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "2": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "3": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "4": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "5": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "6": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}}}, "feature_projection": {"layer_norm": {"bias": 0.265625, "scale": 0.384765625}, "projection": {"bias": 0.1533203125, "kernel": 3.09375}}, "masked_spec_embed": 0.0}}, "layer_param_norm/": {"lm_head": {"bias": 0.04177431762218475, "kernel": 5.20725679397583}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 0.8606863617897034, "scale": 22.605243682861328}, "layers": {"0": {"attention": {"k_proj": {"bias": 0.06318297237157822, "kernel": 25.94121742248535}, "out_proj": {"bias": 1.5376051664352417, "kernel": 25.12331771850586}, "q_proj": {"bias": 1.3132240772247314, "kernel": 26.238143920898438}, "v_proj": {"bias": 0.34682798385620117, "kernel": 25.811904907226562}}, "feed_forward": {"intermediate_dense": {"bias": 1.7679121494293213, "kernel": 95.302001953125}, "output_dense": {"bias": 1.0227036476135254, "kernel": 91.0895767211914}}, "final_layer_norm": {"bias": 1.2829424142837524, "scale": 19.88983726501465}, "layer_norm": {"bias": 3.278278350830078, "scale": 15.982975006103516}}, "1": {"attention": {"k_proj": {"bias": 0.06595329195261002, "kernel": 40.10105895996094}, "out_proj": {"bias": 1.2929775714874268, "kernel": 41.66594314575195}, "q_proj": {"bias": 2.86798357963562, "kernel": 39.95466613769531}, "v_proj": {"bias": 0.2810550332069397, "kernel": 40.132972717285156}}, "feed_forward": {"intermediate_dense": {"bias": 1.5912463665008545, "kernel": 93.38056945800781}, "output_dense": {"bias": 0.801348090171814, "kernel": 84.22969055175781}}, "final_layer_norm": {"bias": 1.1487457752227783, "scale": 18.430646896362305}, "layer_norm": {"bias": 1.7521967887878418, "scale": 19.26400375366211}}, "10": {"attention": {"k_proj": {"bias": 0.0756155252456665, "kernel": 47.322914123535156}, "out_proj": {"bias": 1.2240347862243652, "kernel": 50.1451530456543}, "q_proj": {"bias": 2.4277195930480957, "kernel": 47.28287887573242}, "v_proj": {"bias": 0.3121947944164276, "kernel": 50.34102249145508}}, "feed_forward": {"intermediate_dense": {"bias": 1.630620002746582, "kernel": 97.70709228515625}, "output_dense": {"bias": 0.5621562004089355, "kernel": 91.62409973144531}}, "final_layer_norm": {"bias": 2.204315185546875, "scale": 20.36707305908203}, "layer_norm": {"bias": 1.6952250003814697, "scale": 22.25482749938965}}, "11": {"attention": {"k_proj": {"bias": 0.11612671613693237, "kernel": 47.13056945800781}, "out_proj": {"bias": 1.0761620998382568, "kernel": 49.34977722167969}, "q_proj": {"bias": 2.477080821990967, "kernel": 46.85191345214844}, "v_proj": {"bias": 0.35772353410720825, "kernel": 49.886878967285156}}, "feed_forward": {"intermediate_dense": {"bias": 1.6804685592651367, "kernel": 98.45591735839844}, "output_dense": {"bias": 0.5454298257827759, "kernel": 93.32662963867188}}, "final_layer_norm": {"bias": 2.1902666091918945, "scale": 20.374927520751953}, "layer_norm": {"bias": 1.6745539903640747, "scale": 22.586902618408203}}, "12": {"attention": {"k_proj": {"bias": 0.07888297736644745, "kernel": 47.714202880859375}, "out_proj": {"bias": 1.0603365898132324, "kernel": 49.6598014831543}, "q_proj": {"bias": 2.3620338439941406, "kernel": 47.46424102783203}, "v_proj": {"bias": 0.34527692198753357, "kernel": 50.072296142578125}}, "feed_forward": {"intermediate_dense": {"bias": 1.723632574081421, "kernel": 99.29413604736328}, "output_dense": {"bias": 0.5341137647628784, "kernel": 94.87334442138672}}, "final_layer_norm": {"bias": 2.1468374729156494, "scale": 20.335224151611328}, "layer_norm": {"bias": 1.74271821975708, "scale": 23.122257232666016}}, "13": {"attention": {"k_proj": {"bias": 0.09656181931495667, "kernel": 49.598751068115234}, "out_proj": {"bias": 1.0522843599319458, "kernel": 49.309295654296875}, "q_proj": {"bias": 2.333427667617798, "kernel": 49.457244873046875}, "v_proj": {"bias": 0.3668674826622009, "kernel": 49.486968994140625}}, "feed_forward": {"intermediate_dense": {"bias": 1.778700828552246, "kernel": 99.90986633300781}, "output_dense": {"bias": 0.5496945977210999, "kernel": 95.31092834472656}}, "final_layer_norm": {"bias": 2.032881259918213, "scale": 20.484725952148438}, "layer_norm": {"bias": 1.8432364463806152, "scale": 23.351295471191406}}, "14": {"attention": {"k_proj": {"bias": 0.1631498485803604, "kernel": 49.78543472290039}, "out_proj": {"bias": 1.2131561040878296, "kernel": 47.72584533691406}, "q_proj": {"bias": 2.3975725173950195, "kernel": 49.847618103027344}, "v_proj": {"bias": 0.3659938871860504, "kernel": 47.30793762207031}}, "feed_forward": {"intermediate_dense": {"bias": 1.8164732456207275, "kernel": 100.52674865722656}, "output_dense": {"bias": 0.5669653415679932, "kernel": 96.67445373535156}}, "final_layer_norm": {"bias": 2.1629538536071777, "scale": 20.62717056274414}, "layer_norm": {"bias": 1.9742047786712646, "scale": 23.513076782226562}}, "15": {"attention": {"k_proj": {"bias": 0.10221116989850998, "kernel": 49.92227554321289}, "out_proj": {"bias": 1.2649565935134888, "kernel": 48.34504699707031}, "q_proj": {"bias": 2.538086414337158, "kernel": 49.99497985839844}, "v_proj": {"bias": 0.4016108512878418, "kernel": 47.998756408691406}}, "feed_forward": {"intermediate_dense": {"bias": 1.8245656490325928, "kernel": 100.34303283691406}, "output_dense": {"bias": 0.7157814502716064, "kernel": 97.36470031738281}}, "final_layer_norm": {"bias": 2.092581272125244, "scale": 20.723337173461914}, "layer_norm": {"bias": 2.223583221435547, "scale": 23.69983673095703}}, "16": {"attention": {"k_proj": {"bias": 0.07193054258823395, "kernel": 49.82077407836914}, "out_proj": {"bias": 1.1955220699310303, "kernel": 47.801273345947266}, "q_proj": {"bias": 2.6239218711853027, "kernel": 49.705780029296875}, "v_proj": {"bias": 0.3556715250015259, "kernel": 47.48687744140625}}, "feed_forward": {"intermediate_dense": {"bias": 1.8201916217803955, "kernel": 100.99823760986328}, "output_dense": {"bias": 0.7402896285057068, "kernel": 98.2271957397461}}, "final_layer_norm": {"bias": 2.1531410217285156, "scale": 21.210491180419922}, "layer_norm": {"bias": 2.152494192123413, "scale": 22.587139129638672}}, "17": {"attention": {"k_proj": {"bias": 0.06806065142154694, "kernel": 50.00645446777344}, "out_proj": {"bias": 1.1380884647369385, "kernel": 47.11663055419922}, "q_proj": {"bias": 2.7014288902282715, "kernel": 50.10053253173828}, "v_proj": {"bias": 0.39294928312301636, "kernel": 46.79236602783203}}, "feed_forward": {"intermediate_dense": {"bias": 1.8335134983062744, "kernel": 102.07868957519531}, "output_dense": {"bias": 0.7559198141098022, "kernel": 98.61711120605469}}, "final_layer_norm": {"bias": 2.2452797889709473, "scale": 21.767436981201172}, "layer_norm": {"bias": 2.0604021549224854, "scale": 22.113117218017578}}, "18": {"attention": {"k_proj": {"bias": 0.0945877730846405, "kernel": 50.32107162475586}, "out_proj": {"bias": 1.2434324026107788, "kernel": 48.143550872802734}, "q_proj": {"bias": 2.585094451904297, "kernel": 50.70917510986328}, "v_proj": {"bias": 0.4215960204601288, "kernel": 47.67515563964844}}, "feed_forward": {"intermediate_dense": {"bias": 1.8761192560195923, "kernel": 102.37921905517578}, "output_dense": {"bias": 0.8710594177246094, "kernel": 100.26573181152344}}, "final_layer_norm": {"bias": 2.352644443511963, "scale": 21.729122161865234}, "layer_norm": {"bias": 2.2498598098754883, "scale": 23.874679565429688}}, "19": {"attention": {"k_proj": {"bias": 0.06790818274021149, "kernel": 49.573272705078125}, "out_proj": {"bias": 1.2166444063186646, "kernel": 48.02541732788086}, "q_proj": {"bias": 2.8660926818847656, "kernel": 50.001224517822266}, "v_proj": {"bias": 0.38429099321365356, "kernel": 47.27474594116211}}, "feed_forward": {"intermediate_dense": {"bias": 1.928505301475525, "kernel": 102.95552062988281}, "output_dense": {"bias": 0.9360222816467285, "kernel": 101.17387390136719}}, "final_layer_norm": {"bias": 2.304023265838623, "scale": 22.090991973876953}, "layer_norm": {"bias": 2.1650490760803223, "scale": 23.05562400817871}}, "2": {"attention": {"k_proj": {"bias": 0.07576870918273926, "kernel": 46.154396057128906}, "out_proj": {"bias": 1.2100125551223755, "kernel": 43.89133071899414}, "q_proj": {"bias": 3.055267333984375, "kernel": 45.918235778808594}, "v_proj": {"bias": 0.3107410669326782, "kernel": 43.885040283203125}}, "feed_forward": {"intermediate_dense": {"bias": 1.6276273727416992, "kernel": 98.4322509765625}, "output_dense": {"bias": 0.6901171207427979, "kernel": 87.37322235107422}}, "final_layer_norm": {"bias": 1.4538421630859375, "scale": 21.007522583007812}, "layer_norm": {"bias": 1.6640784740447998, "scale": 21.99614906311035}}, "20": {"attention": {"k_proj": {"bias": 0.06599634885787964, "kernel": 49.55154037475586}, "out_proj": {"bias": 1.2455387115478516, "kernel": 47.410400390625}, "q_proj": {"bias": 2.778987407684326, "kernel": 50.334571838378906}, "v_proj": {"bias": 0.3564513325691223, "kernel": 46.315513610839844}}, "feed_forward": {"intermediate_dense": {"bias": 1.9287090301513672, "kernel": 104.21571350097656}, "output_dense": {"bias": 1.051514983177185, "kernel": 101.80191040039062}}, "final_layer_norm": {"bias": 2.329120397567749, "scale": 23.03461456298828}, "layer_norm": {"bias": 2.1402952671051025, "scale": 23.197338104248047}}, "21": {"attention": {"k_proj": {"bias": 0.07559901475906372, "kernel": 49.98802185058594}, "out_proj": {"bias": 1.2814140319824219, "kernel": 47.44640350341797}, "q_proj": {"bias": 2.729158878326416, "kernel": 50.828125}, "v_proj": {"bias": 0.40822988748550415, "kernel": 46.557373046875}}, "feed_forward": {"intermediate_dense": {"bias": 1.9704020023345947, "kernel": 104.40736389160156}, "output_dense": {"bias": 1.1237080097198486, "kernel": 102.13019561767578}}, "final_layer_norm": {"bias": 2.3562023639678955, "scale": 22.68221092224121}, "layer_norm": {"bias": 2.2092318534851074, "scale": 23.455793380737305}}, "22": {"attention": {"k_proj": {"bias": 0.0718066543340683, "kernel": 50.382484436035156}, "out_proj": {"bias": 1.2022175788879395, "kernel": 46.92112350463867}, "q_proj": {"bias": 2.8022966384887695, "kernel": 50.76280975341797}, "v_proj": {"bias": 0.3634304106235504, "kernel": 46.791908264160156}}, "feed_forward": {"intermediate_dense": {"bias": 1.9008455276489258, "kernel": 104.80474853515625}, "output_dense": {"bias": 1.133584976196289, "kernel": 101.41617584228516}}, "final_layer_norm": {"bias": 2.238583564758301, "scale": 22.203289031982422}, "layer_norm": {"bias": 2.2076501846313477, "scale": 22.47842025756836}}, "23": {"attention": {"k_proj": {"bias": 0.1432589441537857, "kernel": 51.46023178100586}, "out_proj": {"bias": 1.3329226970672607, "kernel": 47.925567626953125}, "q_proj": {"bias": 2.647012710571289, "kernel": 51.571170806884766}, "v_proj": {"bias": 0.5189646482467651, "kernel": 48.566627502441406}}, "feed_forward": {"intermediate_dense": {"bias": 1.8796266317367554, "kernel": 104.60630798339844}, "output_dense": {"bias": 1.111185908317566, "kernel": 102.21832275390625}}, "final_layer_norm": {"bias": 2.498049259185791, "scale": 22.16069793701172}, "layer_norm": {"bias": 2.6903762817382812, "scale": 23.707454681396484}}, "24": {"attention": {"k_proj": {"bias": 0.08844071626663208, "kernel": 49.994483947753906}, "out_proj": {"bias": 1.3890769481658936, "kernel": 49.89257049560547}, "q_proj": {"bias": 2.794116973876953, "kernel": 49.99188995361328}, "v_proj": {"bias": 0.4738750457763672, "kernel": 49.97077178955078}}, "feed_forward": {"intermediate_dense": {"bias": 1.9945287704467773, "kernel": 104.07698059082031}, "output_dense": {"bias": 1.149355173110962, "kernel": 105.09184265136719}}, "final_layer_norm": {"bias": 2.5978496074676514, "scale": 22.21728515625}, "layer_norm": {"bias": 2.4253807067871094, "scale": 23.26264190673828}}, "25": {"attention": {"k_proj": {"bias": 0.08660244196653366, "kernel": 50.500160217285156}, "out_proj": {"bias": 1.1993844509124756, "kernel": 47.8043212890625}, "q_proj": {"bias": 2.877532958984375, "kernel": 50.28739929199219}, "v_proj": {"bias": 0.5545438528060913, "kernel": 48.34324645996094}}, "feed_forward": {"intermediate_dense": {"bias": 1.893567681312561, "kernel": 104.34410095214844}, "output_dense": {"bias": 1.0257065296173096, "kernel": 105.02287292480469}}, "final_layer_norm": {"bias": 2.3044614791870117, "scale": 22.74984359741211}, "layer_norm": {"bias": 2.5752525329589844, "scale": 22.384267807006836}}, "26": {"attention": {"k_proj": {"bias": 0.10195975005626678, "kernel": 50.72240447998047}, "out_proj": {"bias": 1.1352291107177734, "kernel": 48.58570861816406}, "q_proj": {"bias": 2.837502956390381, "kernel": 50.48830795288086}, "v_proj": {"bias": 0.4876839816570282, "kernel": 49.193145751953125}}, "feed_forward": {"intermediate_dense": {"bias": 1.9906024932861328, "kernel": 103.77555847167969}, "output_dense": {"bias": 0.9873849153518677, "kernel": 102.18657684326172}}, "final_layer_norm": {"bias": 1.9404196739196777, "scale": 21.605117797851562}, "layer_norm": {"bias": 2.478506565093994, "scale": 22.824661254882812}}, "27": {"attention": {"k_proj": {"bias": 0.37969350814819336, "kernel": 51.41630554199219}, "out_proj": {"bias": 1.367990255355835, "kernel": 49.914955139160156}, "q_proj": {"bias": 2.6148123741149902, "kernel": 51.28168869018555}, "v_proj": {"bias": 0.5718570947647095, "kernel": 50.37164306640625}}, "feed_forward": {"intermediate_dense": {"bias": 2.15096378326416, "kernel": 102.06016540527344}, "output_dense": {"bias": 0.8653504848480225, "kernel": 101.87760925292969}}, "final_layer_norm": {"bias": 2.2239410877227783, "scale": 20.875133514404297}, "layer_norm": {"bias": 2.5527048110961914, "scale": 23.542659759521484}}, "28": {"attention": {"k_proj": {"bias": 0.41472283005714417, "kernel": 52.33226776123047}, "out_proj": {"bias": 1.3916300535202026, "kernel": 50.67154312133789}, "q_proj": {"bias": 2.7706449031829834, "kernel": 51.977943420410156}, "v_proj": {"bias": 0.45858681201934814, "kernel": 51.003623962402344}}, "feed_forward": {"intermediate_dense": {"bias": 2.0966808795928955, "kernel": 102.07293701171875}, "output_dense": {"bias": 0.7669291496276855, "kernel": 104.06695556640625}}, "final_layer_norm": {"bias": 2.1322994232177734, "scale": 21.200178146362305}, "layer_norm": {"bias": 2.0528011322021484, "scale": 24.387773513793945}}, "29": {"attention": {"k_proj": {"bias": 0.0995764285326004, "kernel": 48.78894805908203}, "out_proj": {"bias": 1.3677316904067993, "kernel": 53.17511749267578}, "q_proj": {"bias": 2.7401580810546875, "kernel": 48.61119842529297}, "v_proj": {"bias": 0.41688603162765503, "kernel": 53.08241271972656}}, "feed_forward": {"intermediate_dense": {"bias": 2.0953128337860107, "kernel": 102.74371337890625}, "output_dense": {"bias": 0.8724071383476257, "kernel": 108.33262634277344}}, "final_layer_norm": {"bias": 2.3651862144470215, "scale": 22.33933448791504}, "layer_norm": {"bias": 2.1517908573150635, "scale": 25.369098663330078}}, "3": {"attention": {"k_proj": {"bias": 0.1391461193561554, "kernel": 50.149742126464844}, "out_proj": {"bias": 1.361741542816162, "kernel": 46.530426025390625}, "q_proj": {"bias": 2.725888252258301, "kernel": 50.37773132324219}, "v_proj": {"bias": 0.29798805713653564, "kernel": 46.92759704589844}}, "feed_forward": {"intermediate_dense": {"bias": 1.6413912773132324, "kernel": 100.0673599243164}, "output_dense": {"bias": 0.6511233448982239, "kernel": 90.23179626464844}}, "final_layer_norm": {"bias": 1.7146446704864502, "scale": 21.102611541748047}, "layer_norm": {"bias": 1.8275432586669922, "scale": 23.560691833496094}}, "30": {"attention": {"k_proj": {"bias": 0.26625239849090576, "kernel": 50.70680236816406}, "out_proj": {"bias": 1.1648390293121338, "kernel": 49.46485900878906}, "q_proj": {"bias": 2.8035969734191895, "kernel": 50.78816223144531}, "v_proj": {"bias": 0.4782582223415375, "kernel": 49.80976867675781}}, "feed_forward": {"intermediate_dense": {"bias": 2.028275966644287, "kernel": 103.24610900878906}, "output_dense": {"bias": 0.825469970703125, "kernel": 107.32813262939453}}, "final_layer_norm": {"bias": 2.1862425804138184, "scale": 23.477405548095703}, "layer_norm": {"bias": 2.2931182384490967, "scale": 25.097976684570312}}, "31": {"attention": {"k_proj": {"bias": 0.3596954345703125, "kernel": 49.22454071044922}, "out_proj": {"bias": 1.0929865837097168, "kernel": 50.31128692626953}, "q_proj": {"bias": 2.5881309509277344, "kernel": 49.32586669921875}, "v_proj": {"bias": 0.5263059139251709, "kernel": 50.44228744506836}}, "feed_forward": {"intermediate_dense": {"bias": 2.1080496311187744, "kernel": 101.92001342773438}, "output_dense": {"bias": 1.0012019872665405, "kernel": 104.70475769042969}}, "final_layer_norm": {"bias": 2.0827205181121826, "scale": 23.364334106445312}, "layer_norm": {"bias": 2.2887473106384277, "scale": 24.86959457397461}}, "32": {"attention": {"k_proj": {"bias": 0.21855297684669495, "kernel": 48.07020950317383}, "out_proj": {"bias": 1.099237322807312, "kernel": 49.491844177246094}, "q_proj": {"bias": 2.8412742614746094, "kernel": 48.05375671386719}, "v_proj": {"bias": 0.39264822006225586, "kernel": 49.77978515625}}, "feed_forward": {"intermediate_dense": {"bias": 2.039989948272705, "kernel": 100.77910614013672}, "output_dense": {"bias": 1.0597267150878906, "kernel": 104.04965209960938}}, "final_layer_norm": {"bias": 2.0426650047302246, "scale": 23.793930053710938}, "layer_norm": {"bias": 2.2444510459899902, "scale": 25.12940216064453}}, "33": {"attention": {"k_proj": {"bias": 0.2213122546672821, "kernel": 48.01631164550781}, "out_proj": {"bias": 1.135854721069336, "kernel": 49.34437942504883}, "q_proj": {"bias": 2.981264114379883, "kernel": 48.00811767578125}, "v_proj": {"bias": 0.42182600498199463, "kernel": 49.611488342285156}}, "feed_forward": {"intermediate_dense": {"bias": 2.0493950843811035, "kernel": 99.14665222167969}, "output_dense": {"bias": 1.0363314151763916, "kernel": 102.80903625488281}}, "final_layer_norm": {"bias": 1.9569058418273926, "scale": 23.55689239501953}, "layer_norm": {"bias": 2.436882972717285, "scale": 25.38343048095703}}, "34": {"attention": {"k_proj": {"bias": 0.2371898591518402, "kernel": 47.21821212768555}, "out_proj": {"bias": 1.3834365606307983, "kernel": 50.825218200683594}, "q_proj": {"bias": 2.872438430786133, "kernel": 47.267547607421875}, "v_proj": {"bias": 0.3901195526123047, "kernel": 50.759151458740234}}, "feed_forward": {"intermediate_dense": {"bias": 2.133671998977661, "kernel": 97.96833038330078}, "output_dense": {"bias": 0.9680823087692261, "kernel": 102.10453796386719}}, "final_layer_norm": {"bias": 1.898040771484375, "scale": 23.20598602294922}, "layer_norm": {"bias": 2.514716386795044, "scale": 25.76838493347168}}, "35": {"attention": {"k_proj": {"bias": 0.36396095156669617, "kernel": 48.94306182861328}, "out_proj": {"bias": 1.3037333488464355, "kernel": 49.6699104309082}, "q_proj": {"bias": 2.611567497253418, "kernel": 49.272621154785156}, "v_proj": {"bias": 0.47717273235321045, "kernel": 49.498260498046875}}, "feed_forward": {"intermediate_dense": {"bias": 2.2121734619140625, "kernel": 96.56979370117188}, "output_dense": {"bias": 0.8638666868209839, "kernel": 100.83570861816406}}, "final_layer_norm": {"bias": 1.973323106765747, "scale": 23.33331298828125}, "layer_norm": {"bias": 2.289663553237915, "scale": 26.261844635009766}}, "36": {"attention": {"k_proj": {"bias": 0.2009318321943283, "kernel": 46.25843048095703}, "out_proj": {"bias": 1.3412315845489502, "kernel": 51.00749588012695}, "q_proj": {"bias": 2.7021098136901855, "kernel": 46.24171447753906}, "v_proj": {"bias": 0.36190247535705566, "kernel": 51.19354248046875}}, "feed_forward": {"intermediate_dense": {"bias": 2.088041305541992, "kernel": 95.65827941894531}, "output_dense": {"bias": 0.8960506916046143, "kernel": 100.51638793945312}}, "final_layer_norm": {"bias": 1.6232030391693115, "scale": 23.855173110961914}, "layer_norm": {"bias": 2.0089900493621826, "scale": 25.782394409179688}}, "37": {"attention": {"k_proj": {"bias": 0.5291366577148438, "kernel": 45.304229736328125}, "out_proj": {"bias": 1.5989298820495605, "kernel": 50.99226379394531}, "q_proj": {"bias": 2.395236015319824, "kernel": 45.37137985229492}, "v_proj": {"bias": 0.3581494688987732, "kernel": 50.86305236816406}}, "feed_forward": {"intermediate_dense": {"bias": 1.983236312866211, "kernel": 94.89313507080078}, "output_dense": {"bias": 0.9043598175048828, "kernel": 100.27244567871094}}, "final_layer_norm": {"bias": 1.4474990367889404, "scale": 24.252254486083984}, "layer_norm": {"bias": 1.9790871143341064, "scale": 25.81557273864746}}, "38": {"attention": {"k_proj": {"bias": 0.6149901747703552, "kernel": 43.495384216308594}, "out_proj": {"bias": 1.2976360321044922, "kernel": 50.47369384765625}, "q_proj": {"bias": 2.327796459197998, "kernel": 43.50128173828125}, "v_proj": {"bias": 0.4169180989265442, "kernel": 50.345977783203125}}, "feed_forward": {"intermediate_dense": {"bias": 1.9261053800582886, "kernel": 92.93551635742188}, "output_dense": {"bias": 0.8913296461105347, "kernel": 98.51844787597656}}, "final_layer_norm": {"bias": 1.4931995868682861, "scale": 24.971830368041992}, "layer_norm": {"bias": 2.1573400497436523, "scale": 26.53833770751953}}, "39": {"attention": {"k_proj": {"bias": 0.6447228193283081, "kernel": 43.26772689819336}, "out_proj": {"bias": 1.5898070335388184, "kernel": 50.3433723449707}, "q_proj": {"bias": 2.1136116981506348, "kernel": 43.651336669921875}, "v_proj": {"bias": 0.3867892622947693, "kernel": 50.01488494873047}}, "feed_forward": {"intermediate_dense": {"bias": 1.9170793294906616, "kernel": 91.245361328125}, "output_dense": {"bias": 0.9705325961112976, "kernel": 98.89170837402344}}, "final_layer_norm": {"bias": 1.6365635395050049, "scale": 25.605688095092773}, "layer_norm": {"bias": 2.1331417560577393, "scale": 27.16867446899414}}, "4": {"attention": {"k_proj": {"bias": 0.15244507789611816, "kernel": 52.71580123901367}, "out_proj": {"bias": 1.5446054935455322, "kernel": 47.943851470947266}, "q_proj": {"bias": 2.5289082527160645, "kernel": 52.898963928222656}, "v_proj": {"bias": 0.34463512897491455, "kernel": 48.29966735839844}}, "feed_forward": {"intermediate_dense": {"bias": 1.6290369033813477, "kernel": 99.65355682373047}, "output_dense": {"bias": 0.8142831921577454, "kernel": 91.45928955078125}}, "final_layer_norm": {"bias": 1.8019299507141113, "scale": 20.63494873046875}, "layer_norm": {"bias": 1.9197347164154053, "scale": 23.959182739257812}}, "40": {"attention": {"k_proj": {"bias": 0.5863298177719116, "kernel": 42.6125602722168}, "out_proj": {"bias": 1.5342514514923096, "kernel": 48.988128662109375}, "q_proj": {"bias": 2.0460410118103027, "kernel": 43.37555694580078}, "v_proj": {"bias": 0.44030535221099854, "kernel": 48.56474304199219}}, "feed_forward": {"intermediate_dense": {"bias": 1.7780985832214355, "kernel": 89.50348663330078}, "output_dense": {"bias": 1.0228041410446167, "kernel": 96.14349365234375}}, "final_layer_norm": {"bias": 1.8028088808059692, "scale": 24.874216079711914}, "layer_norm": {"bias": 2.0758750438690186, "scale": 26.705856323242188}}, "41": {"attention": {"k_proj": {"bias": 1.6716312170028687, "kernel": 39.959205627441406}, "out_proj": {"bias": 1.2966623306274414, "kernel": 50.54540252685547}, "q_proj": {"bias": 1.7283188104629517, "kernel": 40.71592712402344}, "v_proj": {"bias": 0.39489489793777466, "kernel": 49.49800109863281}}, "feed_forward": {"intermediate_dense": {"bias": 1.9181920289993286, "kernel": 86.29412841796875}, "output_dense": {"bias": 1.0486664772033691, "kernel": 95.18574523925781}}, "final_layer_norm": {"bias": 2.298692226409912, "scale": 28.326946258544922}, "layer_norm": {"bias": 2.103621482849121, "scale": 28.495193481445312}}, "42": {"attention": {"k_proj": {"bias": 0.7968032956123352, "kernel": 36.730003356933594}, "out_proj": {"bias": 1.340155839920044, "kernel": 44.78160095214844}, "q_proj": {"bias": 1.5501999855041504, "kernel": 38.07794189453125}, "v_proj": {"bias": 0.5893732309341431, "kernel": 43.12867736816406}}, "feed_forward": {"intermediate_dense": {"bias": 1.6594417095184326, "kernel": 85.28736877441406}, "output_dense": {"bias": 1.100542426109314, "kernel": 93.38209533691406}}, "final_layer_norm": {"bias": 2.0190157890319824, "scale": 29.625246047973633}, "layer_norm": {"bias": 1.570120096206665, "scale": 27.366708755493164}}, "43": {"attention": {"k_proj": {"bias": 1.2106237411499023, "kernel": 33.25742721557617}, "out_proj": {"bias": 1.3313493728637695, "kernel": 41.17510223388672}, "q_proj": {"bias": 1.3641879558563232, "kernel": 34.067970275878906}, "v_proj": {"bias": 0.5160531997680664, "kernel": 39.07243347167969}}, "feed_forward": {"intermediate_dense": {"bias": 1.6928391456604004, "kernel": 84.50951385498047}, "output_dense": {"bias": 0.8673111200332642, "kernel": 91.31503295898438}}, "final_layer_norm": {"bias": 1.9508904218673706, "scale": 31.843149185180664}, "layer_norm": {"bias": 1.6904380321502686, "scale": 25.526779174804688}}, "44": {"attention": {"k_proj": {"bias": 2.4914321899414062, "kernel": 33.84665298461914}, "out_proj": {"bias": 1.0955990552902222, "kernel": 44.89889907836914}, "q_proj": {"bias": 1.2959020137786865, "kernel": 34.21710968017578}, "v_proj": {"bias": 0.37947821617126465, "kernel": 43.99350357055664}}, "feed_forward": {"intermediate_dense": {"bias": 1.7714064121246338, "kernel": 83.4621353149414}, "output_dense": {"bias": 0.8126852512359619, "kernel": 88.96501922607422}}, "final_layer_norm": {"bias": 1.9349080324172974, "scale": 34.0151252746582}, "layer_norm": {"bias": 1.5853739976882935, "scale": 25.543411254882812}}, "45": {"attention": {"k_proj": {"bias": 2.0486693382263184, "kernel": 33.69037628173828}, "out_proj": {"bias": 0.9811146259307861, "kernel": 48.496158599853516}, "q_proj": {"bias": 1.3770025968551636, "kernel": 33.870399475097656}, "v_proj": {"bias": 0.43050453066825867, "kernel": 48.65619659423828}}, "feed_forward": {"intermediate_dense": {"bias": 1.8894906044006348, "kernel": 80.13162231445312}, "output_dense": {"bias": 0.947555422782898, "kernel": 84.36309814453125}}, "final_layer_norm": {"bias": 1.6835851669311523, "scale": 32.72261047363281}, "layer_norm": {"bias": 1.5154564380645752, "scale": 24.061302185058594}}, "46": {"attention": {"k_proj": {"bias": 1.5383987426757812, "kernel": 34.86901092529297}, "out_proj": {"bias": 0.7462900280952454, "kernel": 50.921600341796875}, "q_proj": {"bias": 1.5393162965774536, "kernel": 34.98643112182617}, "v_proj": {"bias": 0.3719345033168793, "kernel": 51.67694854736328}}, "feed_forward": {"intermediate_dense": {"bias": 1.9439404010772705, "kernel": 74.48785400390625}, "output_dense": {"bias": 1.1011567115783691, "kernel": 74.67607879638672}}, "final_layer_norm": {"bias": 1.6659607887268066, "scale": 28.238208770751953}, "layer_norm": {"bias": 1.3336127996444702, "scale": 22.975210189819336}}, "47": {"attention": {"k_proj": {"bias": 0.26446425914764404, "kernel": 37.16825866699219}, "out_proj": {"bias": 0.6312835216522217, "kernel": 45.19498825073242}, "q_proj": {"bias": 1.661026954650879, "kernel": 37.804115295410156}, "v_proj": {"bias": 0.3476800322532654, "kernel": 46.18404769897461}}, "feed_forward": {"intermediate_dense": {"bias": 1.9951825141906738, "kernel": 71.86219787597656}, "output_dense": {"bias": 0.6065097451210022, "kernel": 68.2131118774414}}, "final_layer_norm": {"bias": 1.5113301277160645, "scale": 23.086095809936523}, "layer_norm": {"bias": 1.0591719150543213, "scale": 20.228235244750977}}, "5": {"attention": {"k_proj": {"bias": 0.06965074688196182, "kernel": 48.030853271484375}, "out_proj": {"bias": 1.5286259651184082, "kernel": 49.14524841308594}, "q_proj": {"bias": 2.6216883659362793, "kernel": 48.17328643798828}, "v_proj": {"bias": 0.3084770739078522, "kernel": 49.93551254272461}}, "feed_forward": {"intermediate_dense": {"bias": 1.554388165473938, "kernel": 99.76133728027344}, "output_dense": {"bias": 0.8436105251312256, "kernel": 90.75849914550781}}, "final_layer_norm": {"bias": 2.0791525840759277, "scale": 20.842615127563477}, "layer_norm": {"bias": 1.9531826972961426, "scale": 23.309175491333008}}, "6": {"attention": {"k_proj": {"bias": 0.21384230256080627, "kernel": 49.7178955078125}, "out_proj": {"bias": 1.5188982486724854, "kernel": 48.48472213745117}, "q_proj": {"bias": 2.6720757484436035, "kernel": 50.19920349121094}, "v_proj": {"bias": 0.31003713607788086, "kernel": 49.01288604736328}}, "feed_forward": {"intermediate_dense": {"bias": 1.5325901508331299, "kernel": 98.85545349121094}, "output_dense": {"bias": 0.6974776387214661, "kernel": 90.35247802734375}}, "final_layer_norm": {"bias": 2.3775582313537598, "scale": 20.320110321044922}, "layer_norm": {"bias": 1.9541221857070923, "scale": 23.739662170410156}}, "7": {"attention": {"k_proj": {"bias": 0.20933420956134796, "kernel": 49.48929977416992}, "out_proj": {"bias": 1.3352687358856201, "kernel": 48.73784637451172}, "q_proj": {"bias": 2.4438915252685547, "kernel": 49.88124084472656}, "v_proj": {"bias": 0.39411574602127075, "kernel": 48.703468322753906}}, "feed_forward": {"intermediate_dense": {"bias": 1.5383787155151367, "kernel": 98.59712219238281}, "output_dense": {"bias": 0.5376439094543457, "kernel": 90.091796875}}, "final_layer_norm": {"bias": 2.218599557876587, "scale": 20.553186416625977}, "layer_norm": {"bias": 1.8594433069229126, "scale": 22.461990356445312}}, "8": {"attention": {"k_proj": {"bias": 0.18750429153442383, "kernel": 48.985626220703125}, "out_proj": {"bias": 1.1614996194839478, "kernel": 49.27722930908203}, "q_proj": {"bias": 2.4209980964660645, "kernel": 48.75666427612305}, "v_proj": {"bias": 0.3244088292121887, "kernel": 49.4588508605957}}, "feed_forward": {"intermediate_dense": {"bias": 1.5941160917282104, "kernel": 98.19940185546875}, "output_dense": {"bias": 0.4951457381248474, "kernel": 89.51641845703125}}, "final_layer_norm": {"bias": 2.176692485809326, "scale": 20.33606719970703}, "layer_norm": {"bias": 1.7974094152450562, "scale": 22.91210174560547}}, "9": {"attention": {"k_proj": {"bias": 0.21880677342414856, "kernel": 49.65381622314453}, "out_proj": {"bias": 1.3676939010620117, "kernel": 50.08551788330078}, "q_proj": {"bias": 2.37821102142334, "kernel": 49.81830596923828}, "v_proj": {"bias": 0.3354354500770569, "kernel": 50.494117736816406}}, "feed_forward": {"intermediate_dense": {"bias": 1.6760692596435547, "kernel": 96.8187255859375}, "output_dense": {"bias": 0.634233832359314, "kernel": 90.05500793457031}}, "final_layer_norm": {"bias": 2.0708870887756348, "scale": 19.63043975830078}, "layer_norm": {"bias": 1.8872566223144531, "scale": 24.317705154418945}}}, "pos_conv_embed": {"conv": {"bias": 5.561209678649902, "weight_g": 8.846490859985352, "weight_v": 84.58929443359375}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 2.0290679931640625, "kernel": 20.55536460876465}, "layer_norm": {"bias": 4.550922393798828, "scale": 16.167570114135742}}, "1": {"conv": {"bias": 1.7790228128433228, "kernel": 51.24136734008789}, "layer_norm": {"bias": 5.962646961212158, "scale": 23.268157958984375}}, "2": {"conv": {"bias": 1.140576720237732, "kernel": 46.50312042236328}, "layer_norm": {"bias": 4.176670551300049, "scale": 20.370853424072266}}, "3": {"conv": {"bias": 0.6725863218307495, "kernel": 44.397525787353516}, "layer_norm": {"bias": 3.888174533843994, "scale": 17.53795051574707}}, "4": {"conv": {"bias": 0.6373162269592285, "kernel": 41.314056396484375}, "layer_norm": {"bias": 2.385471820831299, "scale": 16.34571647644043}}, "5": {"conv": {"bias": 0.5147221684455872, "kernel": 37.479759216308594}, "layer_norm": {"bias": 2.020900011062622, "scale": 17.064470291137695}}, "6": {"conv": {"bias": 0.4947893023490906, "kernel": 40.64780044555664}, "layer_norm": {"bias": 0.5876954793930054, "scale": 19.058603286743164}}}}, "feature_projection": {"layer_norm": {"bias": 6.318020820617676, "scale": 16.558513641357422}, "projection": {"bias": 1.6580264568328857, "kernel": 34.771278381347656}}, "masked_spec_embed": 11.914372444152832}}, "train/learning_rate": 7.247499888762832e-05, "train/loss": 0.7996366024017334, "train/param_norm": 1187.21630859375, "_runtime": 16911, "_timestamp": 1659220077, "_step": 2900, "eval/loss": 0.4004497826099396, "eval/wer": 0.322829819462482, "eval/cer": 0.09177810646685618, "eval/step_1k": {"_type": "table-file", "path": "media/table/eval/step_1k_1000_ce5fff7a5f93dcf0e7ad.table.json", "sha256": "ce5fff7a5f93dcf0e7ad34503eff3441fbb5b8dba264f40b90ec55c9f541c9b8", "size": 27574, "artifact_path": "wandb-client-artifact://gbnojfz3g4ziytgztlmj3rztet68prv6za7i9iv5afw7o7p2ubgaq16b8nz4z64t5htgjemcb45hcfit5civ3es5gaj54t6uf5yg5ch9kx5zonikx7tsmwfuciuebrat:latest/eval/step_1k.table.json", "_latest_artifact_path": "wandb-client-artifact://gbnojfz3g4ziytgztlmj3rztet68prv6za7i9iv5afw7o7p2ubgaq16b8nz4z64t5htgjemcb45hcfit5civ3es5gaj54t6uf5yg5ch9kx5zonikx7tsmwfuciuebrat:latest/eval/step_1k.table.json", "ncols": 2, "nrows": 50}, "eval/step_2k": {"_type": "table-file", "path": "media/table/eval/step_2k_2000_a82fd645d93af68dc377.table.json", "sha256": "a82fd645d93af68dc3777f23dc4d2418db2211f60a935f4f10773307f284b5fa", "size": 27421, "artifact_path": "wandb-client-artifact://1lyb8yl2lndykwiz5l6u699ysnzjhc0edoz9dvv6pd1e3u715qj95lrg1ahxaqmc71lqupfoj9t28etsy96b8w3y847ltdlg7cgy1sbqkxri91i3mi1g0odsbcf6r39l:latest/eval/step_2k.table.json", "_latest_artifact_path": "wandb-client-artifact://1lyb8yl2lndykwiz5l6u699ysnzjhc0edoz9dvv6pd1e3u715qj95lrg1ahxaqmc71lqupfoj9t28etsy96b8w3y847ltdlg7cgy1sbqkxri91i3mi1g0odsbcf6r39l:latest/eval/step_2k.table.json", "ncols": 2, "nrows": 50}} \ No newline at end of file +{"train/grad_norm": 4.78125, "layer_grad_norm/": {"lm_head": {"bias": 0.0673828125, "kernel": 1.2578125}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 0.05322265625, "scale": 0.048828125}, "layers": {"0": {"attention": {"k_proj": {"bias": 6.532669067382812e-05, "kernel": 0.05859375}, "out_proj": {"bias": 0.051513671875, "kernel": 0.326171875}, "q_proj": {"bias": 0.008056640625, "kernel": 0.091796875}, "v_proj": {"bias": 0.042236328125, "kernel": 0.34375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0498046875, "kernel": 0.671875}, "output_dense": {"bias": 0.018798828125, "kernel": 0.5546875}}, "final_layer_norm": {"bias": 0.12890625, "scale": 0.1474609375}, "layer_norm": {"bias": 0.07666015625, "scale": 0.13671875}}, "1": {"attention": {"k_proj": {"bias": 4.8160552978515625e-05, "kernel": 0.04833984375}, "out_proj": {"bias": 0.021728515625, "kernel": 0.22265625}, "q_proj": {"bias": 0.00445556640625, "kernel": 0.0478515625}, "v_proj": {"bias": 0.03173828125, "kernel": 0.1787109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.02685546875, "kernel": 0.38671875}, "output_dense": {"bias": 0.019775390625, "kernel": 0.34765625}}, "final_layer_norm": {"bias": 0.04736328125, "scale": 0.035400390625}, "layer_norm": {"bias": 0.048583984375, "scale": 0.05224609375}}, "10": {"attention": {"k_proj": {"bias": 3.266334533691406e-05, "kernel": 0.107421875}, "out_proj": {"bias": 0.016845703125, "kernel": 0.20703125}, "q_proj": {"bias": 0.0064697265625, "kernel": 0.11279296875}, "v_proj": {"bias": 0.0244140625, "kernel": 0.220703125}}, "feed_forward": {"intermediate_dense": {"bias": 0.025634765625, "kernel": 0.40234375}, "output_dense": {"bias": 0.016357421875, "kernel": 0.30859375}}, "final_layer_norm": {"bias": 0.04296875, "scale": 0.03955078125}, "layer_norm": {"bias": 0.045166015625, "scale": 0.04443359375}}, "11": {"attention": {"k_proj": {"bias": 6.67572021484375e-05, "kernel": 0.16015625}, "out_proj": {"bias": 0.0162353515625, "kernel": 0.294921875}, "q_proj": {"bias": 0.0093994140625, "kernel": 0.138671875}, "v_proj": {"bias": 0.02685546875, "kernel": 0.3125}}, "feed_forward": {"intermediate_dense": {"bias": 0.024169921875, "kernel": 0.388671875}, "output_dense": {"bias": 0.015625, "kernel": 0.28125}}, "final_layer_norm": {"bias": 0.03955078125, "scale": 0.0361328125}, "layer_norm": {"bias": 0.0458984375, "scale": 0.0615234375}}, "12": {"attention": {"k_proj": {"bias": 4.291534423828125e-05, "kernel": 0.1142578125}, "out_proj": {"bias": 0.015625, "kernel": 0.23828125}, "q_proj": {"bias": 0.00628662109375, "kernel": 0.1064453125}, "v_proj": {"bias": 0.0233154296875, "kernel": 0.248046875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0244140625, "kernel": 0.3828125}, "output_dense": {"bias": 0.01513671875, "kernel": 0.287109375}}, "final_layer_norm": {"bias": 0.03955078125, "scale": 0.03369140625}, "layer_norm": {"bias": 0.037109375, "scale": 0.0380859375}}, "13": {"attention": {"k_proj": {"bias": 7.390975952148438e-05, "kernel": 0.16015625}, "out_proj": {"bias": 0.015625, "kernel": 0.28515625}, "q_proj": {"bias": 0.0098876953125, "kernel": 0.1474609375}, "v_proj": {"bias": 0.02587890625, "kernel": 0.3125}}, "feed_forward": {"intermediate_dense": {"bias": 0.02392578125, "kernel": 0.365234375}, "output_dense": {"bias": 0.0157470703125, "kernel": 0.287109375}}, "final_layer_norm": {"bias": 0.03857421875, "scale": 0.0419921875}, "layer_norm": {"bias": 0.03857421875, "scale": 0.04931640625}}, "14": {"attention": {"k_proj": {"bias": 0.00010776519775390625, "kernel": 0.10546875}, "out_proj": {"bias": 0.0157470703125, "kernel": 0.25390625}, "q_proj": {"bias": 0.006011962890625, "kernel": 0.09716796875}, "v_proj": {"bias": 0.0245361328125, "kernel": 0.2734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0242919921875, "kernel": 0.37109375}, "output_dense": {"bias": 0.0152587890625, "kernel": 0.296875}}, "final_layer_norm": {"bias": 0.041015625, "scale": 0.03759765625}, "layer_norm": {"bias": 0.035888671875, "scale": 0.034912109375}}, "15": {"attention": {"k_proj": {"bias": 0.0001392364501953125, "kernel": 0.1259765625}, "out_proj": {"bias": 0.015625, "kernel": 0.345703125}, "q_proj": {"bias": 0.006866455078125, "kernel": 0.1142578125}, "v_proj": {"bias": 0.024658203125, "kernel": 0.30078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.022705078125, "kernel": 0.33984375}, "output_dense": {"bias": 0.0155029296875, "kernel": 0.298828125}}, "final_layer_norm": {"bias": 0.036376953125, "scale": 0.030517578125}, "layer_norm": {"bias": 0.035400390625, "scale": 0.02978515625}}, "16": {"attention": {"k_proj": {"bias": 5.555152893066406e-05, "kernel": 0.11669921875}, "out_proj": {"bias": 0.015869140625, "kernel": 0.224609375}, "q_proj": {"bias": 0.00677490234375, "kernel": 0.10546875}, "v_proj": {"bias": 0.023681640625, "kernel": 0.236328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.022216796875, "kernel": 0.33984375}, "output_dense": {"bias": 0.01531982421875, "kernel": 0.298828125}}, "final_layer_norm": {"bias": 0.035400390625, "scale": 0.030029296875}, "layer_norm": {"bias": 0.035888671875, "scale": 0.046142578125}}, "17": {"attention": {"k_proj": {"bias": 5.054473876953125e-05, "kernel": 0.10546875}, "out_proj": {"bias": 0.0166015625, "kernel": 0.1953125}, "q_proj": {"bias": 0.0064697265625, "kernel": 0.10205078125}, "v_proj": {"bias": 0.025390625, "kernel": 0.23046875}}, "feed_forward": {"intermediate_dense": {"bias": 0.02294921875, "kernel": 0.34765625}, "output_dense": {"bias": 0.0162353515625, "kernel": 0.302734375}}, "final_layer_norm": {"bias": 0.037109375, "scale": 0.033447265625}, "layer_norm": {"bias": 0.03857421875, "scale": 0.02783203125}}, "18": {"attention": {"k_proj": {"bias": 7.343292236328125e-05, "kernel": 0.1181640625}, "out_proj": {"bias": 0.0167236328125, "kernel": 0.26171875}, "q_proj": {"bias": 0.00634765625, "kernel": 0.1103515625}, "v_proj": {"bias": 0.02490234375, "kernel": 0.2412109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.022705078125, "kernel": 0.35546875}, "output_dense": {"bias": 0.01611328125, "kernel": 0.32421875}}, "final_layer_norm": {"bias": 0.037353515625, "scale": 0.0322265625}, "layer_norm": {"bias": 0.03662109375, "scale": 0.03515625}}, "19": {"attention": {"k_proj": {"bias": 5.2928924560546875e-05, "kernel": 0.10009765625}, "out_proj": {"bias": 0.0167236328125, "kernel": 0.2109375}, "q_proj": {"bias": 0.00592041015625, "kernel": 0.1025390625}, "v_proj": {"bias": 0.0245361328125, "kernel": 0.2294921875}}, "feed_forward": {"intermediate_dense": {"bias": 0.022705078125, "kernel": 0.375}, "output_dense": {"bias": 0.0162353515625, "kernel": 0.3359375}}, "final_layer_norm": {"bias": 0.0361328125, "scale": 0.033203125}, "layer_norm": {"bias": 0.03662109375, "scale": 0.040283203125}}, "2": {"attention": {"k_proj": {"bias": 4.8160552978515625e-05, "kernel": 0.058349609375}, "out_proj": {"bias": 0.0234375, "kernel": 0.2578125}, "q_proj": {"bias": 0.00494384765625, "kernel": 0.058349609375}, "v_proj": {"bias": 0.0361328125, "kernel": 0.2373046875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0303955078125, "kernel": 0.49609375}, "output_dense": {"bias": 0.021240234375, "kernel": 0.38671875}}, "final_layer_norm": {"bias": 0.052734375, "scale": 0.04150390625}, "layer_norm": {"bias": 0.051025390625, "scale": 0.052978515625}}, "20": {"attention": {"k_proj": {"bias": 1.6927719116210938e-05, "kernel": 0.0654296875}, "out_proj": {"bias": 0.0177001953125, "kernel": 0.12255859375}, "q_proj": {"bias": 0.00396728515625, "kernel": 0.068359375}, "v_proj": {"bias": 0.025634765625, "kernel": 0.1591796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0234375, "kernel": 0.39453125}, "output_dense": {"bias": 0.016845703125, "kernel": 0.328125}}, "final_layer_norm": {"bias": 0.038330078125, "scale": 0.033203125}, "layer_norm": {"bias": 0.0380859375, "scale": 0.0308837890625}}, "21": {"attention": {"k_proj": {"bias": 3.62396240234375e-05, "kernel": 0.08544921875}, "out_proj": {"bias": 0.0174560546875, "kernel": 0.203125}, "q_proj": {"bias": 0.004974365234375, "kernel": 0.087890625}, "v_proj": {"bias": 0.0244140625, "kernel": 0.21484375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0224609375, "kernel": 0.39453125}, "output_dense": {"bias": 0.016845703125, "kernel": 0.33203125}}, "final_layer_norm": {"bias": 0.036865234375, "scale": 0.0322265625}, "layer_norm": {"bias": 0.033203125, "scale": 0.028076171875}}, "22": {"attention": {"k_proj": {"bias": 3.0040740966796875e-05, "kernel": 0.091796875}, "out_proj": {"bias": 0.018310546875, "kernel": 0.1748046875}, "q_proj": {"bias": 0.005126953125, "kernel": 0.095703125}, "v_proj": {"bias": 0.024658203125, "kernel": 0.1923828125}}, "feed_forward": {"intermediate_dense": {"bias": 0.023681640625, "kernel": 0.41015625}, "output_dense": {"bias": 0.0181884765625, "kernel": 0.33203125}}, "final_layer_norm": {"bias": 0.038330078125, "scale": 0.0390625}, "layer_norm": {"bias": 0.0361328125, "scale": 0.03759765625}}, "23": {"attention": {"k_proj": {"bias": 0.0001068115234375, "kernel": 0.1240234375}, "out_proj": {"bias": 0.01953125, "kernel": 0.3359375}, "q_proj": {"bias": 0.00653076171875, "kernel": 0.1220703125}, "v_proj": {"bias": 0.0283203125, "kernel": 0.322265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0235595703125, "kernel": 0.412109375}, "output_dense": {"bias": 0.01904296875, "kernel": 0.328125}}, "final_layer_norm": {"bias": 0.038330078125, "scale": 0.03564453125}, "layer_norm": {"bias": 0.0419921875, "scale": 0.04296875}}, "24": {"attention": {"k_proj": {"bias": 5.626678466796875e-05, "kernel": 0.1279296875}, "out_proj": {"bias": 0.017578125, "kernel": 0.251953125}, "q_proj": {"bias": 0.00762939453125, "kernel": 0.1298828125}, "v_proj": {"bias": 0.0274658203125, "kernel": 0.25390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.022216796875, "kernel": 0.3984375}, "output_dense": {"bias": 0.0169677734375, "kernel": 0.296875}}, "final_layer_norm": {"bias": 0.03662109375, "scale": 0.04345703125}, "layer_norm": {"bias": 0.04541015625, "scale": 0.03564453125}}, "25": {"attention": {"k_proj": {"bias": 6.437301635742188e-05, "kernel": 0.1005859375}, "out_proj": {"bias": 0.0181884765625, "kernel": 0.25390625}, "q_proj": {"bias": 0.005706787109375, "kernel": 0.1025390625}, "v_proj": {"bias": 0.02685546875, "kernel": 0.2734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.022705078125, "kernel": 0.39453125}, "output_dense": {"bias": 0.0179443359375, "kernel": 0.296875}}, "final_layer_norm": {"bias": 0.038330078125, "scale": 0.048583984375}, "layer_norm": {"bias": 0.041259765625, "scale": 0.0380859375}}, "26": {"attention": {"k_proj": {"bias": 4.9114227294921875e-05, "kernel": 0.12109375}, "out_proj": {"bias": 0.0186767578125, "kernel": 0.251953125}, "q_proj": {"bias": 0.00714111328125, "kernel": 0.125}, "v_proj": {"bias": 0.028564453125, "kernel": 0.265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0230712890625, "kernel": 0.375}, "output_dense": {"bias": 0.0189208984375, "kernel": 0.298828125}}, "final_layer_norm": {"bias": 0.038818359375, "scale": 0.0380859375}, "layer_norm": {"bias": 0.04052734375, "scale": 0.0380859375}}, "27": {"attention": {"k_proj": {"bias": 7.2479248046875e-05, "kernel": 0.154296875}, "out_proj": {"bias": 0.0174560546875, "kernel": 0.30078125}, "q_proj": {"bias": 0.00909423828125, "kernel": 0.158203125}, "v_proj": {"bias": 0.0283203125, "kernel": 0.30078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.023681640625, "kernel": 0.375}, "output_dense": {"bias": 0.0179443359375, "kernel": 0.30078125}}, "final_layer_norm": {"bias": 0.041015625, "scale": 0.05029296875}, "layer_norm": {"bias": 0.0458984375, "scale": 0.04638671875}}, "28": {"attention": {"k_proj": {"bias": 7.05718994140625e-05, "kernel": 0.11962890625}, "out_proj": {"bias": 0.016357421875, "kernel": 0.291015625}, "q_proj": {"bias": 0.007110595703125, "kernel": 0.126953125}, "v_proj": {"bias": 0.027099609375, "kernel": 0.294921875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0224609375, "kernel": 0.36328125}, "output_dense": {"bias": 0.016845703125, "kernel": 0.296875}}, "final_layer_norm": {"bias": 0.0380859375, "scale": 0.04345703125}, "layer_norm": {"bias": 0.044189453125, "scale": 0.05517578125}}, "29": {"attention": {"k_proj": {"bias": 4.863739013671875e-05, "kernel": 0.119140625}, "out_proj": {"bias": 0.01556396484375, "kernel": 0.21875}, "q_proj": {"bias": 0.006805419921875, "kernel": 0.125}, "v_proj": {"bias": 0.0247802734375, "kernel": 0.23046875}}, "feed_forward": {"intermediate_dense": {"bias": 0.02099609375, "kernel": 0.3828125}, "output_dense": {"bias": 0.014892578125, "kernel": 0.28515625}}, "final_layer_norm": {"bias": 0.033935546875, "scale": 0.04248046875}, "layer_norm": {"bias": 0.043701171875, "scale": 0.0322265625}}, "3": {"attention": {"k_proj": {"bias": 8.916854858398438e-05, "kernel": 0.107421875}, "out_proj": {"bias": 0.0233154296875, "kernel": 0.37109375}, "q_proj": {"bias": 0.00787353515625, "kernel": 0.1025390625}, "v_proj": {"bias": 0.0361328125, "kernel": 0.328125}}, "feed_forward": {"intermediate_dense": {"bias": 0.030517578125, "kernel": 0.498046875}, "output_dense": {"bias": 0.0211181640625, "kernel": 0.37890625}}, "final_layer_norm": {"bias": 0.0517578125, "scale": 0.0419921875}, "layer_norm": {"bias": 0.052734375, "scale": 0.060546875}}, "30": {"attention": {"k_proj": {"bias": 8.296966552734375e-05, "kernel": 0.1328125}, "out_proj": {"bias": 0.01483154296875, "kernel": 0.2578125}, "q_proj": {"bias": 0.006591796875, "kernel": 0.134765625}, "v_proj": {"bias": 0.02392578125, "kernel": 0.27734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.019775390625, "kernel": 0.365234375}, "output_dense": {"bias": 0.014404296875, "kernel": 0.25390625}}, "final_layer_norm": {"bias": 0.0322265625, "scale": 0.03173828125}, "layer_norm": {"bias": 0.033447265625, "scale": 0.037109375}}, "31": {"attention": {"k_proj": {"bias": 5.626678466796875e-05, "kernel": 0.130859375}, "out_proj": {"bias": 0.013671875, "kernel": 0.2236328125}, "q_proj": {"bias": 0.00653076171875, "kernel": 0.13671875}, "v_proj": {"bias": 0.021484375, "kernel": 0.24609375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0185546875, "kernel": 0.34375}, "output_dense": {"bias": 0.0130615234375, "kernel": 0.234375}}, "final_layer_norm": {"bias": 0.029541015625, "scale": 0.02734375}, "layer_norm": {"bias": 0.03271484375, "scale": 0.027587890625}}, "32": {"attention": {"k_proj": {"bias": 6.437301635742188e-05, "kernel": 0.1015625}, "out_proj": {"bias": 0.0123291015625, "kernel": 0.1923828125}, "q_proj": {"bias": 0.00531005859375, "kernel": 0.10400390625}, "v_proj": {"bias": 0.0186767578125, "kernel": 0.2265625}}, "feed_forward": {"intermediate_dense": {"bias": 0.0169677734375, "kernel": 0.328125}, "output_dense": {"bias": 0.0111083984375, "kernel": 0.2158203125}}, "final_layer_norm": {"bias": 0.028564453125, "scale": 0.0279541015625}, "layer_norm": {"bias": 0.028076171875, "scale": 0.0289306640625}}, "33": {"attention": {"k_proj": {"bias": 8.58306884765625e-05, "kernel": 0.14453125}, "out_proj": {"bias": 0.01055908203125, "kernel": 0.19140625}, "q_proj": {"bias": 0.006927490234375, "kernel": 0.1435546875}, "v_proj": {"bias": 0.0166015625, "kernel": 0.21484375}}, "feed_forward": {"intermediate_dense": {"bias": 0.01458740234375, "kernel": 0.2734375}, "output_dense": {"bias": 0.0096435546875, "kernel": 0.1953125}}, "final_layer_norm": {"bias": 0.025634765625, "scale": 0.030029296875}, "layer_norm": {"bias": 0.025146484375, "scale": 0.03369140625}}, "34": {"attention": {"k_proj": {"bias": 8.392333984375e-05, "kernel": 0.12158203125}, "out_proj": {"bias": 0.00860595703125, "kernel": 0.1796875}, "q_proj": {"bias": 0.00567626953125, "kernel": 0.12060546875}, "v_proj": {"bias": 0.01397705078125, "kernel": 0.189453125}}, "feed_forward": {"intermediate_dense": {"bias": 0.01239013671875, "kernel": 0.232421875}, "output_dense": {"bias": 0.0078125, "kernel": 0.171875}}, "final_layer_norm": {"bias": 0.021240234375, "scale": 0.021484375}, "layer_norm": {"bias": 0.022705078125, "scale": 0.021484375}}, "35": {"attention": {"k_proj": {"bias": 0.00010156631469726562, "kernel": 0.07421875}, "out_proj": {"bias": 0.007293701171875, "kernel": 0.173828125}, "q_proj": {"bias": 0.0032958984375, "kernel": 0.07421875}, "v_proj": {"bias": 0.0101318359375, "kernel": 0.158203125}}, "feed_forward": {"intermediate_dense": {"bias": 0.010009765625, "kernel": 0.1845703125}, "output_dense": {"bias": 0.006866455078125, "kernel": 0.150390625}}, "final_layer_norm": {"bias": 0.016845703125, "scale": 0.01708984375}, "layer_norm": {"bias": 0.0166015625, "scale": 0.020751953125}}, "36": {"attention": {"k_proj": {"bias": 4.6253204345703125e-05, "kernel": 0.072265625}, "out_proj": {"bias": 0.00665283203125, "kernel": 0.1357421875}, "q_proj": {"bias": 0.003265380859375, "kernel": 0.06884765625}, "v_proj": {"bias": 0.00897216796875, "kernel": 0.1240234375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00897216796875, "kernel": 0.166015625}, "output_dense": {"bias": 0.006256103515625, "kernel": 0.1357421875}}, "final_layer_norm": {"bias": 0.01519775390625, "scale": 0.0162353515625}, "layer_norm": {"bias": 0.0150146484375, "scale": 0.0113525390625}}, "37": {"attention": {"k_proj": {"bias": 6.961822509765625e-05, "kernel": 0.0751953125}, "out_proj": {"bias": 0.0059814453125, "kernel": 0.12890625}, "q_proj": {"bias": 0.003326416015625, "kernel": 0.072265625}, "v_proj": {"bias": 0.0087890625, "kernel": 0.1220703125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00811767578125, "kernel": 0.1533203125}, "output_dense": {"bias": 0.00567626953125, "kernel": 0.126953125}}, "final_layer_norm": {"bias": 0.0137939453125, "scale": 0.01513671875}, "layer_norm": {"bias": 0.015625, "scale": 0.019287109375}}, "38": {"attention": {"k_proj": {"bias": 4.76837158203125e-05, "kernel": 0.08447265625}, "out_proj": {"bias": 0.00537109375, "kernel": 0.1171875}, "q_proj": {"bias": 0.0034027099609375, "kernel": 0.078125}, "v_proj": {"bias": 0.007781982421875, "kernel": 0.109375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00738525390625, "kernel": 0.138671875}, "output_dense": {"bias": 0.005126953125, "kernel": 0.1201171875}}, "final_layer_norm": {"bias": 0.01239013671875, "scale": 0.0135498046875}, "layer_norm": {"bias": 0.0140380859375, "scale": 0.01397705078125}}, "39": {"attention": {"k_proj": {"bias": 2.5272369384765625e-05, "kernel": 0.0517578125}, "out_proj": {"bias": 0.004669189453125, "kernel": 0.10693359375}, "q_proj": {"bias": 0.0022430419921875, "kernel": 0.050537109375}, "v_proj": {"bias": 0.00665283203125, "kernel": 0.09716796875}}, "feed_forward": {"intermediate_dense": {"bias": 0.00665283203125, "kernel": 0.12890625}, "output_dense": {"bias": 0.004425048828125, "kernel": 0.1240234375}}, "final_layer_norm": {"bias": 0.0118408203125, "scale": 0.01275634765625}, "layer_norm": {"bias": 0.01171875, "scale": 0.011474609375}}, "4": {"attention": {"k_proj": {"bias": 0.00010204315185546875, "kernel": 0.119140625}, "out_proj": {"bias": 0.0225830078125, "kernel": 0.4140625}, "q_proj": {"bias": 0.007568359375, "kernel": 0.119140625}, "v_proj": {"bias": 0.032470703125, "kernel": 0.375}}, "feed_forward": {"intermediate_dense": {"bias": 0.03076171875, "kernel": 0.478515625}, "output_dense": {"bias": 0.021484375, "kernel": 0.380859375}}, "final_layer_norm": {"bias": 0.05126953125, "scale": 0.047607421875}, "layer_norm": {"bias": 0.04736328125, "scale": 0.062255859375}}, "40": {"attention": {"k_proj": {"bias": 2.0503997802734375e-05, "kernel": 0.045166015625}, "out_proj": {"bias": 0.00433349609375, "kernel": 0.0927734375}, "q_proj": {"bias": 0.0018463134765625, "kernel": 0.04345703125}, "v_proj": {"bias": 0.00579833984375, "kernel": 0.0830078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00628662109375, "kernel": 0.1201171875}, "output_dense": {"bias": 0.0042724609375, "kernel": 0.107421875}}, "final_layer_norm": {"bias": 0.01171875, "scale": 0.0166015625}, "layer_norm": {"bias": 0.0098876953125, "scale": 0.0086669921875}}, "41": {"attention": {"k_proj": {"bias": 1.8596649169921875e-05, "kernel": 0.04296875}, "out_proj": {"bias": 0.00396728515625, "kernel": 0.0966796875}, "q_proj": {"bias": 0.001800537109375, "kernel": 0.04296875}, "v_proj": {"bias": 0.005615234375, "kernel": 0.099609375}}, "feed_forward": {"intermediate_dense": {"bias": 0.00555419921875, "kernel": 0.1171875}, "output_dense": {"bias": 0.00396728515625, "kernel": 0.109375}}, "final_layer_norm": {"bias": 0.01031494140625, "scale": 0.01171875}, "layer_norm": {"bias": 0.01025390625, "scale": 0.010986328125}}, "42": {"attention": {"k_proj": {"bias": 7.510185241699219e-06, "kernel": 0.020751953125}, "out_proj": {"bias": 0.003997802734375, "kernel": 0.06982421875}, "q_proj": {"bias": 0.000926971435546875, "kernel": 0.022705078125}, "v_proj": {"bias": 0.00445556640625, "kernel": 0.0634765625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00494384765625, "kernel": 0.10888671875}, "output_dense": {"bias": 0.004119873046875, "kernel": 0.091796875}}, "final_layer_norm": {"bias": 0.00830078125, "scale": 0.00872802734375}, "layer_norm": {"bias": 0.0069580078125, "scale": 0.0093994140625}}, "43": {"attention": {"k_proj": {"bias": 4.559755325317383e-06, "kernel": 0.0147705078125}, "out_proj": {"bias": 0.004180908203125, "kernel": 0.05126953125}, "q_proj": {"bias": 0.0007781982421875, "kernel": 0.01708984375}, "v_proj": {"bias": 0.0045166015625, "kernel": 0.054931640625}}, "feed_forward": {"intermediate_dense": {"bias": 0.00531005859375, "kernel": 0.12353515625}, "output_dense": {"bias": 0.00445556640625, "kernel": 0.095703125}}, "final_layer_norm": {"bias": 0.0093994140625, "scale": 0.0086669921875}, "layer_norm": {"bias": 0.0079345703125, "scale": 0.00885009765625}}, "44": {"attention": {"k_proj": {"bias": 4.0531158447265625e-06, "kernel": 0.0162353515625}, "out_proj": {"bias": 0.00457763671875, "kernel": 0.05908203125}, "q_proj": {"bias": 0.00081634521484375, "kernel": 0.01806640625}, "v_proj": {"bias": 0.005035400390625, "kernel": 0.06201171875}}, "feed_forward": {"intermediate_dense": {"bias": 0.004852294921875, "kernel": 0.12060546875}, "output_dense": {"bias": 0.004791259765625, "kernel": 0.08642578125}}, "final_layer_norm": {"bias": 0.0078125, "scale": 0.006378173828125}, "layer_norm": {"bias": 0.009765625, "scale": 0.0091552734375}}, "45": {"attention": {"k_proj": {"bias": 3.904104232788086e-06, "kernel": 0.0167236328125}, "out_proj": {"bias": 0.00482177734375, "kernel": 0.062255859375}, "q_proj": {"bias": 0.000873565673828125, "kernel": 0.01806640625}, "v_proj": {"bias": 0.005584716796875, "kernel": 0.0654296875}}, "feed_forward": {"intermediate_dense": {"bias": 0.0045166015625, "kernel": 0.11279296875}, "output_dense": {"bias": 0.005035400390625, "kernel": 0.0830078125}}, "final_layer_norm": {"bias": 0.00732421875, "scale": 0.008056640625}, "layer_norm": {"bias": 0.0126953125, "scale": 0.00860595703125}}, "46": {"attention": {"k_proj": {"bias": 5.8710575103759766e-06, "kernel": 0.019287109375}, "out_proj": {"bias": 0.0050048828125, "kernel": 0.068359375}, "q_proj": {"bias": 0.000926971435546875, "kernel": 0.0184326171875}, "v_proj": {"bias": 0.00616455078125, "kernel": 0.07421875}}, "feed_forward": {"intermediate_dense": {"bias": 0.003875732421875, "kernel": 0.09033203125}, "output_dense": {"bias": 0.00506591796875, "kernel": 0.07958984375}}, "final_layer_norm": {"bias": 0.006866455078125, "scale": 0.0086669921875}, "layer_norm": {"bias": 0.01556396484375, "scale": 0.00885009765625}}, "47": {"attention": {"k_proj": {"bias": 6.794929504394531e-06, "kernel": 0.02099609375}, "out_proj": {"bias": 0.00531005859375, "kernel": 0.06103515625}, "q_proj": {"bias": 0.00107574462890625, "kernel": 0.0174560546875}, "v_proj": {"bias": 0.007415771484375, "kernel": 0.0830078125}}, "feed_forward": {"intermediate_dense": {"bias": 0.00360107421875, "kernel": 0.06787109375}, "output_dense": {"bias": 0.00531005859375, "kernel": 0.07666015625}}, "final_layer_norm": {"bias": 0.006805419921875, "scale": 0.0069580078125}, "layer_norm": {"bias": 0.0201416015625, "scale": 0.0137939453125}}, "5": {"attention": {"k_proj": {"bias": 5.054473876953125e-05, "kernel": 0.10498046875}, "out_proj": {"bias": 0.0228271484375, "kernel": 0.267578125}, "q_proj": {"bias": 0.0064697265625, "kernel": 0.10595703125}, "v_proj": {"bias": 0.03515625, "kernel": 0.27734375}}, "feed_forward": {"intermediate_dense": {"bias": 0.03076171875, "kernel": 0.4609375}, "output_dense": {"bias": 0.022216796875, "kernel": 0.375}}, "final_layer_norm": {"bias": 0.05224609375, "scale": 0.05126953125}, "layer_norm": {"bias": 0.05322265625, "scale": 0.0576171875}}, "6": {"attention": {"k_proj": {"bias": 7.152557373046875e-05, "kernel": 0.1630859375}, "out_proj": {"bias": 0.021484375, "kernel": 0.3671875}, "q_proj": {"bias": 0.0098876953125, "kernel": 0.154296875}, "v_proj": {"bias": 0.03564453125, "kernel": 0.375}}, "feed_forward": {"intermediate_dense": {"bias": 0.0294189453125, "kernel": 0.46875}, "output_dense": {"bias": 0.020751953125, "kernel": 0.353515625}}, "final_layer_norm": {"bias": 0.046875, "scale": 0.053955078125}, "layer_norm": {"bias": 0.055419921875, "scale": 0.08642578125}}, "7": {"attention": {"k_proj": {"bias": 0.00010776519775390625, "kernel": 0.1513671875}, "out_proj": {"bias": 0.02099609375, "kernel": 0.375}, "q_proj": {"bias": 0.0093994140625, "kernel": 0.1455078125}, "v_proj": {"bias": 0.034912109375, "kernel": 0.390625}}, "feed_forward": {"intermediate_dense": {"bias": 0.028564453125, "kernel": 0.46484375}, "output_dense": {"bias": 0.019775390625, "kernel": 0.34765625}}, "final_layer_norm": {"bias": 0.04833984375, "scale": 0.04345703125}, "layer_norm": {"bias": 0.0556640625, "scale": 0.05078125}}, "8": {"attention": {"k_proj": {"bias": 8.726119995117188e-05, "kernel": 0.150390625}, "out_proj": {"bias": 0.0194091796875, "kernel": 0.326171875}, "q_proj": {"bias": 0.009033203125, "kernel": 0.138671875}, "v_proj": {"bias": 0.03125, "kernel": 0.333984375}}, "feed_forward": {"intermediate_dense": {"bias": 0.02880859375, "kernel": 0.462890625}, "output_dense": {"bias": 0.0185546875, "kernel": 0.3359375}}, "final_layer_norm": {"bias": 0.051513671875, "scale": 0.04931640625}, "layer_norm": {"bias": 0.05126953125, "scale": 0.0791015625}}, "9": {"attention": {"k_proj": {"bias": 0.00012683868408203125, "kernel": 0.15625}, "out_proj": {"bias": 0.0174560546875, "kernel": 0.41796875}, "q_proj": {"bias": 0.008544921875, "kernel": 0.1494140625}, "v_proj": {"bias": 0.0283203125, "kernel": 0.419921875}}, "feed_forward": {"intermediate_dense": {"bias": 0.025146484375, "kernel": 0.4140625}, "output_dense": {"bias": 0.016845703125, "kernel": 0.3203125}}, "final_layer_norm": {"bias": 0.041015625, "scale": 0.03759765625}, "layer_norm": {"bias": 0.0478515625, "scale": 0.064453125}}}, "pos_conv_embed": {"conv": {"bias": 0.048583984375, "weight_g": 0.0546875, "weight_v": 0.359375}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "1": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "2": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "3": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "4": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "5": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}, "6": {"conv": {"bias": 0.0, "kernel": 0.0}, "layer_norm": {"bias": 0.0, "scale": 0.0}}}}, "feature_projection": {"layer_norm": {"bias": 0.140625, "scale": 0.2197265625}, "projection": {"bias": 0.07666015625, "kernel": 1.578125}}, "masked_spec_embed": 0.0}}, "layer_param_norm/": {"lm_head": {"bias": 0.04921908676624298, "kernel": 5.413252353668213}, "wav2vec2": {"encoder": {"layer_norm": {"bias": 0.902169942855835, "scale": 22.718978881835938}, "layers": {"0": {"attention": {"k_proj": {"bias": 0.06962637603282928, "kernel": 25.959739685058594}, "out_proj": {"bias": 1.5373609066009521, "kernel": 25.135957717895508}, "q_proj": {"bias": 1.318758249282837, "kernel": 26.255878448486328}, "v_proj": {"bias": 0.34726682305336, "kernel": 25.81730842590332}}, "feed_forward": {"intermediate_dense": {"bias": 1.7692794799804688, "kernel": 95.35393524169922}, "output_dense": {"bias": 1.0224363803863525, "kernel": 91.14517974853516}}, "final_layer_norm": {"bias": 1.2835192680358887, "scale": 19.890865325927734}, "layer_norm": {"bias": 3.273667812347412, "scale": 15.973531723022461}}, "1": {"attention": {"k_proj": {"bias": 0.07245154678821564, "kernel": 40.084129333496094}, "out_proj": {"bias": 1.2936314344406128, "kernel": 41.67698669433594}, "q_proj": {"bias": 2.8694558143615723, "kernel": 39.93986511230469}, "v_proj": {"bias": 0.2807631492614746, "kernel": 40.141090393066406}}, "feed_forward": {"intermediate_dense": {"bias": 1.5952367782592773, "kernel": 93.4329605102539}, "output_dense": {"bias": 0.8014806509017944, "kernel": 84.26225280761719}}, "final_layer_norm": {"bias": 1.1526515483856201, "scale": 18.43830108642578}, "layer_norm": {"bias": 1.7570265531539917, "scale": 19.241287231445312}}, "10": {"attention": {"k_proj": {"bias": 0.08582533895969391, "kernel": 47.338623046875}, "out_proj": {"bias": 1.2253642082214355, "kernel": 50.156185150146484}, "q_proj": {"bias": 2.429431200027466, "kernel": 47.298587799072266}, "v_proj": {"bias": 0.31179967522621155, "kernel": 50.35038375854492}}, "feed_forward": {"intermediate_dense": {"bias": 1.633915662765503, "kernel": 97.74920654296875}, "output_dense": {"bias": 0.5614630579948425, "kernel": 91.66010284423828}}, "final_layer_norm": {"bias": 2.208516836166382, "scale": 20.369182586669922}, "layer_norm": {"bias": 1.6970640420913696, "scale": 22.243541717529297}}, "11": {"attention": {"k_proj": {"bias": 0.12339764833450317, "kernel": 47.152183532714844}, "out_proj": {"bias": 1.0763490200042725, "kernel": 49.36817169189453}, "q_proj": {"bias": 2.4786970615386963, "kernel": 46.87358093261719}, "v_proj": {"bias": 0.3584572374820709, "kernel": 49.90296936035156}}, "feed_forward": {"intermediate_dense": {"bias": 1.6834192276000977, "kernel": 98.49960327148438}, "output_dense": {"bias": 0.544917106628418, "kernel": 93.36177062988281}}, "final_layer_norm": {"bias": 2.1911230087280273, "scale": 20.379417419433594}, "layer_norm": {"bias": 1.675253987312317, "scale": 22.585304260253906}}, "12": {"attention": {"k_proj": {"bias": 0.08773796260356903, "kernel": 47.72996520996094}, "out_proj": {"bias": 1.061636209487915, "kernel": 49.674720764160156}, "q_proj": {"bias": 2.359312057495117, "kernel": 47.480804443359375}, "v_proj": {"bias": 0.34631049633026123, "kernel": 50.0855827331543}}, "feed_forward": {"intermediate_dense": {"bias": 1.7265456914901733, "kernel": 99.33685302734375}, "output_dense": {"bias": 0.5338274240493774, "kernel": 94.90812683105469}}, "final_layer_norm": {"bias": 2.1503143310546875, "scale": 20.338415145874023}, "layer_norm": {"bias": 1.7452950477600098, "scale": 23.114551544189453}}, "13": {"attention": {"k_proj": {"bias": 0.10438726842403412, "kernel": 49.61515808105469}, "out_proj": {"bias": 1.0535494089126587, "kernel": 49.3265380859375}, "q_proj": {"bias": 2.3334383964538574, "kernel": 49.47421646118164}, "v_proj": {"bias": 0.366985559463501, "kernel": 49.502899169921875}}, "feed_forward": {"intermediate_dense": {"bias": 1.7822846174240112, "kernel": 99.95390319824219}, "output_dense": {"bias": 0.5495959520339966, "kernel": 95.34565734863281}}, "final_layer_norm": {"bias": 2.038020610809326, "scale": 20.48843765258789}, "layer_norm": {"bias": 1.8462876081466675, "scale": 23.345609664916992}}, "14": {"attention": {"k_proj": {"bias": 0.1679837554693222, "kernel": 49.79728317260742}, "out_proj": {"bias": 1.214162826538086, "kernel": 47.73725891113281}, "q_proj": {"bias": 2.395333766937256, "kernel": 49.861366271972656}, "v_proj": {"bias": 0.36553746461868286, "kernel": 47.318756103515625}}, "feed_forward": {"intermediate_dense": {"bias": 1.8192400932312012, "kernel": 100.57157135009766}, "output_dense": {"bias": 0.5670008659362793, "kernel": 96.70965576171875}}, "final_layer_norm": {"bias": 2.1648218631744385, "scale": 20.63146209716797}, "layer_norm": {"bias": 1.9780219793319702, "scale": 23.503921508789062}}, "15": {"attention": {"k_proj": {"bias": 0.10829080641269684, "kernel": 49.935768127441406}, "out_proj": {"bias": 1.266237735748291, "kernel": 48.36121368408203}, "q_proj": {"bias": 2.5372462272644043, "kernel": 50.010047912597656}, "v_proj": {"bias": 0.40150129795074463, "kernel": 48.01542663574219}}, "feed_forward": {"intermediate_dense": {"bias": 1.8276853561401367, "kernel": 100.38729858398438}, "output_dense": {"bias": 0.716155469417572, "kernel": 97.40103149414062}}, "final_layer_norm": {"bias": 2.095289468765259, "scale": 20.727493286132812}, "layer_norm": {"bias": 2.2260360717773438, "scale": 23.699451446533203}}, "16": {"attention": {"k_proj": {"bias": 0.07975946366786957, "kernel": 49.828582763671875}, "out_proj": {"bias": 1.1969985961914062, "kernel": 47.813873291015625}, "q_proj": {"bias": 2.623103618621826, "kernel": 49.71464538574219}, "v_proj": {"bias": 0.35514599084854126, "kernel": 47.4999885559082}}, "feed_forward": {"intermediate_dense": {"bias": 1.8235373497009277, "kernel": 101.0421371459961}, "output_dense": {"bias": 0.7403956651687622, "kernel": 98.26219177246094}}, "final_layer_norm": {"bias": 2.1561923027038574, "scale": 21.213504791259766}, "layer_norm": {"bias": 2.15500545501709, "scale": 22.581371307373047}}, "17": {"attention": {"k_proj": {"bias": 0.07650183141231537, "kernel": 50.014312744140625}, "out_proj": {"bias": 1.1387214660644531, "kernel": 47.13007354736328}, "q_proj": {"bias": 2.701202392578125, "kernel": 50.10731506347656}, "v_proj": {"bias": 0.39255812764167786, "kernel": 46.80622100830078}}, "feed_forward": {"intermediate_dense": {"bias": 1.836789846420288, "kernel": 102.122314453125}, "output_dense": {"bias": 0.7562288045883179, "kernel": 98.65126037597656}}, "final_layer_norm": {"bias": 2.2485809326171875, "scale": 21.770286560058594}, "layer_norm": {"bias": 2.0617494583129883, "scale": 22.103214263916016}}, "18": {"attention": {"k_proj": {"bias": 0.1012062206864357, "kernel": 50.335113525390625}, "out_proj": {"bias": 1.2434499263763428, "kernel": 48.15618896484375}, "q_proj": {"bias": 2.5853705406188965, "kernel": 50.72291564941406}, "v_proj": {"bias": 0.42100799083709717, "kernel": 47.68859100341797}}, "feed_forward": {"intermediate_dense": {"bias": 1.8790702819824219, "kernel": 102.42304992675781}, "output_dense": {"bias": 0.8707617521286011, "kernel": 100.30254364013672}}, "final_layer_norm": {"bias": 2.354912757873535, "scale": 21.733640670776367}, "layer_norm": {"bias": 2.2507004737854004, "scale": 23.873477935791016}}, "19": {"attention": {"k_proj": {"bias": 0.07635638117790222, "kernel": 49.57988739013672}, "out_proj": {"bias": 1.2168512344360352, "kernel": 48.036102294921875}, "q_proj": {"bias": 2.8640265464782715, "kernel": 50.00752258300781}, "v_proj": {"bias": 0.3834489583969116, "kernel": 47.286338806152344}}, "feed_forward": {"intermediate_dense": {"bias": 1.930978536605835, "kernel": 102.99739074707031}, "output_dense": {"bias": 0.9363443851470947, "kernel": 101.21075439453125}}, "final_layer_norm": {"bias": 2.305131196975708, "scale": 22.094635009765625}, "layer_norm": {"bias": 2.1665005683898926, "scale": 23.046451568603516}}, "2": {"attention": {"k_proj": {"bias": 0.08556383848190308, "kernel": 46.1557502746582}, "out_proj": {"bias": 1.2100764513015747, "kernel": 43.90522003173828}, "q_proj": {"bias": 3.057514190673828, "kernel": 45.91914367675781}, "v_proj": {"bias": 0.3111116886138916, "kernel": 43.896095275878906}}, "feed_forward": {"intermediate_dense": {"bias": 1.6304028034210205, "kernel": 98.4805679321289}, "output_dense": {"bias": 0.6906487941741943, "kernel": 87.40882873535156}}, "final_layer_norm": {"bias": 1.453829050064087, "scale": 21.0145320892334}, "layer_norm": {"bias": 1.6643424034118652, "scale": 21.979299545288086}}, "20": {"attention": {"k_proj": {"bias": 0.07302132248878479, "kernel": 49.55927276611328}, "out_proj": {"bias": 1.2459075450897217, "kernel": 47.42100524902344}, "q_proj": {"bias": 2.778502941131592, "kernel": 50.34117889404297}, "v_proj": {"bias": 0.3550981879234314, "kernel": 46.327613830566406}}, "feed_forward": {"intermediate_dense": {"bias": 1.9308311939239502, "kernel": 104.25685119628906}, "output_dense": {"bias": 1.0517168045043945, "kernel": 101.8398208618164}}, "final_layer_norm": {"bias": 2.328671932220459, "scale": 23.03919219970703}, "layer_norm": {"bias": 2.140599250793457, "scale": 23.187705993652344}}, "21": {"attention": {"k_proj": {"bias": 0.08338358253240585, "kernel": 49.99740219116211}, "out_proj": {"bias": 1.2818005084991455, "kernel": 47.455970764160156}, "q_proj": {"bias": 2.7295784950256348, "kernel": 50.83648681640625}, "v_proj": {"bias": 0.4071462154388428, "kernel": 46.567935943603516}}, "feed_forward": {"intermediate_dense": {"bias": 1.9719955921173096, "kernel": 104.44792175292969}, "output_dense": {"bias": 1.1241118907928467, "kernel": 102.1684799194336}}, "final_layer_norm": {"bias": 2.3547935485839844, "scale": 22.687164306640625}, "layer_norm": {"bias": 2.2089452743530273, "scale": 23.444211959838867}}, "22": {"attention": {"k_proj": {"bias": 0.08050543069839478, "kernel": 50.391395568847656}, "out_proj": {"bias": 1.202901840209961, "kernel": 46.93544387817383}, "q_proj": {"bias": 2.8019251823425293, "kernel": 50.769989013671875}, "v_proj": {"bias": 0.3627583980560303, "kernel": 46.806793212890625}}, "feed_forward": {"intermediate_dense": {"bias": 1.9032940864562988, "kernel": 104.84452056884766}, "output_dense": {"bias": 1.1339362859725952, "kernel": 101.4557876586914}}, "final_layer_norm": {"bias": 2.2400705814361572, "scale": 22.207096099853516}, "layer_norm": {"bias": 2.207681655883789, "scale": 22.46786880493164}}, "23": {"attention": {"k_proj": {"bias": 0.14860177040100098, "kernel": 51.463966369628906}, "out_proj": {"bias": 1.3338309526443481, "kernel": 47.94309616088867}, "q_proj": {"bias": 2.6465096473693848, "kernel": 51.574920654296875}, "v_proj": {"bias": 0.5188049077987671, "kernel": 48.583961486816406}}, "feed_forward": {"intermediate_dense": {"bias": 1.8811860084533691, "kernel": 104.649658203125}, "output_dense": {"bias": 1.111782193183899, "kernel": 102.2613525390625}}, "final_layer_norm": {"bias": 2.498034715652466, "scale": 22.16802215576172}, "layer_norm": {"bias": 2.689911365509033, "scale": 23.70257568359375}}, "24": {"attention": {"k_proj": {"bias": 0.09581315517425537, "kernel": 50.01167297363281}, "out_proj": {"bias": 1.3900225162506104, "kernel": 49.904876708984375}, "q_proj": {"bias": 2.7940328121185303, "kernel": 50.00927734375}, "v_proj": {"bias": 0.47368431091308594, "kernel": 49.98321533203125}}, "feed_forward": {"intermediate_dense": {"bias": 1.9960441589355469, "kernel": 104.11872863769531}, "output_dense": {"bias": 1.149522066116333, "kernel": 105.13333129882812}}, "final_layer_norm": {"bias": 2.5994277000427246, "scale": 22.22292709350586}, "layer_norm": {"bias": 2.4260878562927246, "scale": 23.260433197021484}}, "25": {"attention": {"k_proj": {"bias": 0.09423495829105377, "kernel": 50.50605010986328}, "out_proj": {"bias": 1.2003486156463623, "kernel": 47.81767272949219}, "q_proj": {"bias": 2.8769443035125732, "kernel": 50.293296813964844}, "v_proj": {"bias": 0.5541197061538696, "kernel": 48.35689163208008}}, "feed_forward": {"intermediate_dense": {"bias": 1.8949710130691528, "kernel": 104.3861083984375}, "output_dense": {"bias": 1.0257925987243652, "kernel": 105.06424713134766}}, "final_layer_norm": {"bias": 2.304706335067749, "scale": 22.754941940307617}, "layer_norm": {"bias": 2.575399160385132, "scale": 22.375782012939453}}, "26": {"attention": {"k_proj": {"bias": 0.11013755202293396, "kernel": 50.73456573486328}, "out_proj": {"bias": 1.1362485885620117, "kernel": 48.601402282714844}, "q_proj": {"bias": 2.8388776779174805, "kernel": 50.500099182128906}, "v_proj": {"bias": 0.4869583547115326, "kernel": 49.20857238769531}}, "feed_forward": {"intermediate_dense": {"bias": 1.9934030771255493, "kernel": 103.81732177734375}, "output_dense": {"bias": 0.9875839948654175, "kernel": 102.2249755859375}}, "final_layer_norm": {"bias": 1.944329857826233, "scale": 21.60901641845703}, "layer_norm": {"bias": 2.480065107345581, "scale": 22.817523956298828}}, "27": {"attention": {"k_proj": {"bias": 0.38318121433258057, "kernel": 51.434200286865234}, "out_proj": {"bias": 1.3689255714416504, "kernel": 49.93168258666992}, "q_proj": {"bias": 2.6149909496307373, "kernel": 51.30052185058594}, "v_proj": {"bias": 0.5731261968612671, "kernel": 50.388309478759766}}, "feed_forward": {"intermediate_dense": {"bias": 2.153050422668457, "kernel": 102.10476684570312}, "output_dense": {"bias": 0.8649272918701172, "kernel": 101.91909790039062}}, "final_layer_norm": {"bias": 2.226088047027588, "scale": 20.881439208984375}, "layer_norm": {"bias": 2.553518772125244, "scale": 23.541961669921875}}, "28": {"attention": {"k_proj": {"bias": 0.4168204069137573, "kernel": 52.348060607910156}, "out_proj": {"bias": 1.3920505046844482, "kernel": 50.6854248046875}, "q_proj": {"bias": 2.772891044616699, "kernel": 51.99403381347656}, "v_proj": {"bias": 0.45851659774780273, "kernel": 51.018001556396484}}, "feed_forward": {"intermediate_dense": {"bias": 2.098593235015869, "kernel": 102.11735534667969}, "output_dense": {"bias": 0.7665820121765137, "kernel": 104.10763549804688}}, "final_layer_norm": {"bias": 2.134143352508545, "scale": 21.207117080688477}, "layer_norm": {"bias": 2.05226469039917, "scale": 24.383384704589844}}, "29": {"attention": {"k_proj": {"bias": 0.10613982379436493, "kernel": 48.80516815185547}, "out_proj": {"bias": 1.3684569597244263, "kernel": 53.18773651123047}, "q_proj": {"bias": 2.739417314529419, "kernel": 48.6279411315918}, "v_proj": {"bias": 0.4170042872428894, "kernel": 53.09568786621094}}, "feed_forward": {"intermediate_dense": {"bias": 2.0967764854431152, "kernel": 102.78724670410156}, "output_dense": {"bias": 0.8726767301559448, "kernel": 108.37275695800781}}, "final_layer_norm": {"bias": 2.365769386291504, "scale": 22.347183227539062}, "layer_norm": {"bias": 2.1533961296081543, "scale": 25.36864471435547}}, "3": {"attention": {"k_proj": {"bias": 0.1442270576953888, "kernel": 50.15776824951172}, "out_proj": {"bias": 1.3631471395492554, "kernel": 46.542762756347656}, "q_proj": {"bias": 2.72804594039917, "kernel": 50.385353088378906}, "v_proj": {"bias": 0.2975858449935913, "kernel": 46.9383544921875}}, "feed_forward": {"intermediate_dense": {"bias": 1.6446895599365234, "kernel": 100.11238098144531}, "output_dense": {"bias": 0.6513658761978149, "kernel": 90.26824951171875}}, "final_layer_norm": {"bias": 1.7181930541992188, "scale": 21.107528686523438}, "layer_norm": {"bias": 1.8293272256851196, "scale": 23.550312042236328}}, "30": {"attention": {"k_proj": {"bias": 0.26914966106414795, "kernel": 50.72053527832031}, "out_proj": {"bias": 1.1657063961029053, "kernel": 49.478797912597656}, "q_proj": {"bias": 2.803964614868164, "kernel": 50.80168151855469}, "v_proj": {"bias": 0.47784772515296936, "kernel": 49.824363708496094}}, "feed_forward": {"intermediate_dense": {"bias": 2.0292789936065674, "kernel": 103.28843688964844}, "output_dense": {"bias": 0.8261035084724426, "kernel": 107.3691635131836}}, "final_layer_norm": {"bias": 2.185847282409668, "scale": 23.4852294921875}, "layer_norm": {"bias": 2.2913970947265625, "scale": 25.094669342041016}}, "31": {"attention": {"k_proj": {"bias": 0.3609456419944763, "kernel": 49.23519515991211}, "out_proj": {"bias": 1.0936834812164307, "kernel": 50.319664001464844}, "q_proj": {"bias": 2.588756799697876, "kernel": 49.33666229248047}, "v_proj": {"bias": 0.5257799625396729, "kernel": 50.451271057128906}}, "feed_forward": {"intermediate_dense": {"bias": 2.108356475830078, "kernel": 101.96115112304688}, "output_dense": {"bias": 1.0010154247283936, "kernel": 104.74467468261719}}, "final_layer_norm": {"bias": 2.0811474323272705, "scale": 23.371543884277344}, "layer_norm": {"bias": 2.287497043609619, "scale": 24.863964080810547}}, "32": {"attention": {"k_proj": {"bias": 0.22056454420089722, "kernel": 48.08464813232422}, "out_proj": {"bias": 1.100999116897583, "kernel": 49.500885009765625}, "q_proj": {"bias": 2.8397715091705322, "kernel": 48.067569732666016}, "v_proj": {"bias": 0.39237940311431885, "kernel": 49.78910827636719}}, "feed_forward": {"intermediate_dense": {"bias": 2.0420303344726562, "kernel": 100.81647491455078}, "output_dense": {"bias": 1.0590380430221558, "kernel": 104.08695983886719}}, "final_layer_norm": {"bias": 2.0450878143310547, "scale": 23.797229766845703}, "layer_norm": {"bias": 2.245396852493286, "scale": 25.125873565673828}}, "33": {"attention": {"k_proj": {"bias": 0.22372880578041077, "kernel": 48.02918243408203}, "out_proj": {"bias": 1.1369173526763916, "kernel": 49.35287094116211}, "q_proj": {"bias": 2.979318618774414, "kernel": 48.02069854736328}, "v_proj": {"bias": 0.42063963413238525, "kernel": 49.619895935058594}}, "feed_forward": {"intermediate_dense": {"bias": 2.0515971183776855, "kernel": 99.18258666992188}, "output_dense": {"bias": 1.0368101596832275, "kernel": 102.84422302246094}}, "final_layer_norm": {"bias": 1.9600660800933838, "scale": 23.559476852416992}, "layer_norm": {"bias": 2.436300277709961, "scale": 25.378341674804688}}, "34": {"attention": {"k_proj": {"bias": 0.23947922885417938, "kernel": 47.22948455810547}, "out_proj": {"bias": 1.38441801071167, "kernel": 50.832462310791016}, "q_proj": {"bias": 2.8738064765930176, "kernel": 47.27733612060547}, "v_proj": {"bias": 0.3886609673500061, "kernel": 50.765830993652344}}, "feed_forward": {"intermediate_dense": {"bias": 2.135632038116455, "kernel": 98.00289916992188}, "output_dense": {"bias": 0.9690021276473999, "kernel": 102.13497161865234}}, "final_layer_norm": {"bias": 1.8987112045288086, "scale": 23.208091735839844}, "layer_norm": {"bias": 2.513838768005371, "scale": 25.764408111572266}}, "35": {"attention": {"k_proj": {"bias": 0.36401885747909546, "kernel": 48.95426940917969}, "out_proj": {"bias": 1.3058595657348633, "kernel": 49.674713134765625}, "q_proj": {"bias": 2.610793113708496, "kernel": 49.28239822387695}, "v_proj": {"bias": 0.4768249988555908, "kernel": 49.50395584106445}}, "feed_forward": {"intermediate_dense": {"bias": 2.214972972869873, "kernel": 96.60084533691406}, "output_dense": {"bias": 0.8646180629730225, "kernel": 100.86213684082031}}, "final_layer_norm": {"bias": 1.9740121364593506, "scale": 23.334243774414062}, "layer_norm": {"bias": 2.292015314102173, "scale": 26.258684158325195}}, "36": {"attention": {"k_proj": {"bias": 0.2038479447364807, "kernel": 46.26953125}, "out_proj": {"bias": 1.3423584699630737, "kernel": 51.01178741455078}, "q_proj": {"bias": 2.702681303024292, "kernel": 46.250885009765625}, "v_proj": {"bias": 0.361502468585968, "kernel": 51.19853210449219}}, "feed_forward": {"intermediate_dense": {"bias": 2.0908637046813965, "kernel": 95.68685913085938}, "output_dense": {"bias": 0.8964763879776001, "kernel": 100.53981018066406}}, "final_layer_norm": {"bias": 1.6254173517227173, "scale": 23.85594940185547}, "layer_norm": {"bias": 2.010197639465332, "scale": 25.782194137573242}}, "37": {"attention": {"k_proj": {"bias": 0.5311571955680847, "kernel": 45.31755828857422}, "out_proj": {"bias": 1.599708080291748, "kernel": 50.996910095214844}, "q_proj": {"bias": 2.3959484100341797, "kernel": 45.383331298828125}, "v_proj": {"bias": 0.35769104957580566, "kernel": 50.8680534362793}}, "feed_forward": {"intermediate_dense": {"bias": 1.9859367609024048, "kernel": 94.91675567626953}, "output_dense": {"bias": 0.9045373201370239, "kernel": 100.29216003417969}}, "final_layer_norm": {"bias": 1.4480338096618652, "scale": 24.25265121459961}, "layer_norm": {"bias": 1.9799282550811768, "scale": 25.815475463867188}}, "38": {"attention": {"k_proj": {"bias": 0.6158674955368042, "kernel": 43.506187438964844}, "out_proj": {"bias": 1.297757625579834, "kernel": 50.477516174316406}, "q_proj": {"bias": 2.327930450439453, "kernel": 43.5106086730957}, "v_proj": {"bias": 0.41636350750923157, "kernel": 50.350196838378906}}, "feed_forward": {"intermediate_dense": {"bias": 1.928141474723816, "kernel": 92.95675659179688}, "output_dense": {"bias": 0.891133189201355, "kernel": 98.53646850585938}}, "final_layer_norm": {"bias": 1.4934136867523193, "scale": 24.972431182861328}, "layer_norm": {"bias": 2.15818452835083, "scale": 26.539714813232422}}, "39": {"attention": {"k_proj": {"bias": 0.6454092860221863, "kernel": 43.2810173034668}, "out_proj": {"bias": 1.5893176794052124, "kernel": 50.346134185791016}, "q_proj": {"bias": 2.114044666290283, "kernel": 43.66300582885742}, "v_proj": {"bias": 0.3863828778266907, "kernel": 50.01773452758789}}, "feed_forward": {"intermediate_dense": {"bias": 1.9186015129089355, "kernel": 91.26253509521484}, "output_dense": {"bias": 0.9703125357627869, "kernel": 98.9065170288086}}, "final_layer_norm": {"bias": 1.6361956596374512, "scale": 25.60651206970215}, "layer_norm": {"bias": 2.1329989433288574, "scale": 27.167842864990234}}, "4": {"attention": {"k_proj": {"bias": 0.15817654132843018, "kernel": 52.728912353515625}, "out_proj": {"bias": 1.5461986064910889, "kernel": 47.960357666015625}, "q_proj": {"bias": 2.528503894805908, "kernel": 52.9129638671875}, "v_proj": {"bias": 0.34476393461227417, "kernel": 48.31501388549805}}, "feed_forward": {"intermediate_dense": {"bias": 1.6319609880447388, "kernel": 99.6986312866211}, "output_dense": {"bias": 0.8142093420028687, "kernel": 91.49694061279297}}, "final_layer_norm": {"bias": 1.8056426048278809, "scale": 20.639925003051758}, "layer_norm": {"bias": 1.9219543933868408, "scale": 23.957796096801758}}, "40": {"attention": {"k_proj": {"bias": 0.5868960022926331, "kernel": 42.622196197509766}, "out_proj": {"bias": 1.5340803861618042, "kernel": 48.98883056640625}, "q_proj": {"bias": 2.0456559658050537, "kernel": 43.38402557373047}, "v_proj": {"bias": 0.44007861614227295, "kernel": 48.56523513793945}}, "feed_forward": {"intermediate_dense": {"bias": 1.7797753810882568, "kernel": 89.51732635498047}, "output_dense": {"bias": 1.022843360900879, "kernel": 96.15615844726562}}, "final_layer_norm": {"bias": 1.8033244609832764, "scale": 24.874635696411133}, "layer_norm": {"bias": 2.076083183288574, "scale": 26.70322608947754}}, "41": {"attention": {"k_proj": {"bias": 1.6710917949676514, "kernel": 39.97071075439453}, "out_proj": {"bias": 1.2964427471160889, "kernel": 50.545196533203125}, "q_proj": {"bias": 1.7293728590011597, "kernel": 40.72488021850586}, "v_proj": {"bias": 0.3948034644126892, "kernel": 49.49766540527344}}, "feed_forward": {"intermediate_dense": {"bias": 1.9193663597106934, "kernel": 86.30714416503906}, "output_dense": {"bias": 1.0492838621139526, "kernel": 95.1964340209961}}, "final_layer_norm": {"bias": 2.298070192337036, "scale": 28.32769012451172}, "layer_norm": {"bias": 2.102863073348999, "scale": 28.49222755432129}}, "42": {"attention": {"k_proj": {"bias": 0.7973123788833618, "kernel": 36.736244201660156}, "out_proj": {"bias": 1.340757131576538, "kernel": 44.781211853027344}, "q_proj": {"bias": 1.5512535572052002, "kernel": 38.0830078125}, "v_proj": {"bias": 0.5902131795883179, "kernel": 43.128143310546875}}, "feed_forward": {"intermediate_dense": {"bias": 1.6618006229400635, "kernel": 85.30094146728516}, "output_dense": {"bias": 1.100982904434204, "kernel": 93.39048767089844}}, "final_layer_norm": {"bias": 2.0181031227111816, "scale": 29.625892639160156}, "layer_norm": {"bias": 1.569514274597168, "scale": 27.363496780395508}}, "43": {"attention": {"k_proj": {"bias": 1.210697889328003, "kernel": 33.26449203491211}, "out_proj": {"bias": 1.3315963745117188, "kernel": 41.17461013793945}, "q_proj": {"bias": 1.365762710571289, "kernel": 34.07386779785156}, "v_proj": {"bias": 0.5160566568374634, "kernel": 39.07225036621094}}, "feed_forward": {"intermediate_dense": {"bias": 1.69490385055542, "kernel": 84.52133178710938}, "output_dense": {"bias": 0.8677738904953003, "kernel": 91.32421875}}, "final_layer_norm": {"bias": 1.9521011114120483, "scale": 31.843727111816406}, "layer_norm": {"bias": 1.690071702003479, "scale": 25.525150299072266}}, "44": {"attention": {"k_proj": {"bias": 2.491560935974121, "kernel": 33.85327911376953}, "out_proj": {"bias": 1.0961835384368896, "kernel": 44.89820098876953}, "q_proj": {"bias": 1.2976069450378418, "kernel": 34.22284698486328}, "v_proj": {"bias": 0.37959951162338257, "kernel": 43.99299621582031}}, "feed_forward": {"intermediate_dense": {"bias": 1.7735038995742798, "kernel": 83.4736099243164}, "output_dense": {"bias": 0.8127532005310059, "kernel": 88.97355651855469}}, "final_layer_norm": {"bias": 1.935543179512024, "scale": 34.01557159423828}, "layer_norm": {"bias": 1.5853848457336426, "scale": 25.541553497314453}}, "45": {"attention": {"k_proj": {"bias": 2.0490241050720215, "kernel": 33.69703674316406}, "out_proj": {"bias": 0.9815094470977783, "kernel": 48.49534606933594}, "q_proj": {"bias": 1.3788188695907593, "kernel": 33.87537384033203}, "v_proj": {"bias": 0.43066418170928955, "kernel": 48.65563201904297}}, "feed_forward": {"intermediate_dense": {"bias": 1.8914554119110107, "kernel": 80.14337921142578}, "output_dense": {"bias": 0.9473761916160583, "kernel": 84.372802734375}}, "final_layer_norm": {"bias": 1.6844468116760254, "scale": 32.72320556640625}, "layer_norm": {"bias": 1.5153565406799316, "scale": 24.05960464477539}}, "46": {"attention": {"k_proj": {"bias": 1.5393372774124146, "kernel": 34.87531661987305}, "out_proj": {"bias": 0.7466506958007812, "kernel": 50.92041778564453}, "q_proj": {"bias": 1.5398750305175781, "kernel": 34.99156951904297}, "v_proj": {"bias": 0.37225598096847534, "kernel": 51.67621612548828}}, "feed_forward": {"intermediate_dense": {"bias": 1.9456932544708252, "kernel": 74.50343322753906}, "output_dense": {"bias": 1.1010990142822266, "kernel": 74.68907165527344}}, "final_layer_norm": {"bias": 1.6661288738250732, "scale": 28.239055633544922}, "layer_norm": {"bias": 1.3337628841400146, "scale": 22.973430633544922}}, "47": {"attention": {"k_proj": {"bias": 0.2661091089248657, "kernel": 37.18138122558594}, "out_proj": {"bias": 0.6315953731536865, "kernel": 45.19499206542969}, "q_proj": {"bias": 1.661759853363037, "kernel": 37.81513977050781}, "v_proj": {"bias": 0.3477732539176941, "kernel": 46.184226989746094}}, "feed_forward": {"intermediate_dense": {"bias": 1.9964029788970947, "kernel": 71.88719177246094}, "output_dense": {"bias": 0.6068949103355408, "kernel": 68.23475646972656}}, "final_layer_norm": {"bias": 1.5102505683898926, "scale": 23.088993072509766}, "layer_norm": {"bias": 1.0595608949661255, "scale": 20.226943969726562}}, "5": {"attention": {"k_proj": {"bias": 0.08035004884004593, "kernel": 48.04241180419922}, "out_proj": {"bias": 1.529860496520996, "kernel": 49.15669631958008}, "q_proj": {"bias": 2.624617099761963, "kernel": 48.18388748168945}, "v_proj": {"bias": 0.30855563282966614, "kernel": 49.94438552856445}}, "feed_forward": {"intermediate_dense": {"bias": 1.5579830408096313, "kernel": 99.80638885498047}, "output_dense": {"bias": 0.843464732170105, "kernel": 90.79640197753906}}, "final_layer_norm": {"bias": 2.082186222076416, "scale": 20.846405029296875}, "layer_norm": {"bias": 1.9546864032745361, "scale": 23.293882369995117}}, "6": {"attention": {"k_proj": {"bias": 0.21805936098098755, "kernel": 49.74000549316406}, "out_proj": {"bias": 1.5195627212524414, "kernel": 48.500099182128906}, "q_proj": {"bias": 2.6734378337860107, "kernel": 50.22113800048828}, "v_proj": {"bias": 0.31005337834358215, "kernel": 49.027130126953125}}, "feed_forward": {"intermediate_dense": {"bias": 1.535608172416687, "kernel": 98.90092468261719}, "output_dense": {"bias": 0.6977838277816772, "kernel": 90.391357421875}}, "final_layer_norm": {"bias": 2.379132032394409, "scale": 20.32402801513672}, "layer_norm": {"bias": 1.9548118114471436, "scale": 23.736774444580078}}, "7": {"attention": {"k_proj": {"bias": 0.21445335447788239, "kernel": 49.50411605834961}, "out_proj": {"bias": 1.336416244506836, "kernel": 48.75510787963867}, "q_proj": {"bias": 2.444244861602783, "kernel": 49.896278381347656}, "v_proj": {"bias": 0.3937806487083435, "kernel": 48.72071075439453}}, "feed_forward": {"intermediate_dense": {"bias": 1.5411996841430664, "kernel": 98.64164733886719}, "output_dense": {"bias": 0.5379694700241089, "kernel": 90.13107299804688}}, "final_layer_norm": {"bias": 2.220088481903076, "scale": 20.556297302246094}, "layer_norm": {"bias": 1.8609169721603394, "scale": 22.460102081298828}}, "8": {"attention": {"k_proj": {"bias": 0.19164219498634338, "kernel": 48.9988899230957}, "out_proj": {"bias": 1.1641738414764404, "kernel": 49.290077209472656}, "q_proj": {"bias": 2.4209132194519043, "kernel": 48.770668029785156}, "v_proj": {"bias": 0.32501697540283203, "kernel": 49.470497131347656}}, "feed_forward": {"intermediate_dense": {"bias": 1.5978820323944092, "kernel": 98.24325561523438}, "output_dense": {"bias": 0.4952393174171448, "kernel": 89.55589294433594}}, "final_layer_norm": {"bias": 2.180640697479248, "scale": 20.338396072387695}, "layer_norm": {"bias": 1.8003993034362793, "scale": 22.905012130737305}}, "9": {"attention": {"k_proj": {"bias": 0.22276516258716583, "kernel": 49.68130874633789}, "out_proj": {"bias": 1.3704197406768799, "kernel": 50.10327911376953}, "q_proj": {"bias": 2.3800644874572754, "kernel": 49.846397399902344}, "v_proj": {"bias": 0.33664780855178833, "kernel": 50.51116180419922}}, "feed_forward": {"intermediate_dense": {"bias": 1.6794133186340332, "kernel": 96.86426544189453}, "output_dense": {"bias": 0.6339425444602966, "kernel": 90.09283447265625}}, "final_layer_norm": {"bias": 2.076014280319214, "scale": 19.634029388427734}, "layer_norm": {"bias": 1.8882291316986084, "scale": 24.322616577148438}}}, "pos_conv_embed": {"conv": {"bias": 5.565891265869141, "weight_g": 8.854345321655273, "weight_v": 84.58731079101562}}}, "feature_extractor": {"conv_layers": {"0": {"conv": {"bias": 2.0290679931640625, "kernel": 20.55536460876465}, "layer_norm": {"bias": 4.550922393798828, "scale": 16.167570114135742}}, "1": {"conv": {"bias": 1.7790228128433228, "kernel": 51.24136734008789}, "layer_norm": {"bias": 5.962646961212158, "scale": 23.268157958984375}}, "2": {"conv": {"bias": 1.140576720237732, "kernel": 46.50312042236328}, "layer_norm": {"bias": 4.176670551300049, "scale": 20.370853424072266}}, "3": {"conv": {"bias": 0.6725863218307495, "kernel": 44.397525787353516}, "layer_norm": {"bias": 3.888174533843994, "scale": 17.53795051574707}}, "4": {"conv": {"bias": 0.6373162269592285, "kernel": 41.314056396484375}, "layer_norm": {"bias": 2.385471820831299, "scale": 16.34571647644043}}, "5": {"conv": {"bias": 0.5147221684455872, "kernel": 37.479759216308594}, "layer_norm": {"bias": 2.020900011062622, "scale": 17.064470291137695}}, "6": {"conv": {"bias": 0.4947893023490906, "kernel": 40.64780044555664}, "layer_norm": {"bias": 0.5876954793930054, "scale": 19.058603286743164}}}}, "feature_projection": {"layer_norm": {"bias": 6.317561149597168, "scale": 16.558958053588867}, "projection": {"bias": 1.6600124835968018, "kernel": 34.79463195800781}}, "masked_spec_embed": 11.914372444152832}}, "train/learning_rate": 7.997499778866768e-05, "train/loss": 0.36255156993865967, "train/param_norm": 1187.5712890625, "_runtime": 19747, "_timestamp": 1659222913, "_step": 3200, "eval/loss": 0.4165652394294739, "eval/wer": 0.3462994456013605, "eval/cer": 0.0990741924291949, "eval/step_1k": {"_type": "table-file", "path": "media/table/eval/step_1k_1000_ce5fff7a5f93dcf0e7ad.table.json", "sha256": "ce5fff7a5f93dcf0e7ad34503eff3441fbb5b8dba264f40b90ec55c9f541c9b8", "size": 27574, "artifact_path": "wandb-client-artifact://gbnojfz3g4ziytgztlmj3rztet68prv6za7i9iv5afw7o7p2ubgaq16b8nz4z64t5htgjemcb45hcfit5civ3es5gaj54t6uf5yg5ch9kx5zonikx7tsmwfuciuebrat:latest/eval/step_1k.table.json", "_latest_artifact_path": "wandb-client-artifact://gbnojfz3g4ziytgztlmj3rztet68prv6za7i9iv5afw7o7p2ubgaq16b8nz4z64t5htgjemcb45hcfit5civ3es5gaj54t6uf5yg5ch9kx5zonikx7tsmwfuciuebrat:latest/eval/step_1k.table.json", "ncols": 2, "nrows": 50}, "eval/step_2k": {"_type": "table-file", "path": "media/table/eval/step_2k_2000_a82fd645d93af68dc377.table.json", "sha256": "a82fd645d93af68dc3777f23dc4d2418db2211f60a935f4f10773307f284b5fa", "size": 27421, "artifact_path": "wandb-client-artifact://1lyb8yl2lndykwiz5l6u699ysnzjhc0edoz9dvv6pd1e3u715qj95lrg1ahxaqmc71lqupfoj9t28etsy96b8w3y847ltdlg7cgy1sbqkxri91i3mi1g0odsbcf6r39l:latest/eval/step_2k.table.json", "_latest_artifact_path": "wandb-client-artifact://1lyb8yl2lndykwiz5l6u699ysnzjhc0edoz9dvv6pd1e3u715qj95lrg1ahxaqmc71lqupfoj9t28etsy96b8w3y847ltdlg7cgy1sbqkxri91i3mi1g0odsbcf6r39l:latest/eval/step_2k.table.json", "ncols": 2, "nrows": 50}, "eval/step_3k": {"_type": "table-file", "path": "media/table/eval/step_3k_3000_7f6d2aff7a45425d179c.table.json", "sha256": "7f6d2aff7a45425d179c71f716b09561bb2e472773f02ae271ef0e4c0169de0f", "size": 27499, "artifact_path": "wandb-client-artifact://jt7jkmhqa68ptaphyatg6dq0rfnqfvpoj4w0zo865n3ihhds2qmyzqkexwq3yi51dyi8wwuukvlly18wybh877clpyus64qrxm0kcdpoeiaxjosg7o5bhhlcoexs8dxh:latest/eval/step_3k.table.json", "_latest_artifact_path": "wandb-client-artifact://jt7jkmhqa68ptaphyatg6dq0rfnqfvpoj4w0zo865n3ihhds2qmyzqkexwq3yi51dyi8wwuukvlly18wybh877clpyus64qrxm0kcdpoeiaxjosg7o5bhhlcoexs8dxh:latest/eval/step_3k.table.json", "ncols": 2, "nrows": 50}, "_wandb": {"runtime": 19748}} \ No newline at end of file