0 |
bert.embeddings.word_embeddings |
Embedding |
weight |
[30522, 768] |
23440896 |
23440896 |
0 |
1 |
bert.embeddings.position_embeddings |
Embedding |
weight |
[512, 768] |
393216 |
393216 |
0 |
2 |
bert.embeddings.token_type_embeddings |
Embedding |
weight |
[2, 768] |
1536 |
1536 |
0 |
3 |
bert.embeddings.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
4 |
bert.embeddings.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
5 |
bert.encoder.layer.0.attention.self.query |
Linear |
weight |
[320, 768] |
245760 |
135168 |
0.45 |
6 |
bert.encoder.layer.0.attention.self.query |
Linear |
bias |
[320] |
320 |
256 |
0.2 |
7 |
bert.encoder.layer.0.attention.self.key |
Linear |
weight |
[320, 768] |
245760 |
149504 |
0.391667 |
8 |
bert.encoder.layer.0.attention.self.key |
Linear |
bias |
[320] |
320 |
256 |
0.2 |
9 |
bert.encoder.layer.0.attention.self.value |
Linear |
weight |
[320, 768] |
245760 |
173056 |
0.295833 |
10 |
bert.encoder.layer.0.attention.self.value |
Linear |
bias |
[320] |
320 |
256 |
0.2 |
11 |
bert.encoder.layer.0.attention.output.dense |
Linear |
weight |
[768, 320] |
245760 |
181248 |
0.2625 |
12 |
bert.encoder.layer.0.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
13 |
bert.encoder.layer.0.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
14 |
bert.encoder.layer.0.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
15 |
bert.encoder.layer.0.intermediate.dense |
Linear |
weight |
[185, 768] |
142080 |
142080 |
0 |
16 |
bert.encoder.layer.0.intermediate.dense |
Linear |
bias |
[185] |
185 |
185 |
0 |
17 |
bert.encoder.layer.0.output.dense |
Linear |
weight |
[768, 185] |
142080 |
142080 |
0 |
18 |
bert.encoder.layer.0.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
19 |
bert.encoder.layer.0.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
20 |
bert.encoder.layer.0.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
21 |
bert.encoder.layer.1.attention.self.query |
Linear |
weight |
[320, 768] |
245760 |
175104 |
0.2875 |
22 |
bert.encoder.layer.1.attention.self.query |
Linear |
bias |
[320] |
320 |
288 |
0.1 |
23 |
bert.encoder.layer.1.attention.self.key |
Linear |
weight |
[320, 768] |
245760 |
177152 |
0.279167 |
24 |
bert.encoder.layer.1.attention.self.key |
Linear |
bias |
[320] |
320 |
288 |
0.1 |
25 |
bert.encoder.layer.1.attention.self.value |
Linear |
weight |
[320, 768] |
245760 |
166912 |
0.320833 |
26 |
bert.encoder.layer.1.attention.self.value |
Linear |
bias |
[320] |
320 |
288 |
0.1 |
27 |
bert.encoder.layer.1.attention.output.dense |
Linear |
weight |
[768, 320] |
245760 |
167936 |
0.316667 |
28 |
bert.encoder.layer.1.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
29 |
bert.encoder.layer.1.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
30 |
bert.encoder.layer.1.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
31 |
bert.encoder.layer.1.intermediate.dense |
Linear |
weight |
[315, 768] |
241920 |
241920 |
0 |
32 |
bert.encoder.layer.1.intermediate.dense |
Linear |
bias |
[315] |
315 |
315 |
0 |
33 |
bert.encoder.layer.1.output.dense |
Linear |
weight |
[768, 315] |
241920 |
241920 |
0 |
34 |
bert.encoder.layer.1.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
35 |
bert.encoder.layer.1.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
36 |
bert.encoder.layer.1.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
37 |
bert.encoder.layer.2.attention.self.query |
Linear |
weight |
[576, 768] |
442368 |
285696 |
0.354167 |
38 |
bert.encoder.layer.2.attention.self.query |
Linear |
bias |
[576] |
576 |
480 |
0.166667 |
39 |
bert.encoder.layer.2.attention.self.key |
Linear |
weight |
[576, 768] |
442368 |
297984 |
0.326389 |
40 |
bert.encoder.layer.2.attention.self.key |
Linear |
bias |
[576] |
576 |
480 |
0.166667 |
41 |
bert.encoder.layer.2.attention.self.value |
Linear |
weight |
[576, 768] |
442368 |
226304 |
0.488426 |
42 |
bert.encoder.layer.2.attention.self.value |
Linear |
bias |
[576] |
576 |
384 |
0.333333 |
43 |
bert.encoder.layer.2.attention.output.dense |
Linear |
weight |
[768, 576] |
442368 |
237568 |
0.462963 |
44 |
bert.encoder.layer.2.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
45 |
bert.encoder.layer.2.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
46 |
bert.encoder.layer.2.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
47 |
bert.encoder.layer.2.intermediate.dense |
Linear |
weight |
[339, 768] |
260352 |
260352 |
0 |
48 |
bert.encoder.layer.2.intermediate.dense |
Linear |
bias |
[339] |
339 |
339 |
0 |
49 |
bert.encoder.layer.2.output.dense |
Linear |
weight |
[768, 339] |
260352 |
260352 |
0 |
50 |
bert.encoder.layer.2.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
51 |
bert.encoder.layer.2.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
52 |
bert.encoder.layer.2.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
53 |
bert.encoder.layer.3.attention.self.query |
Linear |
weight |
[576, 768] |
442368 |
277504 |
0.372685 |
54 |
bert.encoder.layer.3.attention.self.query |
Linear |
bias |
[576] |
576 |
512 |
0.111111 |
55 |
bert.encoder.layer.3.attention.self.key |
Linear |
weight |
[576, 768] |
442368 |
303104 |
0.314815 |
56 |
bert.encoder.layer.3.attention.self.key |
Linear |
bias |
[576] |
576 |
512 |
0.111111 |
57 |
bert.encoder.layer.3.attention.self.value |
Linear |
weight |
[576, 768] |
442368 |
297984 |
0.326389 |
58 |
bert.encoder.layer.3.attention.self.value |
Linear |
bias |
[576] |
576 |
512 |
0.111111 |
59 |
bert.encoder.layer.3.attention.output.dense |
Linear |
weight |
[768, 576] |
442368 |
308224 |
0.303241 |
60 |
bert.encoder.layer.3.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
61 |
bert.encoder.layer.3.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
62 |
bert.encoder.layer.3.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
63 |
bert.encoder.layer.3.intermediate.dense |
Linear |
weight |
[368, 768] |
282624 |
282624 |
0 |
64 |
bert.encoder.layer.3.intermediate.dense |
Linear |
bias |
[368] |
368 |
368 |
0 |
65 |
bert.encoder.layer.3.output.dense |
Linear |
weight |
[768, 368] |
282624 |
282624 |
0 |
66 |
bert.encoder.layer.3.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
67 |
bert.encoder.layer.3.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
68 |
bert.encoder.layer.3.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
69 |
bert.encoder.layer.4.attention.self.query |
Linear |
weight |
[576, 768] |
442368 |
291840 |
0.340278 |
70 |
bert.encoder.layer.4.attention.self.query |
Linear |
bias |
[576] |
576 |
544 |
0.0555555 |
71 |
bert.encoder.layer.4.attention.self.key |
Linear |
weight |
[576, 768] |
442368 |
310272 |
0.298611 |
72 |
bert.encoder.layer.4.attention.self.key |
Linear |
bias |
[576] |
576 |
544 |
0.0555555 |
73 |
bert.encoder.layer.4.attention.self.value |
Linear |
weight |
[576, 768] |
442368 |
272384 |
0.384259 |
74 |
bert.encoder.layer.4.attention.self.value |
Linear |
bias |
[576] |
576 |
480 |
0.166667 |
75 |
bert.encoder.layer.4.attention.output.dense |
Linear |
weight |
[768, 576] |
442368 |
263168 |
0.405093 |
76 |
bert.encoder.layer.4.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
77 |
bert.encoder.layer.4.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
78 |
bert.encoder.layer.4.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
79 |
bert.encoder.layer.4.intermediate.dense |
Linear |
weight |
[386, 768] |
296448 |
296448 |
0 |
80 |
bert.encoder.layer.4.intermediate.dense |
Linear |
bias |
[386] |
386 |
386 |
0 |
81 |
bert.encoder.layer.4.output.dense |
Linear |
weight |
[768, 386] |
296448 |
296448 |
0 |
82 |
bert.encoder.layer.4.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
83 |
bert.encoder.layer.4.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
84 |
bert.encoder.layer.4.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
85 |
bert.encoder.layer.5.attention.self.query |
Linear |
weight |
[384, 768] |
294912 |
171008 |
0.420139 |
86 |
bert.encoder.layer.5.attention.self.query |
Linear |
bias |
[384] |
384 |
352 |
0.0833333 |
87 |
bert.encoder.layer.5.attention.self.key |
Linear |
weight |
[384, 768] |
294912 |
205824 |
0.302083 |
88 |
bert.encoder.layer.5.attention.self.key |
Linear |
bias |
[384] |
384 |
352 |
0.0833333 |
89 |
bert.encoder.layer.5.attention.self.value |
Linear |
weight |
[384, 768] |
294912 |
217088 |
0.263889 |
90 |
bert.encoder.layer.5.attention.self.value |
Linear |
bias |
[384] |
384 |
384 |
0 |
91 |
bert.encoder.layer.5.attention.output.dense |
Linear |
weight |
[768, 384] |
294912 |
223232 |
0.243056 |
92 |
bert.encoder.layer.5.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
93 |
bert.encoder.layer.5.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
94 |
bert.encoder.layer.5.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
95 |
bert.encoder.layer.5.intermediate.dense |
Linear |
weight |
[336, 768] |
258048 |
258048 |
0 |
96 |
bert.encoder.layer.5.intermediate.dense |
Linear |
bias |
[336] |
336 |
336 |
0 |
97 |
bert.encoder.layer.5.output.dense |
Linear |
weight |
[768, 336] |
258048 |
258048 |
0 |
98 |
bert.encoder.layer.5.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
99 |
bert.encoder.layer.5.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
100 |
bert.encoder.layer.5.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
101 |
bert.encoder.layer.6.attention.self.query |
Linear |
weight |
[448, 768] |
344064 |
192512 |
0.440476 |
102 |
bert.encoder.layer.6.attention.self.query |
Linear |
bias |
[448] |
448 |
416 |
0.0714285 |
103 |
bert.encoder.layer.6.attention.self.key |
Linear |
weight |
[448, 768] |
344064 |
224256 |
0.348214 |
104 |
bert.encoder.layer.6.attention.self.key |
Linear |
bias |
[448] |
448 |
416 |
0.0714285 |
105 |
bert.encoder.layer.6.attention.self.value |
Linear |
weight |
[448, 768] |
344064 |
209920 |
0.389881 |
106 |
bert.encoder.layer.6.attention.self.value |
Linear |
bias |
[448] |
448 |
352 |
0.214286 |
107 |
bert.encoder.layer.6.attention.output.dense |
Linear |
weight |
[768, 448] |
344064 |
199680 |
0.419643 |
108 |
bert.encoder.layer.6.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
109 |
bert.encoder.layer.6.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
110 |
bert.encoder.layer.6.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
111 |
bert.encoder.layer.6.intermediate.dense |
Linear |
weight |
[280, 768] |
215040 |
215040 |
0 |
112 |
bert.encoder.layer.6.intermediate.dense |
Linear |
bias |
[280] |
280 |
280 |
0 |
113 |
bert.encoder.layer.6.output.dense |
Linear |
weight |
[768, 280] |
215040 |
215040 |
0 |
114 |
bert.encoder.layer.6.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
115 |
bert.encoder.layer.6.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
116 |
bert.encoder.layer.6.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
117 |
bert.encoder.layer.7.attention.self.query |
Linear |
weight |
[448, 768] |
344064 |
201728 |
0.41369 |
118 |
bert.encoder.layer.7.attention.self.query |
Linear |
bias |
[448] |
448 |
416 |
0.0714285 |
119 |
bert.encoder.layer.7.attention.self.key |
Linear |
weight |
[448, 768] |
344064 |
237568 |
0.309524 |
120 |
bert.encoder.layer.7.attention.self.key |
Linear |
bias |
[448] |
448 |
416 |
0.0714285 |
121 |
bert.encoder.layer.7.attention.self.value |
Linear |
weight |
[448, 768] |
344064 |
218112 |
0.366071 |
122 |
bert.encoder.layer.7.attention.self.value |
Linear |
bias |
[448] |
448 |
352 |
0.214286 |
123 |
bert.encoder.layer.7.attention.output.dense |
Linear |
weight |
[768, 448] |
344064 |
202752 |
0.410714 |
124 |
bert.encoder.layer.7.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
125 |
bert.encoder.layer.7.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
126 |
bert.encoder.layer.7.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
127 |
bert.encoder.layer.7.intermediate.dense |
Linear |
weight |
[211, 768] |
162048 |
162048 |
0 |
128 |
bert.encoder.layer.7.intermediate.dense |
Linear |
bias |
[211] |
211 |
211 |
0 |
129 |
bert.encoder.layer.7.output.dense |
Linear |
weight |
[768, 211] |
162048 |
162048 |
0 |
130 |
bert.encoder.layer.7.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
131 |
bert.encoder.layer.7.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
132 |
bert.encoder.layer.7.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
133 |
bert.encoder.layer.8.attention.self.query |
Linear |
weight |
[448, 768] |
344064 |
186368 |
0.458333 |
134 |
bert.encoder.layer.8.attention.self.query |
Linear |
bias |
[448] |
448 |
416 |
0.0714285 |
135 |
bert.encoder.layer.8.attention.self.key |
Linear |
weight |
[448, 768] |
344064 |
197632 |
0.425595 |
136 |
bert.encoder.layer.8.attention.self.key |
Linear |
bias |
[448] |
448 |
416 |
0.0714285 |
137 |
bert.encoder.layer.8.attention.self.value |
Linear |
weight |
[448, 768] |
344064 |
154624 |
0.550595 |
138 |
bert.encoder.layer.8.attention.self.value |
Linear |
bias |
[448] |
448 |
288 |
0.357143 |
139 |
bert.encoder.layer.8.attention.output.dense |
Linear |
weight |
[768, 448] |
344064 |
148480 |
0.568452 |
140 |
bert.encoder.layer.8.attention.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
141 |
bert.encoder.layer.8.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
142 |
bert.encoder.layer.8.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
143 |
bert.encoder.layer.8.intermediate.dense |
Linear |
weight |
[108, 768] |
82944 |
82944 |
0 |
144 |
bert.encoder.layer.8.intermediate.dense |
Linear |
bias |
[108] |
108 |
108 |
0 |
145 |
bert.encoder.layer.8.output.dense |
Linear |
weight |
[768, 108] |
82944 |
82944 |
0 |
146 |
bert.encoder.layer.8.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
147 |
bert.encoder.layer.8.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
148 |
bert.encoder.layer.8.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
149 |
bert.encoder.layer.9.attention.self.query |
Linear |
weight |
[320, 768] |
245760 |
144384 |
0.4125 |
150 |
bert.encoder.layer.9.attention.self.query |
Linear |
bias |
[320] |
320 |
288 |
0.1 |
151 |
bert.encoder.layer.9.attention.self.key |
Linear |
weight |
[320, 768] |
245760 |
155648 |
0.366667 |
152 |
bert.encoder.layer.9.attention.self.key |
Linear |
bias |
[320] |
320 |
288 |
0.1 |
153 |
bert.encoder.layer.9.attention.self.value |
Linear |
weight |
[320, 768] |
245760 |
63488 |
0.741667 |
154 |
bert.encoder.layer.9.attention.self.value |
Linear |
bias |
[320] |
320 |
160 |
0.5 |
155 |
bert.encoder.layer.9.attention.output.dense |
Linear |
weight |
[768, 320] |
245760 |
65536 |
0.733333 |
156 |
bert.encoder.layer.9.attention.output.dense |
Linear |
bias |
[768] |
768 |
704 |
0.0833333 |
157 |
bert.encoder.layer.9.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
158 |
bert.encoder.layer.9.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
159 |
bert.encoder.layer.9.intermediate.dense |
Linear |
weight |
[53, 768] |
40704 |
40704 |
5.96046e-08 |
160 |
bert.encoder.layer.9.intermediate.dense |
Linear |
bias |
[53] |
53 |
53 |
0 |
161 |
bert.encoder.layer.9.output.dense |
Linear |
weight |
[768, 53] |
40704 |
40704 |
5.96046e-08 |
162 |
bert.encoder.layer.9.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
163 |
bert.encoder.layer.9.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
164 |
bert.encoder.layer.9.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
165 |
bert.encoder.layer.10.attention.self.query |
Linear |
weight |
[384, 768] |
294912 |
158720 |
0.461806 |
166 |
bert.encoder.layer.10.attention.self.query |
Linear |
bias |
[384] |
384 |
320 |
0.166667 |
167 |
bert.encoder.layer.10.attention.self.key |
Linear |
weight |
[384, 768] |
294912 |
158720 |
0.461806 |
168 |
bert.encoder.layer.10.attention.self.key |
Linear |
bias |
[384] |
384 |
320 |
0.166667 |
169 |
bert.encoder.layer.10.attention.self.value |
Linear |
weight |
[384, 768] |
294912 |
77824 |
0.736111 |
170 |
bert.encoder.layer.10.attention.self.value |
Linear |
bias |
[384] |
384 |
192 |
0.5 |
171 |
bert.encoder.layer.10.attention.output.dense |
Linear |
weight |
[768, 384] |
294912 |
78848 |
0.732639 |
172 |
bert.encoder.layer.10.attention.output.dense |
Linear |
bias |
[768] |
768 |
736 |
0.0416666 |
173 |
bert.encoder.layer.10.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
174 |
bert.encoder.layer.10.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
175 |
bert.encoder.layer.10.intermediate.dense |
Linear |
weight |
[86, 768] |
66048 |
66048 |
0 |
176 |
bert.encoder.layer.10.intermediate.dense |
Linear |
bias |
[86] |
86 |
86 |
0 |
177 |
bert.encoder.layer.10.output.dense |
Linear |
weight |
[768, 86] |
66048 |
66048 |
0 |
178 |
bert.encoder.layer.10.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
179 |
bert.encoder.layer.10.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
180 |
bert.encoder.layer.10.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
181 |
bert.encoder.layer.11.attention.self.query |
Linear |
weight |
[384, 768] |
294912 |
107520 |
0.635417 |
182 |
bert.encoder.layer.11.attention.self.query |
Linear |
bias |
[384] |
384 |
256 |
0.333333 |
183 |
bert.encoder.layer.11.attention.self.key |
Linear |
weight |
[384, 768] |
294912 |
118784 |
0.597222 |
184 |
bert.encoder.layer.11.attention.self.key |
Linear |
bias |
[384] |
384 |
256 |
0.333333 |
185 |
bert.encoder.layer.11.attention.self.value |
Linear |
weight |
[384, 768] |
294912 |
62464 |
0.788194 |
186 |
bert.encoder.layer.11.attention.self.value |
Linear |
bias |
[384] |
384 |
192 |
0.5 |
187 |
bert.encoder.layer.11.attention.output.dense |
Linear |
weight |
[768, 384] |
294912 |
54272 |
0.815972 |
188 |
bert.encoder.layer.11.attention.output.dense |
Linear |
bias |
[768] |
768 |
672 |
0.125 |
189 |
bert.encoder.layer.11.attention.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
190 |
bert.encoder.layer.11.attention.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
191 |
bert.encoder.layer.11.intermediate.dense |
Linear |
weight |
[105, 768] |
80640 |
80640 |
0 |
192 |
bert.encoder.layer.11.intermediate.dense |
Linear |
bias |
[105] |
105 |
105 |
0 |
193 |
bert.encoder.layer.11.output.dense |
Linear |
weight |
[768, 105] |
80640 |
80640 |
0 |
194 |
bert.encoder.layer.11.output.dense |
Linear |
bias |
[768] |
768 |
768 |
0 |
195 |
bert.encoder.layer.11.output.LayerNorm |
LayerNorm |
weight |
[768] |
768 |
768 |
0 |
196 |
bert.encoder.layer.11.output.LayerNorm |
LayerNorm |
bias |
[768] |
768 |
768 |
0 |
197 |
qa_outputs |
Linear |
weight |
[2, 768] |
1536 |
1536 |
0 |
198 |
qa_outputs |
Linear |
bias |
[2] |
2 |
2 |
0 |