Hjgugugjhuhjggg
commited on
Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- job_new.json +519 -513
- measurement.json +1 -1
- out_tensor/lm_head.safetensors +1 -1
- out_tensor/model.layers.1.mlp.down_proj.safetensors +2 -2
- out_tensor/model.layers.1.mlp.gate_proj.safetensors +2 -2
- out_tensor/model.layers.1.mlp.up_proj.safetensors +2 -2
- out_tensor/model.layers.1.self_attn.k_proj.safetensors +2 -2
- out_tensor/model.layers.1.self_attn.o_proj.safetensors +2 -2
- out_tensor/model.layers.1.self_attn.q_proj.safetensors +2 -2
- out_tensor/model.layers.1.self_attn.v_proj.safetensors +2 -2
- out_tensor/model.layers.10.mlp.down_proj.safetensors +2 -2
- out_tensor/model.layers.10.mlp.gate_proj.safetensors +2 -2
- out_tensor/model.layers.10.mlp.up_proj.safetensors +2 -2
- out_tensor/model.layers.10.self_attn.k_proj.safetensors +2 -2
- out_tensor/model.layers.10.self_attn.o_proj.safetensors +2 -2
- out_tensor/model.layers.10.self_attn.q_proj.safetensors +2 -2
- out_tensor/model.layers.10.self_attn.v_proj.safetensors +2 -2
- out_tensor/model.layers.11.mlp.down_proj.safetensors +2 -2
- out_tensor/model.layers.11.mlp.gate_proj.safetensors +2 -2
- out_tensor/model.layers.11.mlp.up_proj.safetensors +2 -2
- out_tensor/model.layers.11.self_attn.k_proj.safetensors +2 -2
- out_tensor/model.layers.11.self_attn.o_proj.safetensors +2 -2
- out_tensor/model.layers.11.self_attn.q_proj.safetensors +2 -2
- out_tensor/model.layers.11.self_attn.v_proj.safetensors +2 -2
- out_tensor/model.layers.12.mlp.down_proj.safetensors +2 -2
- out_tensor/model.layers.12.mlp.gate_proj.safetensors +2 -2
- out_tensor/model.layers.12.mlp.up_proj.safetensors +2 -2
- out_tensor/model.layers.12.self_attn.k_proj.safetensors +1 -1
- out_tensor/model.layers.12.self_attn.o_proj.safetensors +1 -1
- out_tensor/model.layers.12.self_attn.q_proj.safetensors +1 -1
- out_tensor/model.layers.12.self_attn.v_proj.safetensors +1 -1
- out_tensor/model.layers.13.mlp.down_proj.safetensors +1 -1
- out_tensor/model.layers.13.mlp.gate_proj.safetensors +1 -1
- out_tensor/model.layers.13.mlp.up_proj.safetensors +1 -1
- out_tensor/model.layers.13.self_attn.k_proj.safetensors +2 -2
- out_tensor/model.layers.13.self_attn.o_proj.safetensors +2 -2
- out_tensor/model.layers.13.self_attn.q_proj.safetensors +2 -2
- out_tensor/model.layers.13.self_attn.v_proj.safetensors +2 -2
- out_tensor/model.layers.14.mlp.down_proj.safetensors +2 -2
- out_tensor/model.layers.14.mlp.gate_proj.safetensors +2 -2
- out_tensor/model.layers.14.mlp.up_proj.safetensors +2 -2
- out_tensor/model.layers.14.self_attn.k_proj.safetensors +2 -2
- out_tensor/model.layers.14.self_attn.o_proj.safetensors +2 -2
- out_tensor/model.layers.14.self_attn.q_proj.safetensors +2 -2
- out_tensor/model.layers.14.self_attn.v_proj.safetensors +2 -2
- out_tensor/model.layers.15.mlp.down_proj.safetensors +2 -2
- out_tensor/model.layers.15.mlp.gate_proj.safetensors +2 -2
- out_tensor/model.layers.15.mlp.up_proj.safetensors +2 -2
- out_tensor/model.layers.15.self_attn.k_proj.safetensors +2 -2
- out_tensor/model.layers.15.self_attn.o_proj.safetensors +2 -2
job_new.json
CHANGED
@@ -57975,7 +57975,7 @@
|
|
57975 |
}
|
57976 |
},
|
57977 |
{
|
57978 |
-
"accuracy": 0.
|
57979 |
"total_bits": 9097152,
|
57980 |
"gate_proj": {
|
57981 |
"group_size": {
|
@@ -58657,12 +58657,12 @@
|
|
58657 |
}
|
58658 |
},
|
58659 |
"model.layers.1.self_attn": {
|
58660 |
-
"accuracy": 0.
|
58661 |
-
"total_bits":
|
58662 |
"q_proj": {
|
58663 |
"group_size": {
|
58664 |
-
"5":
|
58665 |
-
"4":
|
58666 |
},
|
58667 |
"bits": [
|
58668 |
5,
|
@@ -58676,8 +58676,8 @@
|
|
58676 |
},
|
58677 |
"k_proj": {
|
58678 |
"group_size": {
|
58679 |
-
"5":
|
58680 |
-
"4":
|
58681 |
},
|
58682 |
"bits": [
|
58683 |
5,
|
@@ -58691,20 +58691,23 @@
|
|
58691 |
},
|
58692 |
"v_proj": {
|
58693 |
"group_size": {
|
58694 |
-
"5":
|
|
|
58695 |
},
|
58696 |
"bits": [
|
58697 |
-
5
|
|
|
58698 |
],
|
58699 |
"bits_prop": [
|
58700 |
-
1
|
|
|
58701 |
],
|
58702 |
"scale_bits": 4
|
58703 |
},
|
58704 |
"o_proj": {
|
58705 |
"group_size": {
|
58706 |
-
"5":
|
58707 |
-
"4":
|
58708 |
},
|
58709 |
"bits": [
|
58710 |
5,
|
@@ -58718,16 +58721,16 @@
|
|
58718 |
}
|
58719 |
},
|
58720 |
"model.layers.1.mlp": {
|
58721 |
-
"accuracy": 0.
|
58722 |
-
"total_bits":
|
58723 |
"gate_proj": {
|
58724 |
"group_size": {
|
58725 |
-
"
|
58726 |
-
"
|
58727 |
},
|
58728 |
"bits": [
|
58729 |
-
|
58730 |
-
|
58731 |
],
|
58732 |
"bits_prop": [
|
58733 |
0.1,
|
@@ -58737,12 +58740,12 @@
|
|
58737 |
},
|
58738 |
"up_proj": {
|
58739 |
"group_size": {
|
58740 |
-
"
|
58741 |
-
"
|
58742 |
},
|
58743 |
"bits": [
|
58744 |
-
|
58745 |
-
|
58746 |
],
|
58747 |
"bits_prop": [
|
58748 |
0.25,
|
@@ -58753,13 +58756,13 @@
|
|
58753 |
"down_proj": {
|
58754 |
"group_size": {
|
58755 |
"8": 32,
|
58756 |
-
"
|
58757 |
-
"
|
58758 |
},
|
58759 |
"bits": [
|
58760 |
8,
|
58761 |
-
|
58762 |
-
|
58763 |
],
|
58764 |
"bits_prop": [
|
58765 |
0.05,
|
@@ -58770,12 +58773,12 @@
|
|
58770 |
}
|
58771 |
},
|
58772 |
"model.layers.2.self_attn": {
|
58773 |
-
"accuracy": 0.
|
58774 |
-
"total_bits":
|
58775 |
"q_proj": {
|
58776 |
"group_size": {
|
58777 |
-
"5":
|
58778 |
-
"4":
|
58779 |
},
|
58780 |
"bits": [
|
58781 |
5,
|
@@ -58789,8 +58792,8 @@
|
|
58789 |
},
|
58790 |
"k_proj": {
|
58791 |
"group_size": {
|
58792 |
-
"5":
|
58793 |
-
"4":
|
58794 |
},
|
58795 |
"bits": [
|
58796 |
5,
|
@@ -58804,20 +58807,23 @@
|
|
58804 |
},
|
58805 |
"v_proj": {
|
58806 |
"group_size": {
|
58807 |
-
"5": 32
|
|
|
58808 |
},
|
58809 |
"bits": [
|
58810 |
-
5
|
|
|
58811 |
],
|
58812 |
"bits_prop": [
|
58813 |
-
1
|
|
|
58814 |
],
|
58815 |
"scale_bits": 4
|
58816 |
},
|
58817 |
"o_proj": {
|
58818 |
"group_size": {
|
58819 |
-
"5":
|
58820 |
-
"4":
|
58821 |
},
|
58822 |
"bits": [
|
58823 |
5,
|
@@ -58883,76 +58889,64 @@
|
|
58883 |
}
|
58884 |
},
|
58885 |
"model.layers.3.self_attn": {
|
58886 |
-
"accuracy": 0.
|
58887 |
-
"total_bits":
|
58888 |
"q_proj": {
|
58889 |
"group_size": {
|
58890 |
-
"
|
58891 |
-
"4": 64
|
58892 |
},
|
58893 |
"bits": [
|
58894 |
-
|
58895 |
-
4
|
58896 |
],
|
58897 |
"bits_prop": [
|
58898 |
-
|
58899 |
-
0.9
|
58900 |
],
|
58901 |
"scale_bits": 4
|
58902 |
},
|
58903 |
"k_proj": {
|
58904 |
"group_size": {
|
58905 |
-
"
|
58906 |
-
"4": 64
|
58907 |
},
|
58908 |
"bits": [
|
58909 |
-
|
58910 |
-
4
|
58911 |
],
|
58912 |
"bits_prop": [
|
58913 |
-
|
58914 |
-
0.9
|
58915 |
],
|
58916 |
"scale_bits": 4
|
58917 |
},
|
58918 |
"v_proj": {
|
58919 |
"group_size": {
|
58920 |
-
"
|
58921 |
-
"4": 32
|
58922 |
},
|
58923 |
"bits": [
|
58924 |
-
|
58925 |
-
4
|
58926 |
],
|
58927 |
"bits_prop": [
|
58928 |
-
|
58929 |
-
0.9
|
58930 |
],
|
58931 |
"scale_bits": 4
|
58932 |
},
|
58933 |
"o_proj": {
|
58934 |
"group_size": {
|
58935 |
-
"
|
58936 |
-
"4": 64
|
58937 |
},
|
58938 |
"bits": [
|
58939 |
-
|
58940 |
-
4
|
58941 |
],
|
58942 |
"bits_prop": [
|
58943 |
-
|
58944 |
-
0.9
|
58945 |
],
|
58946 |
"scale_bits": 4
|
58947 |
}
|
58948 |
},
|
58949 |
"model.layers.3.mlp": {
|
58950 |
-
"accuracy": 0.
|
58951 |
-
"total_bits":
|
58952 |
"gate_proj": {
|
58953 |
"group_size": {
|
58954 |
-
"6":
|
58955 |
-
"5":
|
58956 |
},
|
58957 |
"bits": [
|
58958 |
6,
|
@@ -58966,8 +58960,8 @@
|
|
58966 |
},
|
58967 |
"up_proj": {
|
58968 |
"group_size": {
|
58969 |
-
"6":
|
58970 |
-
"5":
|
58971 |
},
|
58972 |
"bits": [
|
58973 |
6,
|
@@ -58982,8 +58976,8 @@
|
|
58982 |
"down_proj": {
|
58983 |
"group_size": {
|
58984 |
"8": 32,
|
58985 |
-
"6":
|
58986 |
-
"5":
|
58987 |
},
|
58988 |
"bits": [
|
58989 |
8,
|
@@ -59103,16 +59097,16 @@
|
|
59103 |
}
|
59104 |
},
|
59105 |
"model.layers.5.self_attn": {
|
59106 |
-
"accuracy": 0.
|
59107 |
-
"total_bits":
|
59108 |
"q_proj": {
|
59109 |
"group_size": {
|
59110 |
-
"
|
59111 |
-
"
|
59112 |
},
|
59113 |
"bits": [
|
59114 |
-
|
59115 |
-
|
59116 |
],
|
59117 |
"bits_prop": [
|
59118 |
0.1,
|
@@ -59122,12 +59116,12 @@
|
|
59122 |
},
|
59123 |
"k_proj": {
|
59124 |
"group_size": {
|
59125 |
-
"
|
59126 |
-
"
|
59127 |
},
|
59128 |
"bits": [
|
59129 |
-
|
59130 |
-
|
59131 |
],
|
59132 |
"bits_prop": [
|
59133 |
0.1,
|
@@ -59137,10 +59131,10 @@
|
|
59137 |
},
|
59138 |
"v_proj": {
|
59139 |
"group_size": {
|
59140 |
-
"
|
59141 |
},
|
59142 |
"bits": [
|
59143 |
-
|
59144 |
],
|
59145 |
"bits_prop": [
|
59146 |
1
|
@@ -59149,12 +59143,12 @@
|
|
59149 |
},
|
59150 |
"o_proj": {
|
59151 |
"group_size": {
|
59152 |
-
"
|
59153 |
-
"
|
59154 |
},
|
59155 |
"bits": [
|
59156 |
-
|
59157 |
-
|
59158 |
],
|
59159 |
"bits_prop": [
|
59160 |
0.1,
|
@@ -59164,53 +59158,44 @@
|
|
59164 |
}
|
59165 |
},
|
59166 |
"model.layers.5.mlp": {
|
59167 |
-
"accuracy": 0.
|
59168 |
-
"total_bits":
|
59169 |
"gate_proj": {
|
59170 |
"group_size": {
|
59171 |
-
"5": 32,
|
59172 |
"4": 32
|
59173 |
},
|
59174 |
"bits": [
|
59175 |
-
5,
|
59176 |
4
|
59177 |
],
|
59178 |
"bits_prop": [
|
59179 |
-
|
59180 |
-
0.9
|
59181 |
],
|
59182 |
"scale_bits": 4
|
59183 |
},
|
59184 |
"up_proj": {
|
59185 |
"group_size": {
|
59186 |
-
"5": 32,
|
59187 |
"4": 32
|
59188 |
},
|
59189 |
"bits": [
|
59190 |
-
5,
|
59191 |
4
|
59192 |
],
|
59193 |
"bits_prop": [
|
59194 |
-
|
59195 |
-
0.75
|
59196 |
],
|
59197 |
"scale_bits": 4
|
59198 |
},
|
59199 |
"down_proj": {
|
59200 |
"group_size": {
|
59201 |
"8": 32,
|
59202 |
-
"5": 32,
|
59203 |
"4": 32
|
59204 |
},
|
59205 |
"bits": [
|
59206 |
8,
|
59207 |
-
5,
|
59208 |
4
|
59209 |
],
|
59210 |
"bits_prop": [
|
59211 |
0.05,
|
59212 |
-
0.
|
59213 |
-
0.85
|
59214 |
],
|
59215 |
"scale_bits": 4
|
59216 |
}
|
@@ -59320,41 +59305,35 @@
|
|
59320 |
}
|
59321 |
},
|
59322 |
"model.layers.7.self_attn": {
|
59323 |
-
"accuracy": 0.
|
59324 |
-
"total_bits":
|
59325 |
"q_proj": {
|
59326 |
"group_size": {
|
59327 |
-
"6":
|
59328 |
-
"5": 32
|
59329 |
},
|
59330 |
"bits": [
|
59331 |
-
6
|
59332 |
-
5
|
59333 |
],
|
59334 |
"bits_prop": [
|
59335 |
-
|
59336 |
-
0.9
|
59337 |
],
|
59338 |
"scale_bits": 4
|
59339 |
},
|
59340 |
"k_proj": {
|
59341 |
"group_size": {
|
59342 |
-
"6":
|
59343 |
-
"5": 32
|
59344 |
},
|
59345 |
"bits": [
|
59346 |
-
6
|
59347 |
-
5
|
59348 |
],
|
59349 |
"bits_prop": [
|
59350 |
-
|
59351 |
-
0.9
|
59352 |
],
|
59353 |
"scale_bits": 4
|
59354 |
},
|
59355 |
"v_proj": {
|
59356 |
"group_size": {
|
59357 |
-
"6":
|
59358 |
},
|
59359 |
"bits": [
|
59360 |
6
|
@@ -59366,27 +59345,24 @@
|
|
59366 |
},
|
59367 |
"o_proj": {
|
59368 |
"group_size": {
|
59369 |
-
"6":
|
59370 |
-
"5": 32
|
59371 |
},
|
59372 |
"bits": [
|
59373 |
-
6
|
59374 |
-
5
|
59375 |
],
|
59376 |
"bits_prop": [
|
59377 |
-
|
59378 |
-
0.9
|
59379 |
],
|
59380 |
"scale_bits": 4
|
59381 |
}
|
59382 |
},
|
59383 |
"model.layers.7.mlp": {
|
59384 |
-
"accuracy": 0.
|
59385 |
-
"total_bits":
|
59386 |
"gate_proj": {
|
59387 |
"group_size": {
|
59388 |
-
"6":
|
59389 |
-
"5":
|
59390 |
},
|
59391 |
"bits": [
|
59392 |
6,
|
@@ -59400,8 +59376,8 @@
|
|
59400 |
},
|
59401 |
"up_proj": {
|
59402 |
"group_size": {
|
59403 |
-
"6":
|
59404 |
-
"5":
|
59405 |
},
|
59406 |
"bits": [
|
59407 |
6,
|
@@ -59416,8 +59392,8 @@
|
|
59416 |
"down_proj": {
|
59417 |
"group_size": {
|
59418 |
"8": 32,
|
59419 |
-
"6":
|
59420 |
-
"5":
|
59421 |
},
|
59422 |
"bits": [
|
59423 |
8,
|
@@ -59607,59 +59583,68 @@
|
|
59607 |
}
|
59608 |
},
|
59609 |
"model.layers.9.mlp": {
|
59610 |
-
"accuracy": 0.
|
59611 |
-
"total_bits":
|
59612 |
"gate_proj": {
|
59613 |
"group_size": {
|
59614 |
-
"
|
|
|
59615 |
},
|
59616 |
"bits": [
|
59617 |
-
|
|
|
59618 |
],
|
59619 |
"bits_prop": [
|
59620 |
-
1
|
|
|
59621 |
],
|
59622 |
"scale_bits": 4
|
59623 |
},
|
59624 |
"up_proj": {
|
59625 |
"group_size": {
|
59626 |
-
"
|
|
|
59627 |
},
|
59628 |
"bits": [
|
59629 |
-
|
|
|
59630 |
],
|
59631 |
"bits_prop": [
|
59632 |
-
|
|
|
59633 |
],
|
59634 |
"scale_bits": 4
|
59635 |
},
|
59636 |
"down_proj": {
|
59637 |
"group_size": {
|
59638 |
"8": 32,
|
59639 |
-
"
|
|
|
59640 |
},
|
59641 |
"bits": [
|
59642 |
8,
|
59643 |
-
|
|
|
59644 |
],
|
59645 |
"bits_prop": [
|
59646 |
0.05,
|
59647 |
-
0.
|
|
|
59648 |
],
|
59649 |
"scale_bits": 4
|
59650 |
}
|
59651 |
},
|
59652 |
"model.layers.10.self_attn": {
|
59653 |
-
"accuracy": 0.
|
59654 |
-
"total_bits":
|
59655 |
"q_proj": {
|
59656 |
"group_size": {
|
59657 |
-
"
|
59658 |
-
"
|
59659 |
},
|
59660 |
"bits": [
|
59661 |
-
|
59662 |
-
|
59663 |
],
|
59664 |
"bits_prop": [
|
59665 |
0.1,
|
@@ -59669,12 +59654,12 @@
|
|
59669 |
},
|
59670 |
"k_proj": {
|
59671 |
"group_size": {
|
59672 |
-
"
|
59673 |
-
"
|
59674 |
},
|
59675 |
"bits": [
|
59676 |
-
|
59677 |
-
|
59678 |
],
|
59679 |
"bits_prop": [
|
59680 |
0.1,
|
@@ -59684,10 +59669,10 @@
|
|
59684 |
},
|
59685 |
"v_proj": {
|
59686 |
"group_size": {
|
59687 |
-
"
|
59688 |
},
|
59689 |
"bits": [
|
59690 |
-
|
59691 |
],
|
59692 |
"bits_prop": [
|
59693 |
1
|
@@ -59696,12 +59681,12 @@
|
|
59696 |
},
|
59697 |
"o_proj": {
|
59698 |
"group_size": {
|
59699 |
-
"
|
59700 |
-
"
|
59701 |
},
|
59702 |
"bits": [
|
59703 |
-
|
59704 |
-
|
59705 |
],
|
59706 |
"bits_prop": [
|
59707 |
0.1,
|
@@ -59711,12 +59696,12 @@
|
|
59711 |
}
|
59712 |
},
|
59713 |
"model.layers.10.mlp": {
|
59714 |
-
"accuracy": 0.
|
59715 |
-
"total_bits":
|
59716 |
"gate_proj": {
|
59717 |
"group_size": {
|
59718 |
-
"6":
|
59719 |
-
"5":
|
59720 |
},
|
59721 |
"bits": [
|
59722 |
6,
|
@@ -59730,8 +59715,8 @@
|
|
59730 |
},
|
59731 |
"up_proj": {
|
59732 |
"group_size": {
|
59733 |
-
"6":
|
59734 |
-
"5":
|
59735 |
},
|
59736 |
"bits": [
|
59737 |
6,
|
@@ -59746,8 +59731,8 @@
|
|
59746 |
"down_proj": {
|
59747 |
"group_size": {
|
59748 |
"8": 32,
|
59749 |
-
"6":
|
59750 |
-
"5":
|
59751 |
},
|
59752 |
"bits": [
|
59753 |
8,
|
@@ -59763,44 +59748,38 @@
|
|
59763 |
}
|
59764 |
},
|
59765 |
"model.layers.11.self_attn": {
|
59766 |
-
"accuracy": 0.
|
59767 |
-
"total_bits":
|
59768 |
"q_proj": {
|
59769 |
"group_size": {
|
59770 |
-
"6":
|
59771 |
-
"5": 128
|
59772 |
},
|
59773 |
"bits": [
|
59774 |
-
6
|
59775 |
-
5
|
59776 |
],
|
59777 |
"bits_prop": [
|
59778 |
-
|
59779 |
-
0.9
|
59780 |
],
|
59781 |
"scale_bits": 4
|
59782 |
},
|
59783 |
"k_proj": {
|
59784 |
"group_size": {
|
59785 |
-
"6":
|
59786 |
-
"5": 128
|
59787 |
},
|
59788 |
"bits": [
|
59789 |
-
6
|
59790 |
-
5
|
59791 |
],
|
59792 |
"bits_prop": [
|
59793 |
-
|
59794 |
-
0.9
|
59795 |
],
|
59796 |
"scale_bits": 4
|
59797 |
},
|
59798 |
"v_proj": {
|
59799 |
"group_size": {
|
59800 |
-
"
|
59801 |
},
|
59802 |
"bits": [
|
59803 |
-
|
59804 |
],
|
59805 |
"bits_prop": [
|
59806 |
1
|
@@ -59809,27 +59788,24 @@
|
|
59809 |
},
|
59810 |
"o_proj": {
|
59811 |
"group_size": {
|
59812 |
-
"6":
|
59813 |
-
"5": 128
|
59814 |
},
|
59815 |
"bits": [
|
59816 |
-
6
|
59817 |
-
5
|
59818 |
],
|
59819 |
"bits_prop": [
|
59820 |
-
|
59821 |
-
0.9
|
59822 |
],
|
59823 |
"scale_bits": 4
|
59824 |
}
|
59825 |
},
|
59826 |
"model.layers.11.mlp": {
|
59827 |
-
"accuracy": 0.
|
59828 |
-
"total_bits":
|
59829 |
"gate_proj": {
|
59830 |
"group_size": {
|
59831 |
-
"4":
|
59832 |
-
"3":
|
59833 |
},
|
59834 |
"bits": [
|
59835 |
4,
|
@@ -59843,8 +59819,8 @@
|
|
59843 |
},
|
59844 |
"up_proj": {
|
59845 |
"group_size": {
|
59846 |
-
"4":
|
59847 |
-
"3":
|
59848 |
},
|
59849 |
"bits": [
|
59850 |
4,
|
@@ -59859,8 +59835,8 @@
|
|
59859 |
"down_proj": {
|
59860 |
"group_size": {
|
59861 |
"8": 32,
|
59862 |
-
"4":
|
59863 |
-
"3":
|
59864 |
},
|
59865 |
"bits": [
|
59866 |
8,
|
@@ -59940,16 +59916,16 @@
|
|
59940 |
}
|
59941 |
},
|
59942 |
"model.layers.12.mlp": {
|
59943 |
-
"accuracy": 0.
|
59944 |
-
"total_bits":
|
59945 |
"gate_proj": {
|
59946 |
"group_size": {
|
59947 |
-
"
|
59948 |
-
"
|
59949 |
},
|
59950 |
"bits": [
|
59951 |
-
|
59952 |
-
|
59953 |
],
|
59954 |
"bits_prop": [
|
59955 |
0.1,
|
@@ -59959,45 +59935,42 @@
|
|
59959 |
},
|
59960 |
"up_proj": {
|
59961 |
"group_size": {
|
59962 |
-
"
|
59963 |
-
"
|
59964 |
},
|
59965 |
"bits": [
|
59966 |
-
|
59967 |
-
|
59968 |
],
|
59969 |
"bits_prop": [
|
59970 |
-
0.
|
59971 |
-
0.
|
59972 |
],
|
59973 |
"scale_bits": 4
|
59974 |
},
|
59975 |
"down_proj": {
|
59976 |
"group_size": {
|
59977 |
-
"
|
59978 |
-
"
|
59979 |
-
"3": 128
|
59980 |
},
|
59981 |
"bits": [
|
59982 |
-
|
59983 |
-
4,
|
59984 |
3
|
59985 |
],
|
59986 |
"bits_prop": [
|
59987 |
0.05,
|
59988 |
-
0.
|
59989 |
-
0.85
|
59990 |
],
|
59991 |
"scale_bits": 4
|
59992 |
}
|
59993 |
},
|
59994 |
"model.layers.13.self_attn": {
|
59995 |
-
"accuracy": 0.
|
59996 |
-
"total_bits":
|
59997 |
"q_proj": {
|
59998 |
"group_size": {
|
59999 |
-
"5":
|
60000 |
-
"4":
|
60001 |
},
|
60002 |
"bits": [
|
60003 |
5,
|
@@ -60011,8 +59984,8 @@
|
|
60011 |
},
|
60012 |
"k_proj": {
|
60013 |
"group_size": {
|
60014 |
-
"5":
|
60015 |
-
"4":
|
60016 |
},
|
60017 |
"bits": [
|
60018 |
5,
|
@@ -60026,23 +59999,20 @@
|
|
60026 |
},
|
60027 |
"v_proj": {
|
60028 |
"group_size": {
|
60029 |
-
"5": 32
|
60030 |
-
"4": 32
|
60031 |
},
|
60032 |
"bits": [
|
60033 |
-
5
|
60034 |
-
4
|
60035 |
],
|
60036 |
"bits_prop": [
|
60037 |
-
|
60038 |
-
0.9
|
60039 |
],
|
60040 |
"scale_bits": 4
|
60041 |
},
|
60042 |
"o_proj": {
|
60043 |
"group_size": {
|
60044 |
-
"5":
|
60045 |
-
"4":
|
60046 |
},
|
60047 |
"bits": [
|
60048 |
5,
|
@@ -60105,44 +60075,38 @@
|
|
60105 |
}
|
60106 |
},
|
60107 |
"model.layers.14.self_attn": {
|
60108 |
-
"accuracy": 0.
|
60109 |
-
"total_bits":
|
60110 |
"q_proj": {
|
60111 |
"group_size": {
|
60112 |
-
"
|
60113 |
-
"5": 32
|
60114 |
},
|
60115 |
"bits": [
|
60116 |
-
|
60117 |
-
5
|
60118 |
],
|
60119 |
"bits_prop": [
|
60120 |
-
|
60121 |
-
0.9
|
60122 |
],
|
60123 |
"scale_bits": 4
|
60124 |
},
|
60125 |
"k_proj": {
|
60126 |
"group_size": {
|
60127 |
-
"
|
60128 |
-
"5": 32
|
60129 |
},
|
60130 |
"bits": [
|
60131 |
-
|
60132 |
-
5
|
60133 |
],
|
60134 |
"bits_prop": [
|
60135 |
-
|
60136 |
-
0.9
|
60137 |
],
|
60138 |
"scale_bits": 4
|
60139 |
},
|
60140 |
"v_proj": {
|
60141 |
"group_size": {
|
60142 |
-
"
|
60143 |
},
|
60144 |
"bits": [
|
60145 |
-
|
60146 |
],
|
60147 |
"bits_prop": [
|
60148 |
1
|
@@ -60151,31 +60115,28 @@
|
|
60151 |
},
|
60152 |
"o_proj": {
|
60153 |
"group_size": {
|
60154 |
-
"
|
60155 |
-
"5": 32
|
60156 |
},
|
60157 |
"bits": [
|
60158 |
-
|
60159 |
-
5
|
60160 |
],
|
60161 |
"bits_prop": [
|
60162 |
-
|
60163 |
-
0.9
|
60164 |
],
|
60165 |
"scale_bits": 4
|
60166 |
}
|
60167 |
},
|
60168 |
"model.layers.14.mlp": {
|
60169 |
-
"accuracy": 0.
|
60170 |
-
"total_bits":
|
60171 |
"gate_proj": {
|
60172 |
"group_size": {
|
60173 |
-
"
|
60174 |
-
"
|
60175 |
},
|
60176 |
"bits": [
|
60177 |
-
|
60178 |
-
|
60179 |
],
|
60180 |
"bits_prop": [
|
60181 |
0.1,
|
@@ -60185,42 +60146,45 @@
|
|
60185 |
},
|
60186 |
"up_proj": {
|
60187 |
"group_size": {
|
60188 |
-
"
|
60189 |
-
"
|
60190 |
},
|
60191 |
"bits": [
|
60192 |
-
|
60193 |
-
|
60194 |
],
|
60195 |
"bits_prop": [
|
60196 |
-
0.
|
60197 |
-
0.
|
60198 |
],
|
60199 |
"scale_bits": 4
|
60200 |
},
|
60201 |
"down_proj": {
|
60202 |
"group_size": {
|
60203 |
-
"
|
|
|
60204 |
"3": 32
|
60205 |
},
|
60206 |
"bits": [
|
60207 |
-
|
|
|
60208 |
3
|
60209 |
],
|
60210 |
"bits_prop": [
|
60211 |
0.05,
|
60212 |
-
0.
|
|
|
60213 |
],
|
60214 |
"scale_bits": 4
|
60215 |
}
|
60216 |
},
|
60217 |
"model.layers.15.self_attn": {
|
60218 |
-
"accuracy": 0.
|
60219 |
-
"total_bits":
|
60220 |
"q_proj": {
|
60221 |
"group_size": {
|
60222 |
-
"5":
|
60223 |
-
"4":
|
60224 |
},
|
60225 |
"bits": [
|
60226 |
5,
|
@@ -60234,8 +60198,8 @@
|
|
60234 |
},
|
60235 |
"k_proj": {
|
60236 |
"group_size": {
|
60237 |
-
"5":
|
60238 |
-
"4":
|
60239 |
},
|
60240 |
"bits": [
|
60241 |
5,
|
@@ -60249,23 +60213,20 @@
|
|
60249 |
},
|
60250 |
"v_proj": {
|
60251 |
"group_size": {
|
60252 |
-
"5":
|
60253 |
-
"4": 64
|
60254 |
},
|
60255 |
"bits": [
|
60256 |
-
5
|
60257 |
-
4
|
60258 |
],
|
60259 |
"bits_prop": [
|
60260 |
-
|
60261 |
-
0.9
|
60262 |
],
|
60263 |
"scale_bits": 4
|
60264 |
},
|
60265 |
"o_proj": {
|
60266 |
"group_size": {
|
60267 |
-
"5":
|
60268 |
-
"4":
|
60269 |
},
|
60270 |
"bits": [
|
60271 |
5,
|
@@ -60279,16 +60240,16 @@
|
|
60279 |
}
|
60280 |
},
|
60281 |
"model.layers.15.mlp": {
|
60282 |
-
"accuracy": 0.
|
60283 |
-
"total_bits":
|
60284 |
"gate_proj": {
|
60285 |
"group_size": {
|
60286 |
-
"
|
60287 |
-
"
|
60288 |
},
|
60289 |
"bits": [
|
60290 |
-
|
60291 |
-
|
60292 |
],
|
60293 |
"bits_prop": [
|
60294 |
0.1,
|
@@ -60298,46 +60259,49 @@
|
|
60298 |
},
|
60299 |
"up_proj": {
|
60300 |
"group_size": {
|
60301 |
-
"
|
60302 |
-
"
|
60303 |
},
|
60304 |
"bits": [
|
60305 |
-
|
60306 |
-
|
60307 |
],
|
60308 |
"bits_prop": [
|
60309 |
-
0.
|
60310 |
-
0.
|
60311 |
],
|
60312 |
"scale_bits": 4
|
60313 |
},
|
60314 |
"down_proj": {
|
60315 |
"group_size": {
|
60316 |
-
"
|
60317 |
-
"
|
|
|
60318 |
},
|
60319 |
"bits": [
|
60320 |
-
|
|
|
60321 |
3
|
60322 |
],
|
60323 |
"bits_prop": [
|
60324 |
0.05,
|
60325 |
-
0.
|
|
|
60326 |
],
|
60327 |
"scale_bits": 4
|
60328 |
}
|
60329 |
},
|
60330 |
"model.layers.16.self_attn": {
|
60331 |
-
"accuracy": 0.
|
60332 |
-
"total_bits":
|
60333 |
"q_proj": {
|
60334 |
"group_size": {
|
60335 |
-
"
|
60336 |
-
"
|
60337 |
},
|
60338 |
"bits": [
|
60339 |
-
|
60340 |
-
|
60341 |
],
|
60342 |
"bits_prop": [
|
60343 |
0.1,
|
@@ -60347,12 +60311,12 @@
|
|
60347 |
},
|
60348 |
"k_proj": {
|
60349 |
"group_size": {
|
60350 |
-
"
|
60351 |
-
"
|
60352 |
},
|
60353 |
"bits": [
|
60354 |
-
|
60355 |
-
|
60356 |
],
|
60357 |
"bits_prop": [
|
60358 |
0.1,
|
@@ -60362,24 +60326,27 @@
|
|
60362 |
},
|
60363 |
"v_proj": {
|
60364 |
"group_size": {
|
60365 |
-
"
|
|
|
60366 |
},
|
60367 |
"bits": [
|
60368 |
-
|
|
|
60369 |
],
|
60370 |
"bits_prop": [
|
60371 |
-
1
|
|
|
60372 |
],
|
60373 |
"scale_bits": 4
|
60374 |
},
|
60375 |
"o_proj": {
|
60376 |
"group_size": {
|
60377 |
-
"
|
60378 |
-
"
|
60379 |
},
|
60380 |
"bits": [
|
60381 |
-
|
60382 |
-
|
60383 |
],
|
60384 |
"bits_prop": [
|
60385 |
0.1,
|
@@ -60389,16 +60356,16 @@
|
|
60389 |
}
|
60390 |
},
|
60391 |
"model.layers.16.mlp": {
|
60392 |
-
"accuracy": 0.
|
60393 |
-
"total_bits":
|
60394 |
"gate_proj": {
|
60395 |
"group_size": {
|
60396 |
-
"
|
60397 |
-
"
|
60398 |
},
|
60399 |
"bits": [
|
60400 |
-
|
60401 |
-
|
60402 |
],
|
60403 |
"bits_prop": [
|
60404 |
0.1,
|
@@ -60408,41 +60375,38 @@
|
|
60408 |
},
|
60409 |
"up_proj": {
|
60410 |
"group_size": {
|
60411 |
-
"
|
60412 |
-
"
|
60413 |
},
|
60414 |
"bits": [
|
60415 |
-
|
60416 |
-
|
60417 |
],
|
60418 |
"bits_prop": [
|
60419 |
-
0.
|
60420 |
-
0.
|
60421 |
],
|
60422 |
"scale_bits": 4
|
60423 |
},
|
60424 |
"down_proj": {
|
60425 |
"group_size": {
|
60426 |
-
"
|
60427 |
-
"4": 32,
|
60428 |
"3": 32
|
60429 |
},
|
60430 |
"bits": [
|
60431 |
-
|
60432 |
-
4,
|
60433 |
3
|
60434 |
],
|
60435 |
"bits_prop": [
|
60436 |
0.05,
|
60437 |
-
0.
|
60438 |
-
0.85
|
60439 |
],
|
60440 |
"scale_bits": 4
|
60441 |
}
|
60442 |
},
|
60443 |
"model.layers.17.self_attn": {
|
60444 |
-
"accuracy": 0.
|
60445 |
-
"total_bits":
|
60446 |
"q_proj": {
|
60447 |
"group_size": {
|
60448 |
"5": 64,
|
@@ -60475,16 +60439,13 @@
|
|
60475 |
},
|
60476 |
"v_proj": {
|
60477 |
"group_size": {
|
60478 |
-
"5":
|
60479 |
-
"4": 32
|
60480 |
},
|
60481 |
"bits": [
|
60482 |
-
5
|
60483 |
-
4
|
60484 |
],
|
60485 |
"bits_prop": [
|
60486 |
-
|
60487 |
-
0.9
|
60488 |
],
|
60489 |
"scale_bits": 4
|
60490 |
},
|
@@ -60505,16 +60466,16 @@
|
|
60505 |
}
|
60506 |
},
|
60507 |
"model.layers.17.mlp": {
|
60508 |
-
"accuracy": 0.
|
60509 |
-
"total_bits":
|
60510 |
"gate_proj": {
|
60511 |
"group_size": {
|
60512 |
-
"
|
60513 |
-
"
|
60514 |
},
|
60515 |
"bits": [
|
60516 |
-
|
60517 |
-
|
60518 |
],
|
60519 |
"bits_prop": [
|
60520 |
0.1,
|
@@ -60524,34 +60485,31 @@
|
|
60524 |
},
|
60525 |
"up_proj": {
|
60526 |
"group_size": {
|
60527 |
-
"
|
60528 |
-
"
|
60529 |
},
|
60530 |
"bits": [
|
60531 |
-
|
60532 |
-
|
60533 |
],
|
60534 |
"bits_prop": [
|
60535 |
-
0.
|
60536 |
-
0.
|
60537 |
],
|
60538 |
"scale_bits": 4
|
60539 |
},
|
60540 |
"down_proj": {
|
60541 |
"group_size": {
|
60542 |
-
"
|
60543 |
-
"4": 32,
|
60544 |
"3": 32
|
60545 |
},
|
60546 |
"bits": [
|
60547 |
-
|
60548 |
-
4,
|
60549 |
3
|
60550 |
],
|
60551 |
"bits_prop": [
|
60552 |
0.05,
|
60553 |
-
0.
|
60554 |
-
0.85
|
60555 |
],
|
60556 |
"scale_bits": 4
|
60557 |
}
|
@@ -60734,8 +60692,8 @@
|
|
60734 |
}
|
60735 |
},
|
60736 |
"model.layers.19.mlp": {
|
60737 |
-
"accuracy": 0.
|
60738 |
-
"total_bits":
|
60739 |
"gate_proj": {
|
60740 |
"group_size": {
|
60741 |
"4": 32,
|
@@ -60769,30 +60727,33 @@
|
|
60769 |
"down_proj": {
|
60770 |
"group_size": {
|
60771 |
"8": 32,
|
60772 |
-
"4":
|
|
|
60773 |
},
|
60774 |
"bits": [
|
60775 |
8,
|
60776 |
-
4
|
|
|
60777 |
],
|
60778 |
"bits_prop": [
|
60779 |
0.05,
|
60780 |
-
0.
|
|
|
60781 |
],
|
60782 |
"scale_bits": 4
|
60783 |
}
|
60784 |
},
|
60785 |
"model.layers.20.self_attn": {
|
60786 |
-
"accuracy": 0.
|
60787 |
-
"total_bits":
|
60788 |
"q_proj": {
|
60789 |
"group_size": {
|
60790 |
-
"
|
60791 |
-
"
|
60792 |
},
|
60793 |
"bits": [
|
60794 |
-
|
60795 |
-
|
60796 |
],
|
60797 |
"bits_prop": [
|
60798 |
0.1,
|
@@ -60802,12 +60763,12 @@
|
|
60802 |
},
|
60803 |
"k_proj": {
|
60804 |
"group_size": {
|
60805 |
-
"
|
60806 |
-
"
|
60807 |
},
|
60808 |
"bits": [
|
60809 |
-
|
60810 |
-
|
60811 |
],
|
60812 |
"bits_prop": [
|
60813 |
0.1,
|
@@ -60817,10 +60778,10 @@
|
|
60817 |
},
|
60818 |
"v_proj": {
|
60819 |
"group_size": {
|
60820 |
-
"
|
60821 |
},
|
60822 |
"bits": [
|
60823 |
-
|
60824 |
],
|
60825 |
"bits_prop": [
|
60826 |
1
|
@@ -60829,12 +60790,12 @@
|
|
60829 |
},
|
60830 |
"o_proj": {
|
60831 |
"group_size": {
|
60832 |
-
"
|
60833 |
-
"
|
60834 |
},
|
60835 |
"bits": [
|
60836 |
-
|
60837 |
-
|
60838 |
],
|
60839 |
"bits_prop": [
|
60840 |
0.1,
|
@@ -60844,11 +60805,11 @@
|
|
60844 |
}
|
60845 |
},
|
60846 |
"model.layers.20.mlp": {
|
60847 |
-
"accuracy": 0.
|
60848 |
-
"total_bits":
|
60849 |
"gate_proj": {
|
60850 |
"group_size": {
|
60851 |
-
"4":
|
60852 |
},
|
60853 |
"bits": [
|
60854 |
4
|
@@ -60873,7 +60834,7 @@
|
|
60873 |
"down_proj": {
|
60874 |
"group_size": {
|
60875 |
"8": 32,
|
60876 |
-
"4":
|
60877 |
},
|
60878 |
"bits": [
|
60879 |
8,
|
@@ -60887,8 +60848,8 @@
|
|
60887 |
}
|
60888 |
},
|
60889 |
"model.layers.21.self_attn": {
|
60890 |
-
"accuracy": 0.
|
60891 |
-
"total_bits":
|
60892 |
"q_proj": {
|
60893 |
"group_size": {
|
60894 |
"5": 64,
|
@@ -60921,13 +60882,16 @@
|
|
60921 |
},
|
60922 |
"v_proj": {
|
60923 |
"group_size": {
|
60924 |
-
"5":
|
|
|
60925 |
},
|
60926 |
"bits": [
|
60927 |
-
5
|
|
|
60928 |
],
|
60929 |
"bits_prop": [
|
60930 |
-
1
|
|
|
60931 |
],
|
60932 |
"scale_bits": 4
|
60933 |
},
|
@@ -60948,11 +60912,11 @@
|
|
60948 |
}
|
60949 |
},
|
60950 |
"model.layers.21.mlp": {
|
60951 |
-
"accuracy": 0.
|
60952 |
-
"total_bits":
|
60953 |
"gate_proj": {
|
60954 |
"group_size": {
|
60955 |
-
"4":
|
60956 |
},
|
60957 |
"bits": [
|
60958 |
4
|
@@ -60977,7 +60941,7 @@
|
|
60977 |
"down_proj": {
|
60978 |
"group_size": {
|
60979 |
"8": 32,
|
60980 |
-
"4":
|
60981 |
},
|
60982 |
"bits": [
|
60983 |
8,
|
@@ -60991,12 +60955,12 @@
|
|
60991 |
}
|
60992 |
},
|
60993 |
"model.layers.22.self_attn": {
|
60994 |
-
"accuracy": 0.
|
60995 |
-
"total_bits":
|
60996 |
"q_proj": {
|
60997 |
"group_size": {
|
60998 |
-
"5":
|
60999 |
-
"4":
|
61000 |
},
|
61001 |
"bits": [
|
61002 |
5,
|
@@ -61009,9 +60973,9 @@
|
|
61009 |
"scale_bits": 4
|
61010 |
},
|
61011 |
"k_proj": {
|
61012 |
-
"group_size": {
|
61013 |
-
"5":
|
61014 |
-
"4":
|
61015 |
},
|
61016 |
"bits": [
|
61017 |
5,
|
@@ -61025,8 +60989,8 @@
|
|
61025 |
},
|
61026 |
"v_proj": {
|
61027 |
"group_size": {
|
61028 |
-
"5":
|
61029 |
-
"4":
|
61030 |
},
|
61031 |
"bits": [
|
61032 |
5,
|
@@ -61040,8 +61004,8 @@
|
|
61040 |
},
|
61041 |
"o_proj": {
|
61042 |
"group_size": {
|
61043 |
-
"5":
|
61044 |
-
"4":
|
61045 |
},
|
61046 |
"bits": [
|
61047 |
5,
|
@@ -61098,38 +61062,44 @@
|
|
61098 |
}
|
61099 |
},
|
61100 |
"model.layers.23.self_attn": {
|
61101 |
-
"accuracy": 0.
|
61102 |
-
"total_bits":
|
61103 |
"q_proj": {
|
61104 |
"group_size": {
|
61105 |
-
"6": 32
|
|
|
61106 |
},
|
61107 |
"bits": [
|
61108 |
-
6
|
|
|
61109 |
],
|
61110 |
"bits_prop": [
|
61111 |
-
1
|
|
|
61112 |
],
|
61113 |
"scale_bits": 4
|
61114 |
},
|
61115 |
"k_proj": {
|
61116 |
"group_size": {
|
61117 |
-
"6": 32
|
|
|
61118 |
},
|
61119 |
"bits": [
|
61120 |
-
6
|
|
|
61121 |
],
|
61122 |
"bits_prop": [
|
61123 |
-
1
|
|
|
61124 |
],
|
61125 |
"scale_bits": 4
|
61126 |
},
|
61127 |
"v_proj": {
|
61128 |
"group_size": {
|
61129 |
-
"
|
61130 |
},
|
61131 |
"bits": [
|
61132 |
-
|
61133 |
],
|
61134 |
"bits_prop": [
|
61135 |
1
|
@@ -61138,13 +61108,16 @@
|
|
61138 |
},
|
61139 |
"o_proj": {
|
61140 |
"group_size": {
|
61141 |
-
"6": 32
|
|
|
61142 |
},
|
61143 |
"bits": [
|
61144 |
-
6
|
|
|
61145 |
],
|
61146 |
"bits_prop": [
|
61147 |
-
1
|
|
|
61148 |
],
|
61149 |
"scale_bits": 4
|
61150 |
}
|
@@ -61193,11 +61166,11 @@
|
|
61193 |
}
|
61194 |
},
|
61195 |
"model.layers.24.self_attn": {
|
61196 |
-
"accuracy": 0.
|
61197 |
-
"total_bits":
|
61198 |
"q_proj": {
|
61199 |
"group_size": {
|
61200 |
-
"6":
|
61201 |
},
|
61202 |
"bits": [
|
61203 |
6
|
@@ -61209,7 +61182,7 @@
|
|
61209 |
},
|
61210 |
"k_proj": {
|
61211 |
"group_size": {
|
61212 |
-
"6":
|
61213 |
},
|
61214 |
"bits": [
|
61215 |
6
|
@@ -61221,10 +61194,10 @@
|
|
61221 |
},
|
61222 |
"v_proj": {
|
61223 |
"group_size": {
|
61224 |
-
"
|
61225 |
},
|
61226 |
"bits": [
|
61227 |
-
|
61228 |
],
|
61229 |
"bits_prop": [
|
61230 |
1
|
@@ -61233,7 +61206,7 @@
|
|
61233 |
},
|
61234 |
"o_proj": {
|
61235 |
"group_size": {
|
61236 |
-
"6":
|
61237 |
},
|
61238 |
"bits": [
|
61239 |
6
|
@@ -61245,59 +61218,68 @@
|
|
61245 |
}
|
61246 |
},
|
61247 |
"model.layers.24.mlp": {
|
61248 |
-
"accuracy": 0.
|
61249 |
-
"total_bits":
|
61250 |
"gate_proj": {
|
61251 |
"group_size": {
|
61252 |
-
"
|
|
|
61253 |
},
|
61254 |
"bits": [
|
|
|
61255 |
4
|
61256 |
],
|
61257 |
"bits_prop": [
|
61258 |
-
1
|
|
|
61259 |
],
|
61260 |
"scale_bits": 4
|
61261 |
},
|
61262 |
"up_proj": {
|
61263 |
"group_size": {
|
61264 |
-
"
|
|
|
61265 |
},
|
61266 |
"bits": [
|
|
|
61267 |
4
|
61268 |
],
|
61269 |
"bits_prop": [
|
61270 |
-
|
|
|
61271 |
],
|
61272 |
"scale_bits": 4
|
61273 |
},
|
61274 |
"down_proj": {
|
61275 |
"group_size": {
|
61276 |
"8": 32,
|
61277 |
-
"
|
|
|
61278 |
},
|
61279 |
"bits": [
|
61280 |
8,
|
|
|
61281 |
4
|
61282 |
],
|
61283 |
"bits_prop": [
|
61284 |
0.05,
|
61285 |
-
0.
|
|
|
61286 |
],
|
61287 |
"scale_bits": 4
|
61288 |
}
|
61289 |
},
|
61290 |
"model.layers.25.self_attn": {
|
61291 |
-
"accuracy": 0.
|
61292 |
-
"total_bits":
|
61293 |
"q_proj": {
|
61294 |
"group_size": {
|
61295 |
-
"
|
61296 |
-
"
|
61297 |
},
|
61298 |
"bits": [
|
61299 |
-
|
61300 |
-
|
61301 |
],
|
61302 |
"bits_prop": [
|
61303 |
0.1,
|
@@ -61307,12 +61289,12 @@
|
|
61307 |
},
|
61308 |
"k_proj": {
|
61309 |
"group_size": {
|
61310 |
-
"
|
61311 |
-
"
|
61312 |
},
|
61313 |
"bits": [
|
61314 |
-
|
61315 |
-
|
61316 |
],
|
61317 |
"bits_prop": [
|
61318 |
0.1,
|
@@ -61322,24 +61304,27 @@
|
|
61322 |
},
|
61323 |
"v_proj": {
|
61324 |
"group_size": {
|
61325 |
-
"
|
|
|
61326 |
},
|
61327 |
"bits": [
|
61328 |
-
|
|
|
61329 |
],
|
61330 |
"bits_prop": [
|
61331 |
-
1
|
|
|
61332 |
],
|
61333 |
"scale_bits": 4
|
61334 |
},
|
61335 |
"o_proj": {
|
61336 |
"group_size": {
|
61337 |
-
"
|
61338 |
-
"
|
61339 |
},
|
61340 |
"bits": [
|
61341 |
-
|
61342 |
-
|
61343 |
],
|
61344 |
"bits_prop": [
|
61345 |
0.1,
|
@@ -61349,11 +61334,11 @@
|
|
61349 |
}
|
61350 |
},
|
61351 |
"model.layers.25.mlp": {
|
61352 |
-
"accuracy": 0.
|
61353 |
-
"total_bits":
|
61354 |
"gate_proj": {
|
61355 |
"group_size": {
|
61356 |
-
"4":
|
61357 |
},
|
61358 |
"bits": [
|
61359 |
4
|
@@ -61378,7 +61363,7 @@
|
|
61378 |
"down_proj": {
|
61379 |
"group_size": {
|
61380 |
"8": 32,
|
61381 |
-
"4":
|
61382 |
},
|
61383 |
"bits": [
|
61384 |
8,
|
@@ -61392,12 +61377,12 @@
|
|
61392 |
}
|
61393 |
},
|
61394 |
"model.layers.26.self_attn": {
|
61395 |
-
"accuracy": 0.
|
61396 |
-
"total_bits":
|
61397 |
"q_proj": {
|
61398 |
"group_size": {
|
61399 |
-
"5":
|
61400 |
-
"4":
|
61401 |
},
|
61402 |
"bits": [
|
61403 |
5,
|
@@ -61411,8 +61396,8 @@
|
|
61411 |
},
|
61412 |
"k_proj": {
|
61413 |
"group_size": {
|
61414 |
-
"5":
|
61415 |
-
"4":
|
61416 |
},
|
61417 |
"bits": [
|
61418 |
5,
|
@@ -61426,20 +61411,23 @@
|
|
61426 |
},
|
61427 |
"v_proj": {
|
61428 |
"group_size": {
|
61429 |
-
"5": 32
|
|
|
61430 |
},
|
61431 |
"bits": [
|
61432 |
-
5
|
|
|
61433 |
],
|
61434 |
"bits_prop": [
|
61435 |
-
1
|
|
|
61436 |
],
|
61437 |
"scale_bits": 4
|
61438 |
},
|
61439 |
"o_proj": {
|
61440 |
"group_size": {
|
61441 |
-
"5":
|
61442 |
-
"4":
|
61443 |
},
|
61444 |
"bits": [
|
61445 |
5,
|
@@ -61505,38 +61493,44 @@
|
|
61505 |
}
|
61506 |
},
|
61507 |
"model.layers.27.self_attn": {
|
61508 |
-
"accuracy": 0.
|
61509 |
-
"total_bits":
|
61510 |
"q_proj": {
|
61511 |
"group_size": {
|
61512 |
-
"
|
|
|
61513 |
},
|
61514 |
"bits": [
|
61515 |
-
|
|
|
61516 |
],
|
61517 |
"bits_prop": [
|
61518 |
-
1
|
|
|
61519 |
],
|
61520 |
"scale_bits": 4
|
61521 |
},
|
61522 |
"k_proj": {
|
61523 |
"group_size": {
|
61524 |
-
"
|
|
|
61525 |
},
|
61526 |
"bits": [
|
61527 |
-
|
|
|
61528 |
],
|
61529 |
"bits_prop": [
|
61530 |
-
1
|
|
|
61531 |
],
|
61532 |
"scale_bits": 4
|
61533 |
},
|
61534 |
"v_proj": {
|
61535 |
"group_size": {
|
61536 |
-
"
|
61537 |
},
|
61538 |
"bits": [
|
61539 |
-
|
61540 |
],
|
61541 |
"bits_prop": [
|
61542 |
1
|
@@ -61545,71 +61539,83 @@
|
|
61545 |
},
|
61546 |
"o_proj": {
|
61547 |
"group_size": {
|
61548 |
-
"
|
|
|
61549 |
},
|
61550 |
"bits": [
|
61551 |
-
|
|
|
61552 |
],
|
61553 |
"bits_prop": [
|
61554 |
-
1
|
|
|
61555 |
],
|
61556 |
"scale_bits": 4
|
61557 |
}
|
61558 |
},
|
61559 |
"model.layers.27.mlp": {
|
61560 |
-
"accuracy": 0.
|
61561 |
-
"total_bits":
|
61562 |
"gate_proj": {
|
61563 |
"group_size": {
|
61564 |
-
"
|
|
|
61565 |
},
|
61566 |
"bits": [
|
|
|
61567 |
4
|
61568 |
],
|
61569 |
"bits_prop": [
|
61570 |
-
1
|
|
|
61571 |
],
|
61572 |
"scale_bits": 4
|
61573 |
},
|
61574 |
"up_proj": {
|
61575 |
"group_size": {
|
|
|
61576 |
"4": 32
|
61577 |
},
|
61578 |
"bits": [
|
|
|
61579 |
4
|
61580 |
],
|
61581 |
"bits_prop": [
|
61582 |
-
|
|
|
61583 |
],
|
61584 |
"scale_bits": 4
|
61585 |
},
|
61586 |
"down_proj": {
|
61587 |
"group_size": {
|
61588 |
"8": 32,
|
61589 |
-
"
|
|
|
61590 |
},
|
61591 |
"bits": [
|
61592 |
8,
|
|
|
61593 |
4
|
61594 |
],
|
61595 |
"bits_prop": [
|
61596 |
0.05,
|
61597 |
-
0.
|
|
|
61598 |
],
|
61599 |
"scale_bits": 4
|
61600 |
}
|
61601 |
},
|
61602 |
"model.layers.28.self_attn": {
|
61603 |
-
"accuracy": 0.
|
61604 |
-
"total_bits":
|
61605 |
"q_proj": {
|
61606 |
"group_size": {
|
61607 |
-
"
|
61608 |
-
"
|
61609 |
},
|
61610 |
"bits": [
|
61611 |
-
|
61612 |
-
|
61613 |
],
|
61614 |
"bits_prop": [
|
61615 |
0.1,
|
@@ -61619,12 +61625,12 @@
|
|
61619 |
},
|
61620 |
"k_proj": {
|
61621 |
"group_size": {
|
61622 |
-
"
|
61623 |
-
"
|
61624 |
},
|
61625 |
"bits": [
|
61626 |
-
|
61627 |
-
|
61628 |
],
|
61629 |
"bits_prop": [
|
61630 |
0.1,
|
@@ -61634,10 +61640,10 @@
|
|
61634 |
},
|
61635 |
"v_proj": {
|
61636 |
"group_size": {
|
61637 |
-
"
|
61638 |
},
|
61639 |
"bits": [
|
61640 |
-
|
61641 |
],
|
61642 |
"bits_prop": [
|
61643 |
1
|
@@ -61646,12 +61652,12 @@
|
|
61646 |
},
|
61647 |
"o_proj": {
|
61648 |
"group_size": {
|
61649 |
-
"
|
61650 |
-
"
|
61651 |
},
|
61652 |
"bits": [
|
61653 |
-
|
61654 |
-
|
61655 |
],
|
61656 |
"bits_prop": [
|
61657 |
0.1,
|
@@ -61713,16 +61719,16 @@
|
|
61713 |
}
|
61714 |
},
|
61715 |
"model.layers.29.self_attn": {
|
61716 |
-
"accuracy": 0.
|
61717 |
-
"total_bits":
|
61718 |
"q_proj": {
|
61719 |
"group_size": {
|
61720 |
-
"
|
61721 |
-
"
|
61722 |
},
|
61723 |
"bits": [
|
61724 |
-
|
61725 |
-
|
61726 |
],
|
61727 |
"bits_prop": [
|
61728 |
0.1,
|
@@ -61732,12 +61738,12 @@
|
|
61732 |
},
|
61733 |
"k_proj": {
|
61734 |
"group_size": {
|
61735 |
-
"
|
61736 |
-
"
|
61737 |
},
|
61738 |
"bits": [
|
61739 |
-
|
61740 |
-
|
61741 |
],
|
61742 |
"bits_prop": [
|
61743 |
0.1,
|
@@ -61747,10 +61753,10 @@
|
|
61747 |
},
|
61748 |
"v_proj": {
|
61749 |
"group_size": {
|
61750 |
-
"
|
61751 |
},
|
61752 |
"bits": [
|
61753 |
-
|
61754 |
],
|
61755 |
"bits_prop": [
|
61756 |
1
|
@@ -61759,12 +61765,12 @@
|
|
61759 |
},
|
61760 |
"o_proj": {
|
61761 |
"group_size": {
|
61762 |
-
"
|
61763 |
-
"
|
61764 |
},
|
61765 |
"bits": [
|
61766 |
-
|
61767 |
-
|
61768 |
],
|
61769 |
"bits_prop": [
|
61770 |
0.1,
|
|
|
57975 |
}
|
57976 |
},
|
57977 |
{
|
57978 |
+
"accuracy": 0.945903766900301,
|
57979 |
"total_bits": 9097152,
|
57980 |
"gate_proj": {
|
57981 |
"group_size": {
|
|
|
58657 |
}
|
58658 |
},
|
58659 |
"model.layers.1.self_attn": {
|
58660 |
+
"accuracy": 0.9855817724019289,
|
58661 |
+
"total_bits": 3831936,
|
58662 |
"q_proj": {
|
58663 |
"group_size": {
|
58664 |
+
"5": 128,
|
58665 |
+
"4": 128
|
58666 |
},
|
58667 |
"bits": [
|
58668 |
5,
|
|
|
58676 |
},
|
58677 |
"k_proj": {
|
58678 |
"group_size": {
|
58679 |
+
"5": 128,
|
58680 |
+
"4": 128
|
58681 |
},
|
58682 |
"bits": [
|
58683 |
5,
|
|
|
58691 |
},
|
58692 |
"v_proj": {
|
58693 |
"group_size": {
|
58694 |
+
"5": 64,
|
58695 |
+
"4": 64
|
58696 |
},
|
58697 |
"bits": [
|
58698 |
+
5,
|
58699 |
+
4
|
58700 |
],
|
58701 |
"bits_prop": [
|
58702 |
+
0.1,
|
58703 |
+
0.9
|
58704 |
],
|
58705 |
"scale_bits": 4
|
58706 |
},
|
58707 |
"o_proj": {
|
58708 |
"group_size": {
|
58709 |
+
"5": 128,
|
58710 |
+
"4": 128
|
58711 |
},
|
58712 |
"bits": [
|
58713 |
5,
|
|
|
58721 |
}
|
58722 |
},
|
58723 |
"model.layers.1.mlp": {
|
58724 |
+
"accuracy": 0.9778902214020491,
|
58725 |
+
"total_bits": 11696064,
|
58726 |
"gate_proj": {
|
58727 |
"group_size": {
|
58728 |
+
"5": 32,
|
58729 |
+
"4": 32
|
58730 |
},
|
58731 |
"bits": [
|
58732 |
+
5,
|
58733 |
+
4
|
58734 |
],
|
58735 |
"bits_prop": [
|
58736 |
0.1,
|
|
|
58740 |
},
|
58741 |
"up_proj": {
|
58742 |
"group_size": {
|
58743 |
+
"5": 32,
|
58744 |
+
"4": 32
|
58745 |
},
|
58746 |
"bits": [
|
58747 |
+
5,
|
58748 |
+
4
|
58749 |
],
|
58750 |
"bits_prop": [
|
58751 |
0.25,
|
|
|
58756 |
"down_proj": {
|
58757 |
"group_size": {
|
58758 |
"8": 32,
|
58759 |
+
"5": 32,
|
58760 |
+
"4": 32
|
58761 |
},
|
58762 |
"bits": [
|
58763 |
8,
|
58764 |
+
5,
|
58765 |
+
4
|
58766 |
],
|
58767 |
"bits_prop": [
|
58768 |
0.05,
|
|
|
58773 |
}
|
58774 |
},
|
58775 |
"model.layers.2.self_attn": {
|
58776 |
+
"accuracy": 0.9876447143033147,
|
58777 |
+
"total_bits": 3775344,
|
58778 |
"q_proj": {
|
58779 |
"group_size": {
|
58780 |
+
"5": 64,
|
58781 |
+
"4": 64
|
58782 |
},
|
58783 |
"bits": [
|
58784 |
5,
|
|
|
58792 |
},
|
58793 |
"k_proj": {
|
58794 |
"group_size": {
|
58795 |
+
"5": 64,
|
58796 |
+
"4": 64
|
58797 |
},
|
58798 |
"bits": [
|
58799 |
5,
|
|
|
58807 |
},
|
58808 |
"v_proj": {
|
58809 |
"group_size": {
|
58810 |
+
"5": 32,
|
58811 |
+
"4": 32
|
58812 |
},
|
58813 |
"bits": [
|
58814 |
+
5,
|
58815 |
+
4
|
58816 |
],
|
58817 |
"bits_prop": [
|
58818 |
+
0.1,
|
58819 |
+
0.9
|
58820 |
],
|
58821 |
"scale_bits": 4
|
58822 |
},
|
58823 |
"o_proj": {
|
58824 |
"group_size": {
|
58825 |
+
"5": 64,
|
58826 |
+
"4": 64
|
58827 |
},
|
58828 |
"bits": [
|
58829 |
5,
|
|
|
58889 |
}
|
58890 |
},
|
58891 |
"model.layers.3.self_attn": {
|
58892 |
+
"accuracy": 0.9954097527079284,
|
58893 |
+
"total_bits": 5717376,
|
58894 |
"q_proj": {
|
58895 |
"group_size": {
|
58896 |
+
"6": 32
|
|
|
58897 |
},
|
58898 |
"bits": [
|
58899 |
+
6
|
|
|
58900 |
],
|
58901 |
"bits_prop": [
|
58902 |
+
1
|
|
|
58903 |
],
|
58904 |
"scale_bits": 4
|
58905 |
},
|
58906 |
"k_proj": {
|
58907 |
"group_size": {
|
58908 |
+
"6": 32
|
|
|
58909 |
},
|
58910 |
"bits": [
|
58911 |
+
6
|
|
|
58912 |
],
|
58913 |
"bits_prop": [
|
58914 |
+
1
|
|
|
58915 |
],
|
58916 |
"scale_bits": 4
|
58917 |
},
|
58918 |
"v_proj": {
|
58919 |
"group_size": {
|
58920 |
+
"8": 32
|
|
|
58921 |
},
|
58922 |
"bits": [
|
58923 |
+
8
|
|
|
58924 |
],
|
58925 |
"bits_prop": [
|
58926 |
+
1
|
|
|
58927 |
],
|
58928 |
"scale_bits": 4
|
58929 |
},
|
58930 |
"o_proj": {
|
58931 |
"group_size": {
|
58932 |
+
"6": 32
|
|
|
58933 |
},
|
58934 |
"bits": [
|
58935 |
+
6
|
|
|
58936 |
],
|
58937 |
"bits_prop": [
|
58938 |
+
1
|
|
|
58939 |
],
|
58940 |
"scale_bits": 4
|
58941 |
}
|
58942 |
},
|
58943 |
"model.layers.3.mlp": {
|
58944 |
+
"accuracy": 0.9891038266941905,
|
58945 |
+
"total_bits": 14294976,
|
58946 |
"gate_proj": {
|
58947 |
"group_size": {
|
58948 |
+
"6": 32,
|
58949 |
+
"5": 32
|
58950 |
},
|
58951 |
"bits": [
|
58952 |
6,
|
|
|
58960 |
},
|
58961 |
"up_proj": {
|
58962 |
"group_size": {
|
58963 |
+
"6": 32,
|
58964 |
+
"5": 32
|
58965 |
},
|
58966 |
"bits": [
|
58967 |
6,
|
|
|
58976 |
"down_proj": {
|
58977 |
"group_size": {
|
58978 |
"8": 32,
|
58979 |
+
"6": 32,
|
58980 |
+
"5": 32
|
58981 |
},
|
58982 |
"bits": [
|
58983 |
8,
|
|
|
59097 |
}
|
59098 |
},
|
59099 |
"model.layers.5.self_attn": {
|
59100 |
+
"accuracy": 0.9921023324131966,
|
59101 |
+
"total_bits": 4808064,
|
59102 |
"q_proj": {
|
59103 |
"group_size": {
|
59104 |
+
"6": 32,
|
59105 |
+
"5": 32
|
59106 |
},
|
59107 |
"bits": [
|
59108 |
+
6,
|
59109 |
+
5
|
59110 |
],
|
59111 |
"bits_prop": [
|
59112 |
0.1,
|
|
|
59116 |
},
|
59117 |
"k_proj": {
|
59118 |
"group_size": {
|
59119 |
+
"6": 32,
|
59120 |
+
"5": 32
|
59121 |
},
|
59122 |
"bits": [
|
59123 |
+
6,
|
59124 |
+
5
|
59125 |
],
|
59126 |
"bits_prop": [
|
59127 |
0.1,
|
|
|
59131 |
},
|
59132 |
"v_proj": {
|
59133 |
"group_size": {
|
59134 |
+
"6": 32
|
59135 |
},
|
59136 |
"bits": [
|
59137 |
+
6
|
59138 |
],
|
59139 |
"bits_prop": [
|
59140 |
1
|
|
|
59143 |
},
|
59144 |
"o_proj": {
|
59145 |
"group_size": {
|
59146 |
+
"6": 32,
|
59147 |
+
"5": 32
|
59148 |
},
|
59149 |
"bits": [
|
59150 |
+
6,
|
59151 |
+
5
|
59152 |
],
|
59153 |
"bits_prop": [
|
59154 |
0.1,
|
|
|
59158 |
}
|
59159 |
},
|
59160 |
"model.layers.5.mlp": {
|
59161 |
+
"accuracy": 0.9745967444032431,
|
59162 |
+
"total_bits": 11259840,
|
59163 |
"gate_proj": {
|
59164 |
"group_size": {
|
|
|
59165 |
"4": 32
|
59166 |
},
|
59167 |
"bits": [
|
|
|
59168 |
4
|
59169 |
],
|
59170 |
"bits_prop": [
|
59171 |
+
1
|
|
|
59172 |
],
|
59173 |
"scale_bits": 4
|
59174 |
},
|
59175 |
"up_proj": {
|
59176 |
"group_size": {
|
|
|
59177 |
"4": 32
|
59178 |
},
|
59179 |
"bits": [
|
|
|
59180 |
4
|
59181 |
],
|
59182 |
"bits_prop": [
|
59183 |
+
1
|
|
|
59184 |
],
|
59185 |
"scale_bits": 4
|
59186 |
},
|
59187 |
"down_proj": {
|
59188 |
"group_size": {
|
59189 |
"8": 32,
|
|
|
59190 |
"4": 32
|
59191 |
},
|
59192 |
"bits": [
|
59193 |
8,
|
|
|
59194 |
4
|
59195 |
],
|
59196 |
"bits_prop": [
|
59197 |
0.05,
|
59198 |
+
0.95
|
|
|
59199 |
],
|
59200 |
"scale_bits": 4
|
59201 |
}
|
|
|
59305 |
}
|
59306 |
},
|
59307 |
"model.layers.7.self_attn": {
|
59308 |
+
"accuracy": 0.9904757384210825,
|
59309 |
+
"total_bits": 5413824,
|
59310 |
"q_proj": {
|
59311 |
"group_size": {
|
59312 |
+
"6": 128
|
|
|
59313 |
},
|
59314 |
"bits": [
|
59315 |
+
6
|
|
|
59316 |
],
|
59317 |
"bits_prop": [
|
59318 |
+
1
|
|
|
59319 |
],
|
59320 |
"scale_bits": 4
|
59321 |
},
|
59322 |
"k_proj": {
|
59323 |
"group_size": {
|
59324 |
+
"6": 128
|
|
|
59325 |
},
|
59326 |
"bits": [
|
59327 |
+
6
|
|
|
59328 |
],
|
59329 |
"bits_prop": [
|
59330 |
+
1
|
|
|
59331 |
],
|
59332 |
"scale_bits": 4
|
59333 |
},
|
59334 |
"v_proj": {
|
59335 |
"group_size": {
|
59336 |
+
"6": 128
|
59337 |
},
|
59338 |
"bits": [
|
59339 |
6
|
|
|
59345 |
},
|
59346 |
"o_proj": {
|
59347 |
"group_size": {
|
59348 |
+
"6": 128
|
|
|
59349 |
},
|
59350 |
"bits": [
|
59351 |
+
6
|
|
|
59352 |
],
|
59353 |
"bits_prop": [
|
59354 |
+
1
|
|
|
59355 |
],
|
59356 |
"scale_bits": 4
|
59357 |
}
|
59358 |
},
|
59359 |
"model.layers.7.mlp": {
|
59360 |
+
"accuracy": 0.9878600733354688,
|
59361 |
+
"total_bits": 14294976,
|
59362 |
"gate_proj": {
|
59363 |
"group_size": {
|
59364 |
+
"6": 32,
|
59365 |
+
"5": 32
|
59366 |
},
|
59367 |
"bits": [
|
59368 |
6,
|
|
|
59376 |
},
|
59377 |
"up_proj": {
|
59378 |
"group_size": {
|
59379 |
+
"6": 32,
|
59380 |
+
"5": 32
|
59381 |
},
|
59382 |
"bits": [
|
59383 |
6,
|
|
|
59392 |
"down_proj": {
|
59393 |
"group_size": {
|
59394 |
"8": 32,
|
59395 |
+
"6": 32,
|
59396 |
+
"5": 32
|
59397 |
},
|
59398 |
"bits": [
|
59399 |
8,
|
|
|
59583 |
}
|
59584 |
},
|
59585 |
"model.layers.9.mlp": {
|
59586 |
+
"accuracy": 0.9877331163734198,
|
59587 |
+
"total_bits": 14294976,
|
59588 |
"gate_proj": {
|
59589 |
"group_size": {
|
59590 |
+
"6": 32,
|
59591 |
+
"5": 32
|
59592 |
},
|
59593 |
"bits": [
|
59594 |
+
6,
|
59595 |
+
5
|
59596 |
],
|
59597 |
"bits_prop": [
|
59598 |
+
0.1,
|
59599 |
+
0.9
|
59600 |
],
|
59601 |
"scale_bits": 4
|
59602 |
},
|
59603 |
"up_proj": {
|
59604 |
"group_size": {
|
59605 |
+
"6": 32,
|
59606 |
+
"5": 32
|
59607 |
},
|
59608 |
"bits": [
|
59609 |
+
6,
|
59610 |
+
5
|
59611 |
],
|
59612 |
"bits_prop": [
|
59613 |
+
0.25,
|
59614 |
+
0.75
|
59615 |
],
|
59616 |
"scale_bits": 4
|
59617 |
},
|
59618 |
"down_proj": {
|
59619 |
"group_size": {
|
59620 |
"8": 32,
|
59621 |
+
"6": 32,
|
59622 |
+
"5": 32
|
59623 |
},
|
59624 |
"bits": [
|
59625 |
8,
|
59626 |
+
6,
|
59627 |
+
5
|
59628 |
],
|
59629 |
"bits_prop": [
|
59630 |
0.05,
|
59631 |
+
0.1,
|
59632 |
+
0.85
|
59633 |
],
|
59634 |
"scale_bits": 4
|
59635 |
}
|
59636 |
},
|
59637 |
"model.layers.10.self_attn": {
|
59638 |
+
"accuracy": 0.9927606964483857,
|
59639 |
+
"total_bits": 4808064,
|
59640 |
"q_proj": {
|
59641 |
"group_size": {
|
59642 |
+
"6": 32,
|
59643 |
+
"5": 32
|
59644 |
},
|
59645 |
"bits": [
|
59646 |
+
6,
|
59647 |
+
5
|
59648 |
],
|
59649 |
"bits_prop": [
|
59650 |
0.1,
|
|
|
59654 |
},
|
59655 |
"k_proj": {
|
59656 |
"group_size": {
|
59657 |
+
"6": 32,
|
59658 |
+
"5": 32
|
59659 |
},
|
59660 |
"bits": [
|
59661 |
+
6,
|
59662 |
+
5
|
59663 |
],
|
59664 |
"bits_prop": [
|
59665 |
0.1,
|
|
|
59669 |
},
|
59670 |
"v_proj": {
|
59671 |
"group_size": {
|
59672 |
+
"6": 32
|
59673 |
},
|
59674 |
"bits": [
|
59675 |
+
6
|
59676 |
],
|
59677 |
"bits_prop": [
|
59678 |
1
|
|
|
59681 |
},
|
59682 |
"o_proj": {
|
59683 |
"group_size": {
|
59684 |
+
"6": 32,
|
59685 |
+
"5": 32
|
59686 |
},
|
59687 |
"bits": [
|
59688 |
+
6,
|
59689 |
+
5
|
59690 |
],
|
59691 |
"bits_prop": [
|
59692 |
0.1,
|
|
|
59696 |
}
|
59697 |
},
|
59698 |
"model.layers.10.mlp": {
|
59699 |
+
"accuracy": 0.9865744085982442,
|
59700 |
+
"total_bits": 14357424,
|
59701 |
"gate_proj": {
|
59702 |
"group_size": {
|
59703 |
+
"6": 128,
|
59704 |
+
"5": 128
|
59705 |
},
|
59706 |
"bits": [
|
59707 |
6,
|
|
|
59715 |
},
|
59716 |
"up_proj": {
|
59717 |
"group_size": {
|
59718 |
+
"6": 128,
|
59719 |
+
"5": 128
|
59720 |
},
|
59721 |
"bits": [
|
59722 |
6,
|
|
|
59731 |
"down_proj": {
|
59732 |
"group_size": {
|
59733 |
"8": 32,
|
59734 |
+
"6": 128,
|
59735 |
+
"5": 128
|
59736 |
},
|
59737 |
"bits": [
|
59738 |
8,
|
|
|
59748 |
}
|
59749 |
},
|
59750 |
"model.layers.11.self_attn": {
|
59751 |
+
"accuracy": 0.9963408834300935,
|
59752 |
+
"total_bits": 5717376,
|
59753 |
"q_proj": {
|
59754 |
"group_size": {
|
59755 |
+
"6": 32
|
|
|
59756 |
},
|
59757 |
"bits": [
|
59758 |
+
6
|
|
|
59759 |
],
|
59760 |
"bits_prop": [
|
59761 |
+
1
|
|
|
59762 |
],
|
59763 |
"scale_bits": 4
|
59764 |
},
|
59765 |
"k_proj": {
|
59766 |
"group_size": {
|
59767 |
+
"6": 32
|
|
|
59768 |
},
|
59769 |
"bits": [
|
59770 |
+
6
|
|
|
59771 |
],
|
59772 |
"bits_prop": [
|
59773 |
+
1
|
|
|
59774 |
],
|
59775 |
"scale_bits": 4
|
59776 |
},
|
59777 |
"v_proj": {
|
59778 |
"group_size": {
|
59779 |
+
"8": 32
|
59780 |
},
|
59781 |
"bits": [
|
59782 |
+
8
|
59783 |
],
|
59784 |
"bits_prop": [
|
59785 |
1
|
|
|
59788 |
},
|
59789 |
"o_proj": {
|
59790 |
"group_size": {
|
59791 |
+
"6": 32
|
|
|
59792 |
},
|
59793 |
"bits": [
|
59794 |
+
6
|
|
|
59795 |
],
|
59796 |
"bits_prop": [
|
59797 |
+
1
|
|
|
59798 |
],
|
59799 |
"scale_bits": 4
|
59800 |
}
|
59801 |
},
|
59802 |
"model.layers.11.mlp": {
|
59803 |
+
"accuracy": 0.9859139826148748,
|
59804 |
+
"total_bits": 9159600,
|
59805 |
"gate_proj": {
|
59806 |
"group_size": {
|
59807 |
+
"4": 128,
|
59808 |
+
"3": 128
|
59809 |
},
|
59810 |
"bits": [
|
59811 |
4,
|
|
|
59819 |
},
|
59820 |
"up_proj": {
|
59821 |
"group_size": {
|
59822 |
+
"4": 128,
|
59823 |
+
"3": 128
|
59824 |
},
|
59825 |
"bits": [
|
59826 |
4,
|
|
|
59835 |
"down_proj": {
|
59836 |
"group_size": {
|
59837 |
"8": 32,
|
59838 |
+
"4": 128,
|
59839 |
+
"3": 128
|
59840 |
},
|
59841 |
"bits": [
|
59842 |
8,
|
|
|
59916 |
}
|
59917 |
},
|
59918 |
"model.layers.12.mlp": {
|
59919 |
+
"accuracy": 0.9815687015652657,
|
59920 |
+
"total_bits": 7007328,
|
59921 |
"gate_proj": {
|
59922 |
"group_size": {
|
59923 |
+
"3": 64,
|
59924 |
+
"2": 64
|
59925 |
},
|
59926 |
"bits": [
|
59927 |
+
3,
|
59928 |
+
2
|
59929 |
],
|
59930 |
"bits_prop": [
|
59931 |
0.1,
|
|
|
59935 |
},
|
59936 |
"up_proj": {
|
59937 |
"group_size": {
|
59938 |
+
"3": 64,
|
59939 |
+
"2": 64
|
59940 |
},
|
59941 |
"bits": [
|
59942 |
+
3,
|
59943 |
+
2
|
59944 |
],
|
59945 |
"bits_prop": [
|
59946 |
+
0.3,
|
59947 |
+
0.7
|
59948 |
],
|
59949 |
"scale_bits": 4
|
59950 |
},
|
59951 |
"down_proj": {
|
59952 |
"group_size": {
|
59953 |
+
"5": 32,
|
59954 |
+
"3": 32
|
|
|
59955 |
},
|
59956 |
"bits": [
|
59957 |
+
5,
|
|
|
59958 |
3
|
59959 |
],
|
59960 |
"bits_prop": [
|
59961 |
0.05,
|
59962 |
+
0.95
|
|
|
59963 |
],
|
59964 |
"scale_bits": 4
|
59965 |
}
|
59966 |
},
|
59967 |
"model.layers.13.self_attn": {
|
59968 |
+
"accuracy": 0.992874032817781,
|
59969 |
+
"total_bits": 3923328,
|
59970 |
"q_proj": {
|
59971 |
"group_size": {
|
59972 |
+
"5": 32,
|
59973 |
+
"4": 32
|
59974 |
},
|
59975 |
"bits": [
|
59976 |
5,
|
|
|
59984 |
},
|
59985 |
"k_proj": {
|
59986 |
"group_size": {
|
59987 |
+
"5": 32,
|
59988 |
+
"4": 32
|
59989 |
},
|
59990 |
"bits": [
|
59991 |
5,
|
|
|
59999 |
},
|
60000 |
"v_proj": {
|
60001 |
"group_size": {
|
60002 |
+
"5": 32
|
|
|
60003 |
},
|
60004 |
"bits": [
|
60005 |
+
5
|
|
|
60006 |
],
|
60007 |
"bits_prop": [
|
60008 |
+
1
|
|
|
60009 |
],
|
60010 |
"scale_bits": 4
|
60011 |
},
|
60012 |
"o_proj": {
|
60013 |
"group_size": {
|
60014 |
+
"5": 32,
|
60015 |
+
"4": 32
|
60016 |
},
|
60017 |
"bits": [
|
60018 |
5,
|
|
|
60075 |
}
|
60076 |
},
|
60077 |
"model.layers.14.self_attn": {
|
60078 |
+
"accuracy": 0.9867758862674236,
|
60079 |
+
"total_bits": 3677040,
|
60080 |
"q_proj": {
|
60081 |
"group_size": {
|
60082 |
+
"4": 64
|
|
|
60083 |
},
|
60084 |
"bits": [
|
60085 |
+
4
|
|
|
60086 |
],
|
60087 |
"bits_prop": [
|
60088 |
+
1
|
|
|
60089 |
],
|
60090 |
"scale_bits": 4
|
60091 |
},
|
60092 |
"k_proj": {
|
60093 |
"group_size": {
|
60094 |
+
"4": 64
|
|
|
60095 |
},
|
60096 |
"bits": [
|
60097 |
+
4
|
|
|
60098 |
],
|
60099 |
"bits_prop": [
|
60100 |
+
1
|
|
|
60101 |
],
|
60102 |
"scale_bits": 4
|
60103 |
},
|
60104 |
"v_proj": {
|
60105 |
"group_size": {
|
60106 |
+
"4": 32
|
60107 |
},
|
60108 |
"bits": [
|
60109 |
+
4
|
60110 |
],
|
60111 |
"bits_prop": [
|
60112 |
1
|
|
|
60115 |
},
|
60116 |
"o_proj": {
|
60117 |
"group_size": {
|
60118 |
+
"4": 64
|
|
|
60119 |
},
|
60120 |
"bits": [
|
60121 |
+
4
|
|
|
60122 |
],
|
60123 |
"bits_prop": [
|
60124 |
+
1
|
|
|
60125 |
],
|
60126 |
"scale_bits": 4
|
60127 |
}
|
60128 |
},
|
60129 |
"model.layers.14.mlp": {
|
60130 |
+
"accuracy": 0.9892165157943964,
|
60131 |
+
"total_bits": 9097152,
|
60132 |
"gate_proj": {
|
60133 |
"group_size": {
|
60134 |
+
"4": 32,
|
60135 |
+
"3": 32
|
60136 |
},
|
60137 |
"bits": [
|
60138 |
+
4,
|
60139 |
+
3
|
60140 |
],
|
60141 |
"bits_prop": [
|
60142 |
0.1,
|
|
|
60146 |
},
|
60147 |
"up_proj": {
|
60148 |
"group_size": {
|
60149 |
+
"4": 32,
|
60150 |
+
"3": 32
|
60151 |
},
|
60152 |
"bits": [
|
60153 |
+
4,
|
60154 |
+
3
|
60155 |
],
|
60156 |
"bits_prop": [
|
60157 |
+
0.25,
|
60158 |
+
0.75
|
60159 |
],
|
60160 |
"scale_bits": 4
|
60161 |
},
|
60162 |
"down_proj": {
|
60163 |
"group_size": {
|
60164 |
+
"8": 32,
|
60165 |
+
"4": 32,
|
60166 |
"3": 32
|
60167 |
},
|
60168 |
"bits": [
|
60169 |
+
8,
|
60170 |
+
4,
|
60171 |
3
|
60172 |
],
|
60173 |
"bits_prop": [
|
60174 |
0.05,
|
60175 |
+
0.1,
|
60176 |
+
0.85
|
60177 |
],
|
60178 |
"scale_bits": 4
|
60179 |
}
|
60180 |
},
|
60181 |
"model.layers.15.self_attn": {
|
60182 |
+
"accuracy": 0.9902772186324,
|
60183 |
+
"total_bits": 3923328,
|
60184 |
"q_proj": {
|
60185 |
"group_size": {
|
60186 |
+
"5": 32,
|
60187 |
+
"4": 32
|
60188 |
},
|
60189 |
"bits": [
|
60190 |
5,
|
|
|
60198 |
},
|
60199 |
"k_proj": {
|
60200 |
"group_size": {
|
60201 |
+
"5": 32,
|
60202 |
+
"4": 32
|
60203 |
},
|
60204 |
"bits": [
|
60205 |
5,
|
|
|
60213 |
},
|
60214 |
"v_proj": {
|
60215 |
"group_size": {
|
60216 |
+
"5": 32
|
|
|
60217 |
},
|
60218 |
"bits": [
|
60219 |
+
5
|
|
|
60220 |
],
|
60221 |
"bits_prop": [
|
60222 |
+
1
|
|
|
60223 |
],
|
60224 |
"scale_bits": 4
|
60225 |
},
|
60226 |
"o_proj": {
|
60227 |
"group_size": {
|
60228 |
+
"5": 32,
|
60229 |
+
"4": 32
|
60230 |
},
|
60231 |
"bits": [
|
60232 |
5,
|
|
|
60240 |
}
|
60241 |
},
|
60242 |
"model.layers.15.mlp": {
|
60243 |
+
"accuracy": 0.9878587136045098,
|
60244 |
+
"total_bits": 9159600,
|
60245 |
"gate_proj": {
|
60246 |
"group_size": {
|
60247 |
+
"4": 128,
|
60248 |
+
"3": 128
|
60249 |
},
|
60250 |
"bits": [
|
60251 |
+
4,
|
60252 |
+
3
|
60253 |
],
|
60254 |
"bits_prop": [
|
60255 |
0.1,
|
|
|
60259 |
},
|
60260 |
"up_proj": {
|
60261 |
"group_size": {
|
60262 |
+
"4": 128,
|
60263 |
+
"3": 128
|
60264 |
},
|
60265 |
"bits": [
|
60266 |
+
4,
|
60267 |
+
3
|
60268 |
],
|
60269 |
"bits_prop": [
|
60270 |
+
0.25,
|
60271 |
+
0.75
|
60272 |
],
|
60273 |
"scale_bits": 4
|
60274 |
},
|
60275 |
"down_proj": {
|
60276 |
"group_size": {
|
60277 |
+
"8": 32,
|
60278 |
+
"4": 128,
|
60279 |
+
"3": 128
|
60280 |
},
|
60281 |
"bits": [
|
60282 |
+
8,
|
60283 |
+
4,
|
60284 |
3
|
60285 |
],
|
60286 |
"bits_prop": [
|
60287 |
0.05,
|
60288 |
+
0.1,
|
60289 |
+
0.85
|
60290 |
],
|
60291 |
"scale_bits": 4
|
60292 |
}
|
60293 |
},
|
60294 |
"model.layers.16.self_attn": {
|
60295 |
+
"accuracy": 0.9928999468684196,
|
60296 |
+
"total_bits": 3775344,
|
60297 |
"q_proj": {
|
60298 |
"group_size": {
|
60299 |
+
"5": 64,
|
60300 |
+
"4": 64
|
60301 |
},
|
60302 |
"bits": [
|
60303 |
+
5,
|
60304 |
+
4
|
60305 |
],
|
60306 |
"bits_prop": [
|
60307 |
0.1,
|
|
|
60311 |
},
|
60312 |
"k_proj": {
|
60313 |
"group_size": {
|
60314 |
+
"5": 64,
|
60315 |
+
"4": 64
|
60316 |
},
|
60317 |
"bits": [
|
60318 |
+
5,
|
60319 |
+
4
|
60320 |
],
|
60321 |
"bits_prop": [
|
60322 |
0.1,
|
|
|
60326 |
},
|
60327 |
"v_proj": {
|
60328 |
"group_size": {
|
60329 |
+
"5": 32,
|
60330 |
+
"4": 32
|
60331 |
},
|
60332 |
"bits": [
|
60333 |
+
5,
|
60334 |
+
4
|
60335 |
],
|
60336 |
"bits_prop": [
|
60337 |
+
0.1,
|
60338 |
+
0.9
|
60339 |
],
|
60340 |
"scale_bits": 4
|
60341 |
},
|
60342 |
"o_proj": {
|
60343 |
"group_size": {
|
60344 |
+
"5": 64,
|
60345 |
+
"4": 64
|
60346 |
},
|
60347 |
"bits": [
|
60348 |
+
5,
|
60349 |
+
4
|
60350 |
],
|
60351 |
"bits_prop": [
|
60352 |
0.1,
|
|
|
60356 |
}
|
60357 |
},
|
60358 |
"model.layers.16.mlp": {
|
60359 |
+
"accuracy": 0.9799803365021944,
|
60360 |
+
"total_bits": 7007328,
|
60361 |
"gate_proj": {
|
60362 |
"group_size": {
|
60363 |
+
"3": 64,
|
60364 |
+
"2": 64
|
60365 |
},
|
60366 |
"bits": [
|
60367 |
+
3,
|
60368 |
+
2
|
60369 |
],
|
60370 |
"bits_prop": [
|
60371 |
0.1,
|
|
|
60375 |
},
|
60376 |
"up_proj": {
|
60377 |
"group_size": {
|
60378 |
+
"3": 64,
|
60379 |
+
"2": 64
|
60380 |
},
|
60381 |
"bits": [
|
60382 |
+
3,
|
60383 |
+
2
|
60384 |
],
|
60385 |
"bits_prop": [
|
60386 |
+
0.3,
|
60387 |
+
0.7
|
60388 |
],
|
60389 |
"scale_bits": 4
|
60390 |
},
|
60391 |
"down_proj": {
|
60392 |
"group_size": {
|
60393 |
+
"5": 32,
|
|
|
60394 |
"3": 32
|
60395 |
},
|
60396 |
"bits": [
|
60397 |
+
5,
|
|
|
60398 |
3
|
60399 |
],
|
60400 |
"bits_prop": [
|
60401 |
0.05,
|
60402 |
+
0.95
|
|
|
60403 |
],
|
60404 |
"scale_bits": 4
|
60405 |
}
|
60406 |
},
|
60407 |
"model.layers.17.self_attn": {
|
60408 |
+
"accuracy": 0.9908454418182373,
|
60409 |
+
"total_bits": 3866304,
|
60410 |
"q_proj": {
|
60411 |
"group_size": {
|
60412 |
"5": 64,
|
|
|
60439 |
},
|
60440 |
"v_proj": {
|
60441 |
"group_size": {
|
60442 |
+
"5": 64
|
|
|
60443 |
},
|
60444 |
"bits": [
|
60445 |
+
5
|
|
|
60446 |
],
|
60447 |
"bits_prop": [
|
60448 |
+
1
|
|
|
60449 |
],
|
60450 |
"scale_bits": 4
|
60451 |
},
|
|
|
60466 |
}
|
60467 |
},
|
60468 |
"model.layers.17.mlp": {
|
60469 |
+
"accuracy": 0.9788082204759121,
|
60470 |
+
"total_bits": 7007328,
|
60471 |
"gate_proj": {
|
60472 |
"group_size": {
|
60473 |
+
"3": 64,
|
60474 |
+
"2": 64
|
60475 |
},
|
60476 |
"bits": [
|
60477 |
+
3,
|
60478 |
+
2
|
60479 |
],
|
60480 |
"bits_prop": [
|
60481 |
0.1,
|
|
|
60485 |
},
|
60486 |
"up_proj": {
|
60487 |
"group_size": {
|
60488 |
+
"3": 64,
|
60489 |
+
"2": 64
|
60490 |
},
|
60491 |
"bits": [
|
60492 |
+
3,
|
60493 |
+
2
|
60494 |
],
|
60495 |
"bits_prop": [
|
60496 |
+
0.3,
|
60497 |
+
0.7
|
60498 |
],
|
60499 |
"scale_bits": 4
|
60500 |
},
|
60501 |
"down_proj": {
|
60502 |
"group_size": {
|
60503 |
+
"5": 32,
|
|
|
60504 |
"3": 32
|
60505 |
},
|
60506 |
"bits": [
|
60507 |
+
5,
|
|
|
60508 |
3
|
60509 |
],
|
60510 |
"bits_prop": [
|
60511 |
0.05,
|
60512 |
+
0.95
|
|
|
60513 |
],
|
60514 |
"scale_bits": 4
|
60515 |
}
|
|
|
60692 |
}
|
60693 |
},
|
60694 |
"model.layers.19.mlp": {
|
60695 |
+
"accuracy": 0.9835127778351307,
|
60696 |
+
"total_bits": 9097152,
|
60697 |
"gate_proj": {
|
60698 |
"group_size": {
|
60699 |
"4": 32,
|
|
|
60727 |
"down_proj": {
|
60728 |
"group_size": {
|
60729 |
"8": 32,
|
60730 |
+
"4": 32,
|
60731 |
+
"3": 32
|
60732 |
},
|
60733 |
"bits": [
|
60734 |
8,
|
60735 |
+
4,
|
60736 |
+
3
|
60737 |
],
|
60738 |
"bits_prop": [
|
60739 |
0.05,
|
60740 |
+
0.1,
|
60741 |
+
0.85
|
60742 |
],
|
60743 |
"scale_bits": 4
|
60744 |
}
|
60745 |
},
|
60746 |
"model.layers.20.self_attn": {
|
60747 |
+
"accuracy": 0.9949732795357704,
|
60748 |
+
"total_bits": 4808064,
|
60749 |
"q_proj": {
|
60750 |
"group_size": {
|
60751 |
+
"6": 32,
|
60752 |
+
"5": 32
|
60753 |
},
|
60754 |
"bits": [
|
60755 |
+
6,
|
60756 |
+
5
|
60757 |
],
|
60758 |
"bits_prop": [
|
60759 |
0.1,
|
|
|
60763 |
},
|
60764 |
"k_proj": {
|
60765 |
"group_size": {
|
60766 |
+
"6": 32,
|
60767 |
+
"5": 32
|
60768 |
},
|
60769 |
"bits": [
|
60770 |
+
6,
|
60771 |
+
5
|
60772 |
],
|
60773 |
"bits_prop": [
|
60774 |
0.1,
|
|
|
60778 |
},
|
60779 |
"v_proj": {
|
60780 |
"group_size": {
|
60781 |
+
"6": 32
|
60782 |
},
|
60783 |
"bits": [
|
60784 |
+
6
|
60785 |
],
|
60786 |
"bits_prop": [
|
60787 |
1
|
|
|
60790 |
},
|
60791 |
"o_proj": {
|
60792 |
"group_size": {
|
60793 |
+
"6": 32,
|
60794 |
+
"5": 32
|
60795 |
},
|
60796 |
"bits": [
|
60797 |
+
6,
|
60798 |
+
5
|
60799 |
],
|
60800 |
"bits_prop": [
|
60801 |
0.1,
|
|
|
60805 |
}
|
60806 |
},
|
60807 |
"model.layers.20.mlp": {
|
60808 |
+
"accuracy": 0.9870169991627336,
|
60809 |
+
"total_bits": 11101728,
|
60810 |
"gate_proj": {
|
60811 |
"group_size": {
|
60812 |
+
"4": 128
|
60813 |
},
|
60814 |
"bits": [
|
60815 |
4
|
|
|
60834 |
"down_proj": {
|
60835 |
"group_size": {
|
60836 |
"8": 32,
|
60837 |
+
"4": 128
|
60838 |
},
|
60839 |
"bits": [
|
60840 |
8,
|
|
|
60848 |
}
|
60849 |
},
|
60850 |
"model.layers.21.self_attn": {
|
60851 |
+
"accuracy": 0.9899369664490223,
|
60852 |
+
"total_bits": 3775344,
|
60853 |
"q_proj": {
|
60854 |
"group_size": {
|
60855 |
"5": 64,
|
|
|
60882 |
},
|
60883 |
"v_proj": {
|
60884 |
"group_size": {
|
60885 |
+
"5": 32,
|
60886 |
+
"4": 32
|
60887 |
},
|
60888 |
"bits": [
|
60889 |
+
5,
|
60890 |
+
4
|
60891 |
],
|
60892 |
"bits_prop": [
|
60893 |
+
0.1,
|
60894 |
+
0.9
|
60895 |
],
|
60896 |
"scale_bits": 4
|
60897 |
},
|
|
|
60912 |
}
|
60913 |
},
|
60914 |
"model.layers.21.mlp": {
|
60915 |
+
"accuracy": 0.9894305132329464,
|
60916 |
+
"total_bits": 11259840,
|
60917 |
"gate_proj": {
|
60918 |
"group_size": {
|
60919 |
+
"4": 32
|
60920 |
},
|
60921 |
"bits": [
|
60922 |
4
|
|
|
60941 |
"down_proj": {
|
60942 |
"group_size": {
|
60943 |
"8": 32,
|
60944 |
+
"4": 32
|
60945 |
},
|
60946 |
"bits": [
|
60947 |
8,
|
|
|
60955 |
}
|
60956 |
},
|
60957 |
"model.layers.22.self_attn": {
|
60958 |
+
"accuracy": 0.9895689664408565,
|
60959 |
+
"total_bits": 3775344,
|
60960 |
"q_proj": {
|
60961 |
"group_size": {
|
60962 |
+
"5": 64,
|
60963 |
+
"4": 64
|
60964 |
},
|
60965 |
"bits": [
|
60966 |
5,
|
|
|
60973 |
"scale_bits": 4
|
60974 |
},
|
60975 |
"k_proj": {
|
60976 |
+
"group_size": {
|
60977 |
+
"5": 64,
|
60978 |
+
"4": 64
|
60979 |
},
|
60980 |
"bits": [
|
60981 |
5,
|
|
|
60989 |
},
|
60990 |
"v_proj": {
|
60991 |
"group_size": {
|
60992 |
+
"5": 32,
|
60993 |
+
"4": 32
|
60994 |
},
|
60995 |
"bits": [
|
60996 |
5,
|
|
|
61004 |
},
|
61005 |
"o_proj": {
|
61006 |
"group_size": {
|
61007 |
+
"5": 64,
|
61008 |
+
"4": 64
|
61009 |
},
|
61010 |
"bits": [
|
61011 |
5,
|
|
|
61062 |
}
|
61063 |
},
|
61064 |
"model.layers.23.self_attn": {
|
61065 |
+
"accuracy": 0.9891995070502162,
|
61066 |
+
"total_bits": 4808064,
|
61067 |
"q_proj": {
|
61068 |
"group_size": {
|
61069 |
+
"6": 32,
|
61070 |
+
"5": 32
|
61071 |
},
|
61072 |
"bits": [
|
61073 |
+
6,
|
61074 |
+
5
|
61075 |
],
|
61076 |
"bits_prop": [
|
61077 |
+
0.1,
|
61078 |
+
0.9
|
61079 |
],
|
61080 |
"scale_bits": 4
|
61081 |
},
|
61082 |
"k_proj": {
|
61083 |
"group_size": {
|
61084 |
+
"6": 32,
|
61085 |
+
"5": 32
|
61086 |
},
|
61087 |
"bits": [
|
61088 |
+
6,
|
61089 |
+
5
|
61090 |
],
|
61091 |
"bits_prop": [
|
61092 |
+
0.1,
|
61093 |
+
0.9
|
61094 |
],
|
61095 |
"scale_bits": 4
|
61096 |
},
|
61097 |
"v_proj": {
|
61098 |
"group_size": {
|
61099 |
+
"6": 32
|
61100 |
},
|
61101 |
"bits": [
|
61102 |
+
6
|
61103 |
],
|
61104 |
"bits_prop": [
|
61105 |
1
|
|
|
61108 |
},
|
61109 |
"o_proj": {
|
61110 |
"group_size": {
|
61111 |
+
"6": 32,
|
61112 |
+
"5": 32
|
61113 |
},
|
61114 |
"bits": [
|
61115 |
+
6,
|
61116 |
+
5
|
61117 |
],
|
61118 |
"bits_prop": [
|
61119 |
+
0.1,
|
61120 |
+
0.9
|
61121 |
],
|
61122 |
"scale_bits": 4
|
61123 |
}
|
|
|
61166 |
}
|
61167 |
},
|
61168 |
"model.layers.24.self_attn": {
|
61169 |
+
"accuracy": 0.990554129704833,
|
61170 |
+
"total_bits": 5413824,
|
61171 |
"q_proj": {
|
61172 |
"group_size": {
|
61173 |
+
"6": 128
|
61174 |
},
|
61175 |
"bits": [
|
61176 |
6
|
|
|
61182 |
},
|
61183 |
"k_proj": {
|
61184 |
"group_size": {
|
61185 |
+
"6": 128
|
61186 |
},
|
61187 |
"bits": [
|
61188 |
6
|
|
|
61194 |
},
|
61195 |
"v_proj": {
|
61196 |
"group_size": {
|
61197 |
+
"6": 128
|
61198 |
},
|
61199 |
"bits": [
|
61200 |
+
6
|
61201 |
],
|
61202 |
"bits_prop": [
|
61203 |
1
|
|
|
61206 |
},
|
61207 |
"o_proj": {
|
61208 |
"group_size": {
|
61209 |
+
"6": 128
|
61210 |
},
|
61211 |
"bits": [
|
61212 |
6
|
|
|
61218 |
}
|
61219 |
},
|
61220 |
"model.layers.24.mlp": {
|
61221 |
+
"accuracy": 0.9851180734112859,
|
61222 |
+
"total_bits": 11758512,
|
61223 |
"gate_proj": {
|
61224 |
"group_size": {
|
61225 |
+
"5": 128,
|
61226 |
+
"4": 128
|
61227 |
},
|
61228 |
"bits": [
|
61229 |
+
5,
|
61230 |
4
|
61231 |
],
|
61232 |
"bits_prop": [
|
61233 |
+
0.1,
|
61234 |
+
0.9
|
61235 |
],
|
61236 |
"scale_bits": 4
|
61237 |
},
|
61238 |
"up_proj": {
|
61239 |
"group_size": {
|
61240 |
+
"5": 128,
|
61241 |
+
"4": 128
|
61242 |
},
|
61243 |
"bits": [
|
61244 |
+
5,
|
61245 |
4
|
61246 |
],
|
61247 |
"bits_prop": [
|
61248 |
+
0.25,
|
61249 |
+
0.75
|
61250 |
],
|
61251 |
"scale_bits": 4
|
61252 |
},
|
61253 |
"down_proj": {
|
61254 |
"group_size": {
|
61255 |
"8": 32,
|
61256 |
+
"5": 128,
|
61257 |
+
"4": 128
|
61258 |
},
|
61259 |
"bits": [
|
61260 |
8,
|
61261 |
+
5,
|
61262 |
4
|
61263 |
],
|
61264 |
"bits_prop": [
|
61265 |
0.05,
|
61266 |
+
0.1,
|
61267 |
+
0.85
|
61268 |
],
|
61269 |
"scale_bits": 4
|
61270 |
}
|
61271 |
},
|
61272 |
"model.layers.25.self_attn": {
|
61273 |
+
"accuracy": 0.9908542837947607,
|
61274 |
+
"total_bits": 3775344,
|
61275 |
"q_proj": {
|
61276 |
"group_size": {
|
61277 |
+
"5": 64,
|
61278 |
+
"4": 64
|
61279 |
},
|
61280 |
"bits": [
|
61281 |
+
5,
|
61282 |
+
4
|
61283 |
],
|
61284 |
"bits_prop": [
|
61285 |
0.1,
|
|
|
61289 |
},
|
61290 |
"k_proj": {
|
61291 |
"group_size": {
|
61292 |
+
"5": 64,
|
61293 |
+
"4": 64
|
61294 |
},
|
61295 |
"bits": [
|
61296 |
+
5,
|
61297 |
+
4
|
61298 |
],
|
61299 |
"bits_prop": [
|
61300 |
0.1,
|
|
|
61304 |
},
|
61305 |
"v_proj": {
|
61306 |
"group_size": {
|
61307 |
+
"5": 32,
|
61308 |
+
"4": 32
|
61309 |
},
|
61310 |
"bits": [
|
61311 |
+
5,
|
61312 |
+
4
|
61313 |
],
|
61314 |
"bits_prop": [
|
61315 |
+
0.1,
|
61316 |
+
0.9
|
61317 |
],
|
61318 |
"scale_bits": 4
|
61319 |
},
|
61320 |
"o_proj": {
|
61321 |
"group_size": {
|
61322 |
+
"5": 64,
|
61323 |
+
"4": 64
|
61324 |
},
|
61325 |
"bits": [
|
61326 |
+
5,
|
61327 |
+
4
|
61328 |
],
|
61329 |
"bits_prop": [
|
61330 |
0.1,
|
|
|
61334 |
}
|
61335 |
},
|
61336 |
"model.layers.25.mlp": {
|
61337 |
+
"accuracy": 0.9851860506460071,
|
61338 |
+
"total_bits": 11259840,
|
61339 |
"gate_proj": {
|
61340 |
"group_size": {
|
61341 |
+
"4": 32
|
61342 |
},
|
61343 |
"bits": [
|
61344 |
4
|
|
|
61363 |
"down_proj": {
|
61364 |
"group_size": {
|
61365 |
"8": 32,
|
61366 |
+
"4": 32
|
61367 |
},
|
61368 |
"bits": [
|
61369 |
8,
|
|
|
61377 |
}
|
61378 |
},
|
61379 |
"model.layers.26.self_attn": {
|
61380 |
+
"accuracy": 0.9889037329703569,
|
61381 |
+
"total_bits": 3775344,
|
61382 |
"q_proj": {
|
61383 |
"group_size": {
|
61384 |
+
"5": 64,
|
61385 |
+
"4": 64
|
61386 |
},
|
61387 |
"bits": [
|
61388 |
5,
|
|
|
61396 |
},
|
61397 |
"k_proj": {
|
61398 |
"group_size": {
|
61399 |
+
"5": 64,
|
61400 |
+
"4": 64
|
61401 |
},
|
61402 |
"bits": [
|
61403 |
5,
|
|
|
61411 |
},
|
61412 |
"v_proj": {
|
61413 |
"group_size": {
|
61414 |
+
"5": 32,
|
61415 |
+
"4": 32
|
61416 |
},
|
61417 |
"bits": [
|
61418 |
+
5,
|
61419 |
+
4
|
61420 |
],
|
61421 |
"bits_prop": [
|
61422 |
+
0.1,
|
61423 |
+
0.9
|
61424 |
],
|
61425 |
"scale_bits": 4
|
61426 |
},
|
61427 |
"o_proj": {
|
61428 |
"group_size": {
|
61429 |
+
"5": 64,
|
61430 |
+
"4": 64
|
61431 |
},
|
61432 |
"bits": [
|
61433 |
5,
|
|
|
61493 |
}
|
61494 |
},
|
61495 |
"model.layers.27.self_attn": {
|
61496 |
+
"accuracy": 0.9844594039022923,
|
61497 |
+
"total_bits": 3866304,
|
61498 |
"q_proj": {
|
61499 |
"group_size": {
|
61500 |
+
"5": 64,
|
61501 |
+
"4": 64
|
61502 |
},
|
61503 |
"bits": [
|
61504 |
+
5,
|
61505 |
+
4
|
61506 |
],
|
61507 |
"bits_prop": [
|
61508 |
+
0.1,
|
61509 |
+
0.9
|
61510 |
],
|
61511 |
"scale_bits": 4
|
61512 |
},
|
61513 |
"k_proj": {
|
61514 |
"group_size": {
|
61515 |
+
"5": 64,
|
61516 |
+
"4": 64
|
61517 |
},
|
61518 |
"bits": [
|
61519 |
+
5,
|
61520 |
+
4
|
61521 |
],
|
61522 |
"bits_prop": [
|
61523 |
+
0.1,
|
61524 |
+
0.9
|
61525 |
],
|
61526 |
"scale_bits": 4
|
61527 |
},
|
61528 |
"v_proj": {
|
61529 |
"group_size": {
|
61530 |
+
"5": 64
|
61531 |
},
|
61532 |
"bits": [
|
61533 |
+
5
|
61534 |
],
|
61535 |
"bits_prop": [
|
61536 |
1
|
|
|
61539 |
},
|
61540 |
"o_proj": {
|
61541 |
"group_size": {
|
61542 |
+
"5": 64,
|
61543 |
+
"4": 64
|
61544 |
},
|
61545 |
"bits": [
|
61546 |
+
5,
|
61547 |
+
4
|
61548 |
],
|
61549 |
"bits_prop": [
|
61550 |
+
0.1,
|
61551 |
+
0.9
|
61552 |
],
|
61553 |
"scale_bits": 4
|
61554 |
}
|
61555 |
},
|
61556 |
"model.layers.27.mlp": {
|
61557 |
+
"accuracy": 0.9841680098325014,
|
61558 |
+
"total_bits": 11696064,
|
61559 |
"gate_proj": {
|
61560 |
"group_size": {
|
61561 |
+
"5": 32,
|
61562 |
+
"4": 32
|
61563 |
},
|
61564 |
"bits": [
|
61565 |
+
5,
|
61566 |
4
|
61567 |
],
|
61568 |
"bits_prop": [
|
61569 |
+
0.1,
|
61570 |
+
0.9
|
61571 |
],
|
61572 |
"scale_bits": 4
|
61573 |
},
|
61574 |
"up_proj": {
|
61575 |
"group_size": {
|
61576 |
+
"5": 32,
|
61577 |
"4": 32
|
61578 |
},
|
61579 |
"bits": [
|
61580 |
+
5,
|
61581 |
4
|
61582 |
],
|
61583 |
"bits_prop": [
|
61584 |
+
0.25,
|
61585 |
+
0.75
|
61586 |
],
|
61587 |
"scale_bits": 4
|
61588 |
},
|
61589 |
"down_proj": {
|
61590 |
"group_size": {
|
61591 |
"8": 32,
|
61592 |
+
"5": 32,
|
61593 |
+
"4": 32
|
61594 |
},
|
61595 |
"bits": [
|
61596 |
8,
|
61597 |
+
5,
|
61598 |
4
|
61599 |
],
|
61600 |
"bits_prop": [
|
61601 |
0.05,
|
61602 |
+
0.1,
|
61603 |
+
0.85
|
61604 |
],
|
61605 |
"scale_bits": 4
|
61606 |
}
|
61607 |
},
|
61608 |
"model.layers.28.self_attn": {
|
61609 |
+
"accuracy": 0.9928587442263961,
|
61610 |
+
"total_bits": 4811712,
|
61611 |
"q_proj": {
|
61612 |
"group_size": {
|
61613 |
+
"6": 128,
|
61614 |
+
"5": 128
|
61615 |
},
|
61616 |
"bits": [
|
61617 |
+
6,
|
61618 |
+
5
|
61619 |
],
|
61620 |
"bits_prop": [
|
61621 |
0.1,
|
|
|
61625 |
},
|
61626 |
"k_proj": {
|
61627 |
"group_size": {
|
61628 |
+
"6": 128,
|
61629 |
+
"5": 128
|
61630 |
},
|
61631 |
"bits": [
|
61632 |
+
6,
|
61633 |
+
5
|
61634 |
],
|
61635 |
"bits_prop": [
|
61636 |
0.1,
|
|
|
61640 |
},
|
61641 |
"v_proj": {
|
61642 |
"group_size": {
|
61643 |
+
"6": 128
|
61644 |
},
|
61645 |
"bits": [
|
61646 |
+
6
|
61647 |
],
|
61648 |
"bits_prop": [
|
61649 |
1
|
|
|
61652 |
},
|
61653 |
"o_proj": {
|
61654 |
"group_size": {
|
61655 |
+
"6": 128,
|
61656 |
+
"5": 128
|
61657 |
},
|
61658 |
"bits": [
|
61659 |
+
6,
|
61660 |
+
5
|
61661 |
],
|
61662 |
"bits_prop": [
|
61663 |
0.1,
|
|
|
61719 |
}
|
61720 |
},
|
61721 |
"model.layers.29.self_attn": {
|
61722 |
+
"accuracy": 0.9930798751302063,
|
61723 |
+
"total_bits": 4811712,
|
61724 |
"q_proj": {
|
61725 |
"group_size": {
|
61726 |
+
"6": 128,
|
61727 |
+
"5": 128
|
61728 |
},
|
61729 |
"bits": [
|
61730 |
+
6,
|
61731 |
+
5
|
61732 |
],
|
61733 |
"bits_prop": [
|
61734 |
0.1,
|
|
|
61738 |
},
|
61739 |
"k_proj": {
|
61740 |
"group_size": {
|
61741 |
+
"6": 128,
|
61742 |
+
"5": 128
|
61743 |
},
|
61744 |
"bits": [
|
61745 |
+
6,
|
61746 |
+
5
|
61747 |
],
|
61748 |
"bits_prop": [
|
61749 |
0.1,
|
|
|
61753 |
},
|
61754 |
"v_proj": {
|
61755 |
"group_size": {
|
61756 |
+
"6": 128
|
61757 |
},
|
61758 |
"bits": [
|
61759 |
+
6
|
61760 |
],
|
61761 |
"bits_prop": [
|
61762 |
1
|
|
|
61765 |
},
|
61766 |
"o_proj": {
|
61767 |
"group_size": {
|
61768 |
+
"6": 128,
|
61769 |
+
"5": 128
|
61770 |
},
|
61771 |
"bits": [
|
61772 |
+
6,
|
61773 |
+
5
|
61774 |
],
|
61775 |
"bits_prop": [
|
61776 |
0.1,
|
measurement.json
CHANGED
@@ -57957,7 +57957,7 @@
|
|
57957 |
}
|
57958 |
},
|
57959 |
{
|
57960 |
-
"accuracy": 0.
|
57961 |
"total_bits": 9097152,
|
57962 |
"gate_proj": {
|
57963 |
"group_size": {
|
|
|
57957 |
}
|
57958 |
},
|
57959 |
{
|
57960 |
+
"accuracy": 0.945903766900301,
|
57961 |
"total_bits": 9097152,
|
57962 |
"gate_proj": {
|
57963 |
"group_size": {
|
out_tensor/lm_head.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 22932166
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b00e9a9ad58653c8b753d64bc96345482aeb377528d8a49a22664021233fe51
|
3 |
size 22932166
|
out_tensor/model.layers.1.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:122e426acb718ec8500b571a642bc6798b45e486dab0e383596af8cd42bc9b6c
|
3 |
+
size 502304
|
out_tensor/model.layers.1.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:89cbc78f723f7c5f29a38cb9d42ae84c09a7332f9b6984869ff801409e97fd4a
|
3 |
+
size 471404
|
out_tensor/model.layers.1.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:acbc58437663f0da9c0a7310111436b0c35546cfedffce7f76c5c44e967ff095
|
3 |
+
size 489828
|
out_tensor/model.layers.1.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e1443175687f038577cbe5062f7fa9b9954b66111bde7a407efb50391040dc1a
|
3 |
+
size 61702
|
out_tensor/model.layers.1.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:43634971f137ca8d8fd563c564f2bcc0842893c013811019932debc28b3d00a0
|
3 |
+
size 179398
|
out_tensor/model.layers.1.self_attn.q_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bdcbbb552d31497058166738d8966175d730c296fb7c454cb10b975e8164fca9
|
3 |
+
size 179398
|
out_tensor/model.layers.1.self_attn.v_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ba2e26e21b021c7497dc7ba363a1a70c8a303c7c9737a5a6c4ba4d4eb26fe28a
|
3 |
+
size 60574
|
out_tensor/model.layers.10.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5c358e71b6a5272eec2d6a732b7b380d5ee7eedf455156b908e3be611d17c4c2
|
3 |
+
size 603202
|
out_tensor/model.layers.10.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fa1b81408ba9e82a77eef6916698df21c2dab7d12994f61609052e81e32478ad
|
3 |
+
size 584222
|
out_tensor/model.layers.10.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c97dc4ce1700dd6567bb94db216c3fad52f90b46c5c46a4699f3175c226cc40e
|
3 |
+
size 608790
|
out_tensor/model.layers.10.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b108b75252b492881465e9a18807ef9db35f89219728b81082ec7d052f014ac3
|
3 |
+
size 75324
|
out_tensor/model.layers.10.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:91e1efa7d78ad6b225042ae5c222f822d4fddfe77618d0cba3142c8868aa16c3
|
3 |
+
size 220092
|
out_tensor/model.layers.10.self_attn.q_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:55391444fa5d42d80e7a36478be427cf84f47fb2d7f3e057cbd3a4bb25a03708
|
3 |
+
size 220092
|
out_tensor/model.layers.10.self_attn.v_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2991ae45c537c35acbc7c90b65a56beeb96ed7ee04e26d7b6351f7ac4ae91487
|
3 |
+
size 87612
|
out_tensor/model.layers.11.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7a4774d491065db90b70e619e961fb32e7daa1e3a49ca79d7f48b7fb9a192b1f
|
3 |
+
size 395842
|
out_tensor/model.layers.11.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:67db48f5754d31acf835016b4104cf11cb24377d2180e3583f73c2f82b0acad0
|
3 |
+
size 363038
|
out_tensor/model.layers.11.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cda2c19e5ad0b12b1209c3366676220530ffbe3632341ce55e2a5af4e991a265
|
3 |
+
size 387606
|
out_tensor/model.layers.11.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:67f6a5bba9db03f095850355740399329783f04bcc7f9156e0d7dac0d81d932e
|
3 |
+
size 87612
|
out_tensor/model.layers.11.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0194a099426cb0bca691fd745deef804210e613abf061d8ca92b5845be6a389b
|
3 |
+
size 256956
|
out_tensor/model.layers.11.self_attn.q_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f8076ee358670d24c330e78f9afce938888b911505d8ee3460dcfc0c7a14f16a
|
3 |
+
size 256956
|
out_tensor/model.layers.11.self_attn.v_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:31af5f20948529d81afd585e97530db8c60144fed28093fb8f5db5d407b57d85
|
3 |
+
size 115260
|
out_tensor/model.layers.12.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ed46b77c0bc2988c313c255091c65acc8cc59a8d05ec50b9aa5f63ed58e73ac2
|
3 |
+
size 366376
|
out_tensor/model.layers.12.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b493b4a727dba8ac8b8230f0eaefce9a4dbfda010828a0154d5914b46d6e521f
|
3 |
+
size 243254
|
out_tensor/model.layers.12.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:871fb605c1b775b7e2d02d90b176e00b020fc2ca32ea0876ae7495d415b7a4f4
|
3 |
+
size 267822
|
out_tensor/model.layers.12.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 47878
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:53f10533ac283494d2c2bbe892ca4dc7d5073c1c12b4568e1814de94dd1ff96d
|
3 |
size 47878
|
out_tensor/model.layers.12.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 137934
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:99755f5f45d5d503853951ce3d3d2edebdc7b2493d6acd71db123a63c4467c1d
|
3 |
size 137934
|
out_tensor/model.layers.12.self_attn.q_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 137934
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3eaa8f4fa56d69177cb934fc01aa651b981285d39c125bef8fa66674608fab8c
|
3 |
size 137934
|
out_tensor/model.layers.12.self_attn.v_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 47878
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6f48ea439ba50a51c0d91c08de3fd6e9b76ec47a4dd59f30863dae0072375463
|
3 |
size 47878
|
out_tensor/model.layers.13.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 366376
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:737d63835c5aca628c5a7e321483998228baa6785801a116eee92bcf91ebb13a
|
3 |
size 366376
|
out_tensor/model.layers.13.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 243254
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ff3791d92088b0ca1c4301061f204b8e3770c9f1ced43c9ddb517a1ab1d743ed
|
3 |
size 243254
|
out_tensor/model.layers.13.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 267822
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:09ebef69a690bf0c60ff790aea58aa59d30a1381c677366978519de7ff004b41
|
3 |
size 267822
|
out_tensor/model.layers.13.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fae7e013ba28b81c3855c1081d39965240b57e748411513785d6a1da13213455
|
3 |
+
size 61500
|
out_tensor/model.layers.13.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:da1189fdc505da691e74d795e5cfd7414f9f130272c37a90268a03507ef54dd3
|
3 |
+
size 178620
|
out_tensor/model.layers.13.self_attn.q_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f9acee0cc9d7b56d3fe6170865b58a18d13234897678522dccd47ed0268364f6
|
3 |
+
size 178620
|
out_tensor/model.layers.13.self_attn.v_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:028771ce4245f1bbd1cc1e491e8b347294794dbad7b6e886d2c42d555109b6bf
|
3 |
+
size 73788
|
out_tensor/model.layers.14.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:75f2ef701d3132beb3dd02868cbd0d78324bffb99630e317b8fe2ce36881af39
|
3 |
+
size 398632
|
out_tensor/model.layers.14.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b1b5396c1ababf6b1dcfa47dee5201903a5c3fce4e0bd371a0ae8af2ca0346bc
|
3 |
+
size 360820
|
out_tensor/model.layers.14.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2c92070124ca1ad26746aa06bd01b33d417b168299fc53de3d9a78dcc612b3e7
|
3 |
+
size 379244
|
out_tensor/model.layers.14.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ceddc15d97327cc721f0d0a618f706eca97d5681e842c6cae7f9044aa69337db
|
3 |
+
size 59038
|
out_tensor/model.layers.14.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8390a3cd72112867844755fa9873aba89a3eadf1b1c943f0cac917df14840b6e
|
3 |
+
size 171366
|
out_tensor/model.layers.14.self_attn.q_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:88ef8213b80f02857314bbdbf8f9179d0673359b17ab3e8ae939833010cae44b
|
3 |
+
size 171366
|
out_tensor/model.layers.14.self_attn.v_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f1b8da419b6b51187373d5a25600ea36e54ff0df8de806549707493dd782bfaa
|
3 |
+
size 59964
|
out_tensor/model.layers.15.mlp.down_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ebe548bd61bcd705769456e9f78ae96eeb56c02de700526e1d93821565ba4724
|
3 |
+
size 395842
|
out_tensor/model.layers.15.mlp.gate_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1e54257e0c2e223ab062386de204e58f4c32e80b325c92261f8d29bd4c67d316
|
3 |
+
size 363038
|
out_tensor/model.layers.15.mlp.up_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46fe76eb5fefb7aca451179e64ea7526fabdc94c128ce48cee5b3fd69309507b
|
3 |
+
size 387606
|
out_tensor/model.layers.15.self_attn.k_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b7b78c9cbd3001fbfea4bd679ec48ee931583b2e11daf1f7b3624781e9d8b77e
|
3 |
+
size 61500
|
out_tensor/model.layers.15.self_attn.o_proj.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:80a3fffdb056e6da72b19011e017ef18bd310c5e613bfafd17e3819c9da94ddf
|
3 |
+
size 178620
|