Hjgugugjhuhjggg commited on
Commit
726b1cc
·
verified ·
1 Parent(s): 9caa874

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. job_new.json +519 -513
  2. measurement.json +1 -1
  3. out_tensor/lm_head.safetensors +1 -1
  4. out_tensor/model.layers.1.mlp.down_proj.safetensors +2 -2
  5. out_tensor/model.layers.1.mlp.gate_proj.safetensors +2 -2
  6. out_tensor/model.layers.1.mlp.up_proj.safetensors +2 -2
  7. out_tensor/model.layers.1.self_attn.k_proj.safetensors +2 -2
  8. out_tensor/model.layers.1.self_attn.o_proj.safetensors +2 -2
  9. out_tensor/model.layers.1.self_attn.q_proj.safetensors +2 -2
  10. out_tensor/model.layers.1.self_attn.v_proj.safetensors +2 -2
  11. out_tensor/model.layers.10.mlp.down_proj.safetensors +2 -2
  12. out_tensor/model.layers.10.mlp.gate_proj.safetensors +2 -2
  13. out_tensor/model.layers.10.mlp.up_proj.safetensors +2 -2
  14. out_tensor/model.layers.10.self_attn.k_proj.safetensors +2 -2
  15. out_tensor/model.layers.10.self_attn.o_proj.safetensors +2 -2
  16. out_tensor/model.layers.10.self_attn.q_proj.safetensors +2 -2
  17. out_tensor/model.layers.10.self_attn.v_proj.safetensors +2 -2
  18. out_tensor/model.layers.11.mlp.down_proj.safetensors +2 -2
  19. out_tensor/model.layers.11.mlp.gate_proj.safetensors +2 -2
  20. out_tensor/model.layers.11.mlp.up_proj.safetensors +2 -2
  21. out_tensor/model.layers.11.self_attn.k_proj.safetensors +2 -2
  22. out_tensor/model.layers.11.self_attn.o_proj.safetensors +2 -2
  23. out_tensor/model.layers.11.self_attn.q_proj.safetensors +2 -2
  24. out_tensor/model.layers.11.self_attn.v_proj.safetensors +2 -2
  25. out_tensor/model.layers.12.mlp.down_proj.safetensors +2 -2
  26. out_tensor/model.layers.12.mlp.gate_proj.safetensors +2 -2
  27. out_tensor/model.layers.12.mlp.up_proj.safetensors +2 -2
  28. out_tensor/model.layers.12.self_attn.k_proj.safetensors +1 -1
  29. out_tensor/model.layers.12.self_attn.o_proj.safetensors +1 -1
  30. out_tensor/model.layers.12.self_attn.q_proj.safetensors +1 -1
  31. out_tensor/model.layers.12.self_attn.v_proj.safetensors +1 -1
  32. out_tensor/model.layers.13.mlp.down_proj.safetensors +1 -1
  33. out_tensor/model.layers.13.mlp.gate_proj.safetensors +1 -1
  34. out_tensor/model.layers.13.mlp.up_proj.safetensors +1 -1
  35. out_tensor/model.layers.13.self_attn.k_proj.safetensors +2 -2
  36. out_tensor/model.layers.13.self_attn.o_proj.safetensors +2 -2
  37. out_tensor/model.layers.13.self_attn.q_proj.safetensors +2 -2
  38. out_tensor/model.layers.13.self_attn.v_proj.safetensors +2 -2
  39. out_tensor/model.layers.14.mlp.down_proj.safetensors +2 -2
  40. out_tensor/model.layers.14.mlp.gate_proj.safetensors +2 -2
  41. out_tensor/model.layers.14.mlp.up_proj.safetensors +2 -2
  42. out_tensor/model.layers.14.self_attn.k_proj.safetensors +2 -2
  43. out_tensor/model.layers.14.self_attn.o_proj.safetensors +2 -2
  44. out_tensor/model.layers.14.self_attn.q_proj.safetensors +2 -2
  45. out_tensor/model.layers.14.self_attn.v_proj.safetensors +2 -2
  46. out_tensor/model.layers.15.mlp.down_proj.safetensors +2 -2
  47. out_tensor/model.layers.15.mlp.gate_proj.safetensors +2 -2
  48. out_tensor/model.layers.15.mlp.up_proj.safetensors +2 -2
  49. out_tensor/model.layers.15.self_attn.k_proj.safetensors +2 -2
  50. out_tensor/model.layers.15.self_attn.o_proj.safetensors +2 -2
job_new.json CHANGED
@@ -57975,7 +57975,7 @@
57975
  }
57976
  },
57977
  {
57978
- "accuracy": 0.9459031000733376,
57979
  "total_bits": 9097152,
57980
  "gate_proj": {
57981
  "group_size": {
@@ -58657,12 +58657,12 @@
58657
  }
58658
  },
58659
  "model.layers.1.self_attn": {
58660
- "accuracy": 0.9875497296452522,
58661
- "total_bits": 3923328,
58662
  "q_proj": {
58663
  "group_size": {
58664
- "5": 32,
58665
- "4": 32
58666
  },
58667
  "bits": [
58668
  5,
@@ -58676,8 +58676,8 @@
58676
  },
58677
  "k_proj": {
58678
  "group_size": {
58679
- "5": 32,
58680
- "4": 32
58681
  },
58682
  "bits": [
58683
  5,
@@ -58691,20 +58691,23 @@
58691
  },
58692
  "v_proj": {
58693
  "group_size": {
58694
- "5": 32
 
58695
  },
58696
  "bits": [
58697
- 5
 
58698
  ],
58699
  "bits_prop": [
58700
- 1
 
58701
  ],
58702
  "scale_bits": 4
58703
  },
58704
  "o_proj": {
58705
  "group_size": {
58706
- "5": 32,
58707
- "4": 32
58708
  },
58709
  "bits": [
58710
  5,
@@ -58718,16 +58721,16 @@
58718
  }
58719
  },
58720
  "model.layers.1.mlp": {
58721
- "accuracy": 0.9877292113378644,
58722
- "total_bits": 14357424,
58723
  "gate_proj": {
58724
  "group_size": {
58725
- "6": 128,
58726
- "5": 128
58727
  },
58728
  "bits": [
58729
- 6,
58730
- 5
58731
  ],
58732
  "bits_prop": [
58733
  0.1,
@@ -58737,12 +58740,12 @@
58737
  },
58738
  "up_proj": {
58739
  "group_size": {
58740
- "6": 128,
58741
- "5": 128
58742
  },
58743
  "bits": [
58744
- 6,
58745
- 5
58746
  ],
58747
  "bits_prop": [
58748
  0.25,
@@ -58753,13 +58756,13 @@
58753
  "down_proj": {
58754
  "group_size": {
58755
  "8": 32,
58756
- "6": 128,
58757
- "5": 128
58758
  },
58759
  "bits": [
58760
  8,
58761
- 6,
58762
- 5
58763
  ],
58764
  "bits_prop": [
58765
  0.05,
@@ -58770,12 +58773,12 @@
58770
  }
58771
  },
58772
  "model.layers.2.self_attn": {
58773
- "accuracy": 0.9897214015945792,
58774
- "total_bits": 3923328,
58775
  "q_proj": {
58776
  "group_size": {
58777
- "5": 32,
58778
- "4": 32
58779
  },
58780
  "bits": [
58781
  5,
@@ -58789,8 +58792,8 @@
58789
  },
58790
  "k_proj": {
58791
  "group_size": {
58792
- "5": 32,
58793
- "4": 32
58794
  },
58795
  "bits": [
58796
  5,
@@ -58804,20 +58807,23 @@
58804
  },
58805
  "v_proj": {
58806
  "group_size": {
58807
- "5": 32
 
58808
  },
58809
  "bits": [
58810
- 5
 
58811
  ],
58812
  "bits_prop": [
58813
- 1
 
58814
  ],
58815
  "scale_bits": 4
58816
  },
58817
  "o_proj": {
58818
  "group_size": {
58819
- "5": 32,
58820
- "4": 32
58821
  },
58822
  "bits": [
58823
  5,
@@ -58883,76 +58889,64 @@
58883
  }
58884
  },
58885
  "model.layers.3.self_attn": {
58886
- "accuracy": 0.9815531522035599,
58887
- "total_bits": 3775344,
58888
  "q_proj": {
58889
  "group_size": {
58890
- "5": 64,
58891
- "4": 64
58892
  },
58893
  "bits": [
58894
- 5,
58895
- 4
58896
  ],
58897
  "bits_prop": [
58898
- 0.1,
58899
- 0.9
58900
  ],
58901
  "scale_bits": 4
58902
  },
58903
  "k_proj": {
58904
  "group_size": {
58905
- "5": 64,
58906
- "4": 64
58907
  },
58908
  "bits": [
58909
- 5,
58910
- 4
58911
  ],
58912
  "bits_prop": [
58913
- 0.1,
58914
- 0.9
58915
  ],
58916
  "scale_bits": 4
58917
  },
58918
  "v_proj": {
58919
  "group_size": {
58920
- "5": 32,
58921
- "4": 32
58922
  },
58923
  "bits": [
58924
- 5,
58925
- 4
58926
  ],
58927
  "bits_prop": [
58928
- 0.1,
58929
- 0.9
58930
  ],
58931
  "scale_bits": 4
58932
  },
58933
  "o_proj": {
58934
  "group_size": {
58935
- "5": 64,
58936
- "4": 64
58937
  },
58938
  "bits": [
58939
- 5,
58940
- 4
58941
  ],
58942
  "bits_prop": [
58943
- 0.1,
58944
- 0.9
58945
  ],
58946
  "scale_bits": 4
58947
  }
58948
  },
58949
  "model.layers.3.mlp": {
58950
- "accuracy": 0.9879447687417269,
58951
- "total_bits": 14357424,
58952
  "gate_proj": {
58953
  "group_size": {
58954
- "6": 128,
58955
- "5": 128
58956
  },
58957
  "bits": [
58958
  6,
@@ -58966,8 +58960,8 @@
58966
  },
58967
  "up_proj": {
58968
  "group_size": {
58969
- "6": 128,
58970
- "5": 128
58971
  },
58972
  "bits": [
58973
  6,
@@ -58982,8 +58976,8 @@
58982
  "down_proj": {
58983
  "group_size": {
58984
  "8": 32,
58985
- "6": 128,
58986
- "5": 128
58987
  },
58988
  "bits": [
58989
  8,
@@ -59103,16 +59097,16 @@
59103
  }
59104
  },
59105
  "model.layers.5.self_attn": {
59106
- "accuracy": 0.9857744798064232,
59107
- "total_bits": 3923328,
59108
  "q_proj": {
59109
  "group_size": {
59110
- "5": 32,
59111
- "4": 32
59112
  },
59113
  "bits": [
59114
- 5,
59115
- 4
59116
  ],
59117
  "bits_prop": [
59118
  0.1,
@@ -59122,12 +59116,12 @@
59122
  },
59123
  "k_proj": {
59124
  "group_size": {
59125
- "5": 32,
59126
- "4": 32
59127
  },
59128
  "bits": [
59129
- 5,
59130
- 4
59131
  ],
59132
  "bits_prop": [
59133
  0.1,
@@ -59137,10 +59131,10 @@
59137
  },
59138
  "v_proj": {
59139
  "group_size": {
59140
- "5": 32
59141
  },
59142
  "bits": [
59143
- 5
59144
  ],
59145
  "bits_prop": [
59146
  1
@@ -59149,12 +59143,12 @@
59149
  },
59150
  "o_proj": {
59151
  "group_size": {
59152
- "5": 32,
59153
- "4": 32
59154
  },
59155
  "bits": [
59156
- 5,
59157
- 4
59158
  ],
59159
  "bits_prop": [
59160
  0.1,
@@ -59164,53 +59158,44 @@
59164
  }
59165
  },
59166
  "model.layers.5.mlp": {
59167
- "accuracy": 0.9763468410819769,
59168
- "total_bits": 11696064,
59169
  "gate_proj": {
59170
  "group_size": {
59171
- "5": 32,
59172
  "4": 32
59173
  },
59174
  "bits": [
59175
- 5,
59176
  4
59177
  ],
59178
  "bits_prop": [
59179
- 0.1,
59180
- 0.9
59181
  ],
59182
  "scale_bits": 4
59183
  },
59184
  "up_proj": {
59185
  "group_size": {
59186
- "5": 32,
59187
  "4": 32
59188
  },
59189
  "bits": [
59190
- 5,
59191
  4
59192
  ],
59193
  "bits_prop": [
59194
- 0.25,
59195
- 0.75
59196
  ],
59197
  "scale_bits": 4
59198
  },
59199
  "down_proj": {
59200
  "group_size": {
59201
  "8": 32,
59202
- "5": 32,
59203
  "4": 32
59204
  },
59205
  "bits": [
59206
  8,
59207
- 5,
59208
  4
59209
  ],
59210
  "bits_prop": [
59211
  0.05,
59212
- 0.1,
59213
- 0.85
59214
  ],
59215
  "scale_bits": 4
59216
  }
@@ -59320,41 +59305,35 @@
59320
  }
59321
  },
59322
  "model.layers.7.self_attn": {
59323
- "accuracy": 0.991921853274107,
59324
- "total_bits": 4808064,
59325
  "q_proj": {
59326
  "group_size": {
59327
- "6": 32,
59328
- "5": 32
59329
  },
59330
  "bits": [
59331
- 6,
59332
- 5
59333
  ],
59334
  "bits_prop": [
59335
- 0.1,
59336
- 0.9
59337
  ],
59338
  "scale_bits": 4
59339
  },
59340
  "k_proj": {
59341
  "group_size": {
59342
- "6": 32,
59343
- "5": 32
59344
  },
59345
  "bits": [
59346
- 6,
59347
- 5
59348
  ],
59349
  "bits_prop": [
59350
- 0.1,
59351
- 0.9
59352
  ],
59353
  "scale_bits": 4
59354
  },
59355
  "v_proj": {
59356
  "group_size": {
59357
- "6": 32
59358
  },
59359
  "bits": [
59360
  6
@@ -59366,27 +59345,24 @@
59366
  },
59367
  "o_proj": {
59368
  "group_size": {
59369
- "6": 32,
59370
- "5": 32
59371
  },
59372
  "bits": [
59373
- 6,
59374
- 5
59375
  ],
59376
  "bits_prop": [
59377
- 0.1,
59378
- 0.9
59379
  ],
59380
  "scale_bits": 4
59381
  }
59382
  },
59383
  "model.layers.7.mlp": {
59384
- "accuracy": 0.9865649975836277,
59385
- "total_bits": 14357424,
59386
  "gate_proj": {
59387
  "group_size": {
59388
- "6": 128,
59389
- "5": 128
59390
  },
59391
  "bits": [
59392
  6,
@@ -59400,8 +59376,8 @@
59400
  },
59401
  "up_proj": {
59402
  "group_size": {
59403
- "6": 128,
59404
- "5": 128
59405
  },
59406
  "bits": [
59407
  6,
@@ -59416,8 +59392,8 @@
59416
  "down_proj": {
59417
  "group_size": {
59418
  "8": 32,
59419
- "6": 128,
59420
- "5": 128
59421
  },
59422
  "bits": [
59423
  8,
@@ -59607,59 +59583,68 @@
59607
  }
59608
  },
59609
  "model.layers.9.mlp": {
59610
- "accuracy": 0.973205866292119,
59611
- "total_bits": 11259840,
59612
  "gate_proj": {
59613
  "group_size": {
59614
- "4": 32
 
59615
  },
59616
  "bits": [
59617
- 4
 
59618
  ],
59619
  "bits_prop": [
59620
- 1
 
59621
  ],
59622
  "scale_bits": 4
59623
  },
59624
  "up_proj": {
59625
  "group_size": {
59626
- "4": 32
 
59627
  },
59628
  "bits": [
59629
- 4
 
59630
  ],
59631
  "bits_prop": [
59632
- 1
 
59633
  ],
59634
  "scale_bits": 4
59635
  },
59636
  "down_proj": {
59637
  "group_size": {
59638
  "8": 32,
59639
- "4": 32
 
59640
  },
59641
  "bits": [
59642
  8,
59643
- 4
 
59644
  ],
59645
  "bits_prop": [
59646
  0.05,
59647
- 0.95
 
59648
  ],
59649
  "scale_bits": 4
59650
  }
59651
  },
59652
  "model.layers.10.self_attn": {
59653
- "accuracy": 0.9858452072367072,
59654
- "total_bits": 3923328,
59655
  "q_proj": {
59656
  "group_size": {
59657
- "5": 32,
59658
- "4": 32
59659
  },
59660
  "bits": [
59661
- 5,
59662
- 4
59663
  ],
59664
  "bits_prop": [
59665
  0.1,
@@ -59669,12 +59654,12 @@
59669
  },
59670
  "k_proj": {
59671
  "group_size": {
59672
- "5": 32,
59673
- "4": 32
59674
  },
59675
  "bits": [
59676
- 5,
59677
- 4
59678
  ],
59679
  "bits_prop": [
59680
  0.1,
@@ -59684,10 +59669,10 @@
59684
  },
59685
  "v_proj": {
59686
  "group_size": {
59687
- "5": 32
59688
  },
59689
  "bits": [
59690
- 5
59691
  ],
59692
  "bits_prop": [
59693
  1
@@ -59696,12 +59681,12 @@
59696
  },
59697
  "o_proj": {
59698
  "group_size": {
59699
- "5": 32,
59700
- "4": 32
59701
  },
59702
  "bits": [
59703
- 5,
59704
- 4
59705
  ],
59706
  "bits_prop": [
59707
  0.1,
@@ -59711,12 +59696,12 @@
59711
  }
59712
  },
59713
  "model.layers.10.mlp": {
59714
- "accuracy": 0.9880222668871284,
59715
- "total_bits": 14294976,
59716
  "gate_proj": {
59717
  "group_size": {
59718
- "6": 32,
59719
- "5": 32
59720
  },
59721
  "bits": [
59722
  6,
@@ -59730,8 +59715,8 @@
59730
  },
59731
  "up_proj": {
59732
  "group_size": {
59733
- "6": 32,
59734
- "5": 32
59735
  },
59736
  "bits": [
59737
  6,
@@ -59746,8 +59731,8 @@
59746
  "down_proj": {
59747
  "group_size": {
59748
  "8": 32,
59749
- "6": 32,
59750
- "5": 32
59751
  },
59752
  "bits": [
59753
  8,
@@ -59763,44 +59748,38 @@
59763
  }
59764
  },
59765
  "model.layers.11.self_attn": {
59766
- "accuracy": 0.9919869741424918,
59767
- "total_bits": 4811712,
59768
  "q_proj": {
59769
  "group_size": {
59770
- "6": 128,
59771
- "5": 128
59772
  },
59773
  "bits": [
59774
- 6,
59775
- 5
59776
  ],
59777
  "bits_prop": [
59778
- 0.1,
59779
- 0.9
59780
  ],
59781
  "scale_bits": 4
59782
  },
59783
  "k_proj": {
59784
  "group_size": {
59785
- "6": 128,
59786
- "5": 128
59787
  },
59788
  "bits": [
59789
- 6,
59790
- 5
59791
  ],
59792
  "bits_prop": [
59793
- 0.1,
59794
- 0.9
59795
  ],
59796
  "scale_bits": 4
59797
  },
59798
  "v_proj": {
59799
  "group_size": {
59800
- "6": 128
59801
  },
59802
  "bits": [
59803
- 6
59804
  ],
59805
  "bits_prop": [
59806
  1
@@ -59809,27 +59788,24 @@
59809
  },
59810
  "o_proj": {
59811
  "group_size": {
59812
- "6": 128,
59813
- "5": 128
59814
  },
59815
  "bits": [
59816
- 6,
59817
- 5
59818
  ],
59819
  "bits_prop": [
59820
- 0.1,
59821
- 0.9
59822
  ],
59823
  "scale_bits": 4
59824
  }
59825
  },
59826
  "model.layers.11.mlp": {
59827
- "accuracy": 0.9839992877095938,
59828
- "total_bits": 9097152,
59829
  "gate_proj": {
59830
  "group_size": {
59831
- "4": 32,
59832
- "3": 32
59833
  },
59834
  "bits": [
59835
  4,
@@ -59843,8 +59819,8 @@
59843
  },
59844
  "up_proj": {
59845
  "group_size": {
59846
- "4": 32,
59847
- "3": 32
59848
  },
59849
  "bits": [
59850
  4,
@@ -59859,8 +59835,8 @@
59859
  "down_proj": {
59860
  "group_size": {
59861
  "8": 32,
59862
- "4": 32,
59863
- "3": 32
59864
  },
59865
  "bits": [
59866
  8,
@@ -59940,16 +59916,16 @@
59940
  }
59941
  },
59942
  "model.layers.12.mlp": {
59943
- "accuracy": 0.9890716709196568,
59944
- "total_bits": 9159600,
59945
  "gate_proj": {
59946
  "group_size": {
59947
- "4": 128,
59948
- "3": 128
59949
  },
59950
  "bits": [
59951
- 4,
59952
- 3
59953
  ],
59954
  "bits_prop": [
59955
  0.1,
@@ -59959,45 +59935,42 @@
59959
  },
59960
  "up_proj": {
59961
  "group_size": {
59962
- "4": 128,
59963
- "3": 128
59964
  },
59965
  "bits": [
59966
- 4,
59967
- 3
59968
  ],
59969
  "bits_prop": [
59970
- 0.25,
59971
- 0.75
59972
  ],
59973
  "scale_bits": 4
59974
  },
59975
  "down_proj": {
59976
  "group_size": {
59977
- "8": 32,
59978
- "4": 128,
59979
- "3": 128
59980
  },
59981
  "bits": [
59982
- 8,
59983
- 4,
59984
  3
59985
  ],
59986
  "bits_prop": [
59987
  0.05,
59988
- 0.1,
59989
- 0.85
59990
  ],
59991
  "scale_bits": 4
59992
  }
59993
  },
59994
  "model.layers.13.self_attn": {
59995
- "accuracy": 0.992383630014956,
59996
- "total_bits": 3775344,
59997
  "q_proj": {
59998
  "group_size": {
59999
- "5": 64,
60000
- "4": 64
60001
  },
60002
  "bits": [
60003
  5,
@@ -60011,8 +59984,8 @@
60011
  },
60012
  "k_proj": {
60013
  "group_size": {
60014
- "5": 64,
60015
- "4": 64
60016
  },
60017
  "bits": [
60018
  5,
@@ -60026,23 +59999,20 @@
60026
  },
60027
  "v_proj": {
60028
  "group_size": {
60029
- "5": 32,
60030
- "4": 32
60031
  },
60032
  "bits": [
60033
- 5,
60034
- 4
60035
  ],
60036
  "bits_prop": [
60037
- 0.1,
60038
- 0.9
60039
  ],
60040
  "scale_bits": 4
60041
  },
60042
  "o_proj": {
60043
  "group_size": {
60044
- "5": 64,
60045
- "4": 64
60046
  },
60047
  "bits": [
60048
  5,
@@ -60105,44 +60075,38 @@
60105
  }
60106
  },
60107
  "model.layers.14.self_attn": {
60108
- "accuracy": 0.9959560642018914,
60109
- "total_bits": 4808064,
60110
  "q_proj": {
60111
  "group_size": {
60112
- "6": 32,
60113
- "5": 32
60114
  },
60115
  "bits": [
60116
- 6,
60117
- 5
60118
  ],
60119
  "bits_prop": [
60120
- 0.1,
60121
- 0.9
60122
  ],
60123
  "scale_bits": 4
60124
  },
60125
  "k_proj": {
60126
  "group_size": {
60127
- "6": 32,
60128
- "5": 32
60129
  },
60130
  "bits": [
60131
- 6,
60132
- 5
60133
  ],
60134
  "bits_prop": [
60135
- 0.1,
60136
- 0.9
60137
  ],
60138
  "scale_bits": 4
60139
  },
60140
  "v_proj": {
60141
  "group_size": {
60142
- "6": 32
60143
  },
60144
  "bits": [
60145
- 6
60146
  ],
60147
  "bits_prop": [
60148
  1
@@ -60151,31 +60115,28 @@
60151
  },
60152
  "o_proj": {
60153
  "group_size": {
60154
- "6": 32,
60155
- "5": 32
60156
  },
60157
  "bits": [
60158
- 6,
60159
- 5
60160
  ],
60161
  "bits_prop": [
60162
- 0.1,
60163
- 0.9
60164
  ],
60165
  "scale_bits": 4
60166
  }
60167
  },
60168
  "model.layers.14.mlp": {
60169
- "accuracy": 0.9811211936175823,
60170
- "total_bits": 7007328,
60171
  "gate_proj": {
60172
  "group_size": {
60173
- "3": 64,
60174
- "2": 64
60175
  },
60176
  "bits": [
60177
- 3,
60178
- 2
60179
  ],
60180
  "bits_prop": [
60181
  0.1,
@@ -60185,42 +60146,45 @@
60185
  },
60186
  "up_proj": {
60187
  "group_size": {
60188
- "3": 64,
60189
- "2": 64
60190
  },
60191
  "bits": [
60192
- 3,
60193
- 2
60194
  ],
60195
  "bits_prop": [
60196
- 0.3,
60197
- 0.7
60198
  ],
60199
  "scale_bits": 4
60200
  },
60201
  "down_proj": {
60202
  "group_size": {
60203
- "5": 32,
 
60204
  "3": 32
60205
  },
60206
  "bits": [
60207
- 5,
 
60208
  3
60209
  ],
60210
  "bits_prop": [
60211
  0.05,
60212
- 0.95
 
60213
  ],
60214
  "scale_bits": 4
60215
  }
60216
  },
60217
  "model.layers.15.self_attn": {
60218
- "accuracy": 0.9887736244127154,
60219
- "total_bits": 3831936,
60220
  "q_proj": {
60221
  "group_size": {
60222
- "5": 128,
60223
- "4": 128
60224
  },
60225
  "bits": [
60226
  5,
@@ -60234,8 +60198,8 @@
60234
  },
60235
  "k_proj": {
60236
  "group_size": {
60237
- "5": 128,
60238
- "4": 128
60239
  },
60240
  "bits": [
60241
  5,
@@ -60249,23 +60213,20 @@
60249
  },
60250
  "v_proj": {
60251
  "group_size": {
60252
- "5": 64,
60253
- "4": 64
60254
  },
60255
  "bits": [
60256
- 5,
60257
- 4
60258
  ],
60259
  "bits_prop": [
60260
- 0.1,
60261
- 0.9
60262
  ],
60263
  "scale_bits": 4
60264
  },
60265
  "o_proj": {
60266
  "group_size": {
60267
- "5": 128,
60268
- "4": 128
60269
  },
60270
  "bits": [
60271
  5,
@@ -60279,16 +60240,16 @@
60279
  }
60280
  },
60281
  "model.layers.15.mlp": {
60282
- "accuracy": 0.9795254543423653,
60283
- "total_bits": 7007328,
60284
  "gate_proj": {
60285
  "group_size": {
60286
- "3": 64,
60287
- "2": 64
60288
  },
60289
  "bits": [
60290
- 3,
60291
- 2
60292
  ],
60293
  "bits_prop": [
60294
  0.1,
@@ -60298,46 +60259,49 @@
60298
  },
60299
  "up_proj": {
60300
  "group_size": {
60301
- "3": 64,
60302
- "2": 64
60303
  },
60304
  "bits": [
60305
- 3,
60306
- 2
60307
  ],
60308
  "bits_prop": [
60309
- 0.3,
60310
- 0.7
60311
  ],
60312
  "scale_bits": 4
60313
  },
60314
  "down_proj": {
60315
  "group_size": {
60316
- "5": 32,
60317
- "3": 32
 
60318
  },
60319
  "bits": [
60320
- 5,
 
60321
  3
60322
  ],
60323
  "bits_prop": [
60324
  0.05,
60325
- 0.95
 
60326
  ],
60327
  "scale_bits": 4
60328
  }
60329
  },
60330
  "model.layers.16.self_attn": {
60331
- "accuracy": 0.9966306854039431,
60332
- "total_bits": 4808064,
60333
  "q_proj": {
60334
  "group_size": {
60335
- "6": 32,
60336
- "5": 32
60337
  },
60338
  "bits": [
60339
- 6,
60340
- 5
60341
  ],
60342
  "bits_prop": [
60343
  0.1,
@@ -60347,12 +60311,12 @@
60347
  },
60348
  "k_proj": {
60349
  "group_size": {
60350
- "6": 32,
60351
- "5": 32
60352
  },
60353
  "bits": [
60354
- 6,
60355
- 5
60356
  ],
60357
  "bits_prop": [
60358
  0.1,
@@ -60362,24 +60326,27 @@
60362
  },
60363
  "v_proj": {
60364
  "group_size": {
60365
- "6": 32
 
60366
  },
60367
  "bits": [
60368
- 6
 
60369
  ],
60370
  "bits_prop": [
60371
- 1
 
60372
  ],
60373
  "scale_bits": 4
60374
  },
60375
  "o_proj": {
60376
  "group_size": {
60377
- "6": 32,
60378
- "5": 32
60379
  },
60380
  "bits": [
60381
- 6,
60382
- 5
60383
  ],
60384
  "bits_prop": [
60385
  0.1,
@@ -60389,16 +60356,16 @@
60389
  }
60390
  },
60391
  "model.layers.16.mlp": {
60392
- "accuracy": 0.988614184781909,
60393
- "total_bits": 9097152,
60394
  "gate_proj": {
60395
  "group_size": {
60396
- "4": 32,
60397
- "3": 32
60398
  },
60399
  "bits": [
60400
- 4,
60401
- 3
60402
  ],
60403
  "bits_prop": [
60404
  0.1,
@@ -60408,41 +60375,38 @@
60408
  },
60409
  "up_proj": {
60410
  "group_size": {
60411
- "4": 32,
60412
- "3": 32
60413
  },
60414
  "bits": [
60415
- 4,
60416
- 3
60417
  ],
60418
  "bits_prop": [
60419
- 0.25,
60420
- 0.75
60421
  ],
60422
  "scale_bits": 4
60423
  },
60424
  "down_proj": {
60425
  "group_size": {
60426
- "8": 32,
60427
- "4": 32,
60428
  "3": 32
60429
  },
60430
  "bits": [
60431
- 8,
60432
- 4,
60433
  3
60434
  ],
60435
  "bits_prop": [
60436
  0.05,
60437
- 0.1,
60438
- 0.85
60439
  ],
60440
  "scale_bits": 4
60441
  }
60442
  },
60443
  "model.layers.17.self_attn": {
60444
- "accuracy": 0.990932559594512,
60445
- "total_bits": 3775344,
60446
  "q_proj": {
60447
  "group_size": {
60448
  "5": 64,
@@ -60475,16 +60439,13 @@
60475
  },
60476
  "v_proj": {
60477
  "group_size": {
60478
- "5": 32,
60479
- "4": 32
60480
  },
60481
  "bits": [
60482
- 5,
60483
- 4
60484
  ],
60485
  "bits_prop": [
60486
- 0.1,
60487
- 0.9
60488
  ],
60489
  "scale_bits": 4
60490
  },
@@ -60505,16 +60466,16 @@
60505
  }
60506
  },
60507
  "model.layers.17.mlp": {
60508
- "accuracy": 0.9879879523068666,
60509
- "total_bits": 9097152,
60510
  "gate_proj": {
60511
  "group_size": {
60512
- "4": 32,
60513
- "3": 32
60514
  },
60515
  "bits": [
60516
- 4,
60517
- 3
60518
  ],
60519
  "bits_prop": [
60520
  0.1,
@@ -60524,34 +60485,31 @@
60524
  },
60525
  "up_proj": {
60526
  "group_size": {
60527
- "4": 32,
60528
- "3": 32
60529
  },
60530
  "bits": [
60531
- 4,
60532
- 3
60533
  ],
60534
  "bits_prop": [
60535
- 0.25,
60536
- 0.75
60537
  ],
60538
  "scale_bits": 4
60539
  },
60540
  "down_proj": {
60541
  "group_size": {
60542
- "8": 32,
60543
- "4": 32,
60544
  "3": 32
60545
  },
60546
  "bits": [
60547
- 8,
60548
- 4,
60549
  3
60550
  ],
60551
  "bits_prop": [
60552
  0.05,
60553
- 0.1,
60554
- 0.85
60555
  ],
60556
  "scale_bits": 4
60557
  }
@@ -60734,8 +60692,8 @@
60734
  }
60735
  },
60736
  "model.layers.19.mlp": {
60737
- "accuracy": 0.9854860184714198,
60738
- "total_bits": 9756816,
60739
  "gate_proj": {
60740
  "group_size": {
60741
  "4": 32,
@@ -60769,30 +60727,33 @@
60769
  "down_proj": {
60770
  "group_size": {
60771
  "8": 32,
60772
- "4": 128
 
60773
  },
60774
  "bits": [
60775
  8,
60776
- 4
 
60777
  ],
60778
  "bits_prop": [
60779
  0.05,
60780
- 0.95
 
60781
  ],
60782
  "scale_bits": 4
60783
  }
60784
  },
60785
  "model.layers.20.self_attn": {
60786
- "accuracy": 0.9898217963054776,
60787
- "total_bits": 3923328,
60788
  "q_proj": {
60789
  "group_size": {
60790
- "5": 32,
60791
- "4": 32
60792
  },
60793
  "bits": [
60794
- 5,
60795
- 4
60796
  ],
60797
  "bits_prop": [
60798
  0.1,
@@ -60802,12 +60763,12 @@
60802
  },
60803
  "k_proj": {
60804
  "group_size": {
60805
- "5": 32,
60806
- "4": 32
60807
  },
60808
  "bits": [
60809
- 5,
60810
- 4
60811
  ],
60812
  "bits_prop": [
60813
  0.1,
@@ -60817,10 +60778,10 @@
60817
  },
60818
  "v_proj": {
60819
  "group_size": {
60820
- "5": 32
60821
  },
60822
  "bits": [
60823
- 5
60824
  ],
60825
  "bits_prop": [
60826
  1
@@ -60829,12 +60790,12 @@
60829
  },
60830
  "o_proj": {
60831
  "group_size": {
60832
- "5": 32,
60833
- "4": 32
60834
  },
60835
  "bits": [
60836
- 5,
60837
- 4
60838
  ],
60839
  "bits_prop": [
60840
  0.1,
@@ -60844,11 +60805,11 @@
60844
  }
60845
  },
60846
  "model.layers.20.mlp": {
60847
- "accuracy": 0.9882188197225332,
60848
- "total_bits": 11259840,
60849
  "gate_proj": {
60850
  "group_size": {
60851
- "4": 32
60852
  },
60853
  "bits": [
60854
  4
@@ -60873,7 +60834,7 @@
60873
  "down_proj": {
60874
  "group_size": {
60875
  "8": 32,
60876
- "4": 32
60877
  },
60878
  "bits": [
60879
  8,
@@ -60887,8 +60848,8 @@
60887
  }
60888
  },
60889
  "model.layers.21.self_attn": {
60890
- "accuracy": 0.9886166863143444,
60891
- "total_bits": 3866304,
60892
  "q_proj": {
60893
  "group_size": {
60894
  "5": 64,
@@ -60921,13 +60882,16 @@
60921
  },
60922
  "v_proj": {
60923
  "group_size": {
60924
- "5": 64
 
60925
  },
60926
  "bits": [
60927
- 5
 
60928
  ],
60929
  "bits_prop": [
60930
- 1
 
60931
  ],
60932
  "scale_bits": 4
60933
  },
@@ -60948,11 +60912,11 @@
60948
  }
60949
  },
60950
  "model.layers.21.mlp": {
60951
- "accuracy": 0.9882833734154701,
60952
- "total_bits": 11101728,
60953
  "gate_proj": {
60954
  "group_size": {
60955
- "4": 128
60956
  },
60957
  "bits": [
60958
  4
@@ -60977,7 +60941,7 @@
60977
  "down_proj": {
60978
  "group_size": {
60979
  "8": 32,
60980
- "4": 128
60981
  },
60982
  "bits": [
60983
  8,
@@ -60991,12 +60955,12 @@
60991
  }
60992
  },
60993
  "model.layers.22.self_attn": {
60994
- "accuracy": 0.9880260303616524,
60995
- "total_bits": 3831936,
60996
  "q_proj": {
60997
  "group_size": {
60998
- "5": 128,
60999
- "4": 128
61000
  },
61001
  "bits": [
61002
  5,
@@ -61009,9 +60973,9 @@
61009
  "scale_bits": 4
61010
  },
61011
  "k_proj": {
61012
- "group_size": {
61013
- "5": 128,
61014
- "4": 128
61015
  },
61016
  "bits": [
61017
  5,
@@ -61025,8 +60989,8 @@
61025
  },
61026
  "v_proj": {
61027
  "group_size": {
61028
- "5": 64,
61029
- "4": 64
61030
  },
61031
  "bits": [
61032
  5,
@@ -61040,8 +61004,8 @@
61040
  },
61041
  "o_proj": {
61042
  "group_size": {
61043
- "5": 128,
61044
- "4": 128
61045
  },
61046
  "bits": [
61047
  5,
@@ -61098,38 +61062,44 @@
61098
  }
61099
  },
61100
  "model.layers.23.self_attn": {
61101
- "accuracy": 0.9944212981499732,
61102
- "total_bits": 5717376,
61103
  "q_proj": {
61104
  "group_size": {
61105
- "6": 32
 
61106
  },
61107
  "bits": [
61108
- 6
 
61109
  ],
61110
  "bits_prop": [
61111
- 1
 
61112
  ],
61113
  "scale_bits": 4
61114
  },
61115
  "k_proj": {
61116
  "group_size": {
61117
- "6": 32
 
61118
  },
61119
  "bits": [
61120
- 6
 
61121
  ],
61122
  "bits_prop": [
61123
- 1
 
61124
  ],
61125
  "scale_bits": 4
61126
  },
61127
  "v_proj": {
61128
  "group_size": {
61129
- "8": 32
61130
  },
61131
  "bits": [
61132
- 8
61133
  ],
61134
  "bits_prop": [
61135
  1
@@ -61138,13 +61108,16 @@
61138
  },
61139
  "o_proj": {
61140
  "group_size": {
61141
- "6": 32
 
61142
  },
61143
  "bits": [
61144
- 6
 
61145
  ],
61146
  "bits_prop": [
61147
- 1
 
61148
  ],
61149
  "scale_bits": 4
61150
  }
@@ -61193,11 +61166,11 @@
61193
  }
61194
  },
61195
  "model.layers.24.self_attn": {
61196
- "accuracy": 0.9966863535810262,
61197
- "total_bits": 5717376,
61198
  "q_proj": {
61199
  "group_size": {
61200
- "6": 32
61201
  },
61202
  "bits": [
61203
  6
@@ -61209,7 +61182,7 @@
61209
  },
61210
  "k_proj": {
61211
  "group_size": {
61212
- "6": 32
61213
  },
61214
  "bits": [
61215
  6
@@ -61221,10 +61194,10 @@
61221
  },
61222
  "v_proj": {
61223
  "group_size": {
61224
- "8": 32
61225
  },
61226
  "bits": [
61227
- 8
61228
  ],
61229
  "bits_prop": [
61230
  1
@@ -61233,7 +61206,7 @@
61233
  },
61234
  "o_proj": {
61235
  "group_size": {
61236
- "6": 32
61237
  },
61238
  "bits": [
61239
  6
@@ -61245,59 +61218,68 @@
61245
  }
61246
  },
61247
  "model.layers.24.mlp": {
61248
- "accuracy": 0.9846452893689275,
61249
- "total_bits": 11259840,
61250
  "gate_proj": {
61251
  "group_size": {
61252
- "4": 32
 
61253
  },
61254
  "bits": [
 
61255
  4
61256
  ],
61257
  "bits_prop": [
61258
- 1
 
61259
  ],
61260
  "scale_bits": 4
61261
  },
61262
  "up_proj": {
61263
  "group_size": {
61264
- "4": 32
 
61265
  },
61266
  "bits": [
 
61267
  4
61268
  ],
61269
  "bits_prop": [
61270
- 1
 
61271
  ],
61272
  "scale_bits": 4
61273
  },
61274
  "down_proj": {
61275
  "group_size": {
61276
  "8": 32,
61277
- "4": 32
 
61278
  },
61279
  "bits": [
61280
  8,
 
61281
  4
61282
  ],
61283
  "bits_prop": [
61284
  0.05,
61285
- 0.95
 
61286
  ],
61287
  "scale_bits": 4
61288
  }
61289
  },
61290
  "model.layers.25.self_attn": {
61291
- "accuracy": 0.993717601057142,
61292
- "total_bits": 4811712,
61293
  "q_proj": {
61294
  "group_size": {
61295
- "6": 128,
61296
- "5": 128
61297
  },
61298
  "bits": [
61299
- 6,
61300
- 5
61301
  ],
61302
  "bits_prop": [
61303
  0.1,
@@ -61307,12 +61289,12 @@
61307
  },
61308
  "k_proj": {
61309
  "group_size": {
61310
- "6": 128,
61311
- "5": 128
61312
  },
61313
  "bits": [
61314
- 6,
61315
- 5
61316
  ],
61317
  "bits_prop": [
61318
  0.1,
@@ -61322,24 +61304,27 @@
61322
  },
61323
  "v_proj": {
61324
  "group_size": {
61325
- "6": 128
 
61326
  },
61327
  "bits": [
61328
- 6
 
61329
  ],
61330
  "bits_prop": [
61331
- 1
 
61332
  ],
61333
  "scale_bits": 4
61334
  },
61335
  "o_proj": {
61336
  "group_size": {
61337
- "6": 128,
61338
- "5": 128
61339
  },
61340
  "bits": [
61341
- 6,
61342
- 5
61343
  ],
61344
  "bits_prop": [
61345
  0.1,
@@ -61349,11 +61334,11 @@
61349
  }
61350
  },
61351
  "model.layers.25.mlp": {
61352
- "accuracy": 0.98378418199718,
61353
- "total_bits": 11101728,
61354
  "gate_proj": {
61355
  "group_size": {
61356
- "4": 128
61357
  },
61358
  "bits": [
61359
  4
@@ -61378,7 +61363,7 @@
61378
  "down_proj": {
61379
  "group_size": {
61380
  "8": 32,
61381
- "4": 128
61382
  },
61383
  "bits": [
61384
  8,
@@ -61392,12 +61377,12 @@
61392
  }
61393
  },
61394
  "model.layers.26.self_attn": {
61395
- "accuracy": 0.9888241868466139,
61396
- "total_bits": 3923328,
61397
  "q_proj": {
61398
  "group_size": {
61399
- "5": 32,
61400
- "4": 32
61401
  },
61402
  "bits": [
61403
  5,
@@ -61411,8 +61396,8 @@
61411
  },
61412
  "k_proj": {
61413
  "group_size": {
61414
- "5": 32,
61415
- "4": 32
61416
  },
61417
  "bits": [
61418
  5,
@@ -61426,20 +61411,23 @@
61426
  },
61427
  "v_proj": {
61428
  "group_size": {
61429
- "5": 32
 
61430
  },
61431
  "bits": [
61432
- 5
 
61433
  ],
61434
  "bits_prop": [
61435
- 1
 
61436
  ],
61437
  "scale_bits": 4
61438
  },
61439
  "o_proj": {
61440
  "group_size": {
61441
- "5": 32,
61442
- "4": 32
61443
  },
61444
  "bits": [
61445
  5,
@@ -61505,38 +61493,44 @@
61505
  }
61506
  },
61507
  "model.layers.27.self_attn": {
61508
- "accuracy": 0.9915196811780334,
61509
- "total_bits": 5413824,
61510
  "q_proj": {
61511
  "group_size": {
61512
- "6": 128
 
61513
  },
61514
  "bits": [
61515
- 6
 
61516
  ],
61517
  "bits_prop": [
61518
- 1
 
61519
  ],
61520
  "scale_bits": 4
61521
  },
61522
  "k_proj": {
61523
  "group_size": {
61524
- "6": 128
 
61525
  },
61526
  "bits": [
61527
- 6
 
61528
  ],
61529
  "bits_prop": [
61530
- 1
 
61531
  ],
61532
  "scale_bits": 4
61533
  },
61534
  "v_proj": {
61535
  "group_size": {
61536
- "6": 128
61537
  },
61538
  "bits": [
61539
- 6
61540
  ],
61541
  "bits_prop": [
61542
  1
@@ -61545,71 +61539,83 @@
61545
  },
61546
  "o_proj": {
61547
  "group_size": {
61548
- "6": 128
 
61549
  },
61550
  "bits": [
61551
- 6
 
61552
  ],
61553
  "bits_prop": [
61554
- 1
 
61555
  ],
61556
  "scale_bits": 4
61557
  }
61558
  },
61559
  "model.layers.27.mlp": {
61560
- "accuracy": 0.9808992575854063,
61561
- "total_bits": 11101728,
61562
  "gate_proj": {
61563
  "group_size": {
61564
- "4": 128
 
61565
  },
61566
  "bits": [
 
61567
  4
61568
  ],
61569
  "bits_prop": [
61570
- 1
 
61571
  ],
61572
  "scale_bits": 4
61573
  },
61574
  "up_proj": {
61575
  "group_size": {
 
61576
  "4": 32
61577
  },
61578
  "bits": [
 
61579
  4
61580
  ],
61581
  "bits_prop": [
61582
- 1
 
61583
  ],
61584
  "scale_bits": 4
61585
  },
61586
  "down_proj": {
61587
  "group_size": {
61588
  "8": 32,
61589
- "4": 128
 
61590
  },
61591
  "bits": [
61592
  8,
 
61593
  4
61594
  ],
61595
  "bits_prop": [
61596
  0.05,
61597
- 0.95
 
61598
  ],
61599
  "scale_bits": 4
61600
  }
61601
  },
61602
  "model.layers.28.self_attn": {
61603
- "accuracy": 0.9871587986126542,
61604
- "total_bits": 3866304,
61605
  "q_proj": {
61606
  "group_size": {
61607
- "5": 64,
61608
- "4": 64
61609
  },
61610
  "bits": [
61611
- 5,
61612
- 4
61613
  ],
61614
  "bits_prop": [
61615
  0.1,
@@ -61619,12 +61625,12 @@
61619
  },
61620
  "k_proj": {
61621
  "group_size": {
61622
- "5": 64,
61623
- "4": 64
61624
  },
61625
  "bits": [
61626
- 5,
61627
- 4
61628
  ],
61629
  "bits_prop": [
61630
  0.1,
@@ -61634,10 +61640,10 @@
61634
  },
61635
  "v_proj": {
61636
  "group_size": {
61637
- "5": 64
61638
  },
61639
  "bits": [
61640
- 5
61641
  ],
61642
  "bits_prop": [
61643
  1
@@ -61646,12 +61652,12 @@
61646
  },
61647
  "o_proj": {
61648
  "group_size": {
61649
- "5": 64,
61650
- "4": 64
61651
  },
61652
  "bits": [
61653
- 5,
61654
- 4
61655
  ],
61656
  "bits_prop": [
61657
  0.1,
@@ -61713,16 +61719,16 @@
61713
  }
61714
  },
61715
  "model.layers.29.self_attn": {
61716
- "accuracy": 0.9892830261960626,
61717
- "total_bits": 3923328,
61718
  "q_proj": {
61719
  "group_size": {
61720
- "5": 32,
61721
- "4": 32
61722
  },
61723
  "bits": [
61724
- 5,
61725
- 4
61726
  ],
61727
  "bits_prop": [
61728
  0.1,
@@ -61732,12 +61738,12 @@
61732
  },
61733
  "k_proj": {
61734
  "group_size": {
61735
- "5": 32,
61736
- "4": 32
61737
  },
61738
  "bits": [
61739
- 5,
61740
- 4
61741
  ],
61742
  "bits_prop": [
61743
  0.1,
@@ -61747,10 +61753,10 @@
61747
  },
61748
  "v_proj": {
61749
  "group_size": {
61750
- "5": 32
61751
  },
61752
  "bits": [
61753
- 5
61754
  ],
61755
  "bits_prop": [
61756
  1
@@ -61759,12 +61765,12 @@
61759
  },
61760
  "o_proj": {
61761
  "group_size": {
61762
- "5": 32,
61763
- "4": 32
61764
  },
61765
  "bits": [
61766
- 5,
61767
- 4
61768
  ],
61769
  "bits_prop": [
61770
  0.1,
 
57975
  }
57976
  },
57977
  {
57978
+ "accuracy": 0.945903766900301,
57979
  "total_bits": 9097152,
57980
  "gate_proj": {
57981
  "group_size": {
 
58657
  }
58658
  },
58659
  "model.layers.1.self_attn": {
58660
+ "accuracy": 0.9855817724019289,
58661
+ "total_bits": 3831936,
58662
  "q_proj": {
58663
  "group_size": {
58664
+ "5": 128,
58665
+ "4": 128
58666
  },
58667
  "bits": [
58668
  5,
 
58676
  },
58677
  "k_proj": {
58678
  "group_size": {
58679
+ "5": 128,
58680
+ "4": 128
58681
  },
58682
  "bits": [
58683
  5,
 
58691
  },
58692
  "v_proj": {
58693
  "group_size": {
58694
+ "5": 64,
58695
+ "4": 64
58696
  },
58697
  "bits": [
58698
+ 5,
58699
+ 4
58700
  ],
58701
  "bits_prop": [
58702
+ 0.1,
58703
+ 0.9
58704
  ],
58705
  "scale_bits": 4
58706
  },
58707
  "o_proj": {
58708
  "group_size": {
58709
+ "5": 128,
58710
+ "4": 128
58711
  },
58712
  "bits": [
58713
  5,
 
58721
  }
58722
  },
58723
  "model.layers.1.mlp": {
58724
+ "accuracy": 0.9778902214020491,
58725
+ "total_bits": 11696064,
58726
  "gate_proj": {
58727
  "group_size": {
58728
+ "5": 32,
58729
+ "4": 32
58730
  },
58731
  "bits": [
58732
+ 5,
58733
+ 4
58734
  ],
58735
  "bits_prop": [
58736
  0.1,
 
58740
  },
58741
  "up_proj": {
58742
  "group_size": {
58743
+ "5": 32,
58744
+ "4": 32
58745
  },
58746
  "bits": [
58747
+ 5,
58748
+ 4
58749
  ],
58750
  "bits_prop": [
58751
  0.25,
 
58756
  "down_proj": {
58757
  "group_size": {
58758
  "8": 32,
58759
+ "5": 32,
58760
+ "4": 32
58761
  },
58762
  "bits": [
58763
  8,
58764
+ 5,
58765
+ 4
58766
  ],
58767
  "bits_prop": [
58768
  0.05,
 
58773
  }
58774
  },
58775
  "model.layers.2.self_attn": {
58776
+ "accuracy": 0.9876447143033147,
58777
+ "total_bits": 3775344,
58778
  "q_proj": {
58779
  "group_size": {
58780
+ "5": 64,
58781
+ "4": 64
58782
  },
58783
  "bits": [
58784
  5,
 
58792
  },
58793
  "k_proj": {
58794
  "group_size": {
58795
+ "5": 64,
58796
+ "4": 64
58797
  },
58798
  "bits": [
58799
  5,
 
58807
  },
58808
  "v_proj": {
58809
  "group_size": {
58810
+ "5": 32,
58811
+ "4": 32
58812
  },
58813
  "bits": [
58814
+ 5,
58815
+ 4
58816
  ],
58817
  "bits_prop": [
58818
+ 0.1,
58819
+ 0.9
58820
  ],
58821
  "scale_bits": 4
58822
  },
58823
  "o_proj": {
58824
  "group_size": {
58825
+ "5": 64,
58826
+ "4": 64
58827
  },
58828
  "bits": [
58829
  5,
 
58889
  }
58890
  },
58891
  "model.layers.3.self_attn": {
58892
+ "accuracy": 0.9954097527079284,
58893
+ "total_bits": 5717376,
58894
  "q_proj": {
58895
  "group_size": {
58896
+ "6": 32
 
58897
  },
58898
  "bits": [
58899
+ 6
 
58900
  ],
58901
  "bits_prop": [
58902
+ 1
 
58903
  ],
58904
  "scale_bits": 4
58905
  },
58906
  "k_proj": {
58907
  "group_size": {
58908
+ "6": 32
 
58909
  },
58910
  "bits": [
58911
+ 6
 
58912
  ],
58913
  "bits_prop": [
58914
+ 1
 
58915
  ],
58916
  "scale_bits": 4
58917
  },
58918
  "v_proj": {
58919
  "group_size": {
58920
+ "8": 32
 
58921
  },
58922
  "bits": [
58923
+ 8
 
58924
  ],
58925
  "bits_prop": [
58926
+ 1
 
58927
  ],
58928
  "scale_bits": 4
58929
  },
58930
  "o_proj": {
58931
  "group_size": {
58932
+ "6": 32
 
58933
  },
58934
  "bits": [
58935
+ 6
 
58936
  ],
58937
  "bits_prop": [
58938
+ 1
 
58939
  ],
58940
  "scale_bits": 4
58941
  }
58942
  },
58943
  "model.layers.3.mlp": {
58944
+ "accuracy": 0.9891038266941905,
58945
+ "total_bits": 14294976,
58946
  "gate_proj": {
58947
  "group_size": {
58948
+ "6": 32,
58949
+ "5": 32
58950
  },
58951
  "bits": [
58952
  6,
 
58960
  },
58961
  "up_proj": {
58962
  "group_size": {
58963
+ "6": 32,
58964
+ "5": 32
58965
  },
58966
  "bits": [
58967
  6,
 
58976
  "down_proj": {
58977
  "group_size": {
58978
  "8": 32,
58979
+ "6": 32,
58980
+ "5": 32
58981
  },
58982
  "bits": [
58983
  8,
 
59097
  }
59098
  },
59099
  "model.layers.5.self_attn": {
59100
+ "accuracy": 0.9921023324131966,
59101
+ "total_bits": 4808064,
59102
  "q_proj": {
59103
  "group_size": {
59104
+ "6": 32,
59105
+ "5": 32
59106
  },
59107
  "bits": [
59108
+ 6,
59109
+ 5
59110
  ],
59111
  "bits_prop": [
59112
  0.1,
 
59116
  },
59117
  "k_proj": {
59118
  "group_size": {
59119
+ "6": 32,
59120
+ "5": 32
59121
  },
59122
  "bits": [
59123
+ 6,
59124
+ 5
59125
  ],
59126
  "bits_prop": [
59127
  0.1,
 
59131
  },
59132
  "v_proj": {
59133
  "group_size": {
59134
+ "6": 32
59135
  },
59136
  "bits": [
59137
+ 6
59138
  ],
59139
  "bits_prop": [
59140
  1
 
59143
  },
59144
  "o_proj": {
59145
  "group_size": {
59146
+ "6": 32,
59147
+ "5": 32
59148
  },
59149
  "bits": [
59150
+ 6,
59151
+ 5
59152
  ],
59153
  "bits_prop": [
59154
  0.1,
 
59158
  }
59159
  },
59160
  "model.layers.5.mlp": {
59161
+ "accuracy": 0.9745967444032431,
59162
+ "total_bits": 11259840,
59163
  "gate_proj": {
59164
  "group_size": {
 
59165
  "4": 32
59166
  },
59167
  "bits": [
 
59168
  4
59169
  ],
59170
  "bits_prop": [
59171
+ 1
 
59172
  ],
59173
  "scale_bits": 4
59174
  },
59175
  "up_proj": {
59176
  "group_size": {
 
59177
  "4": 32
59178
  },
59179
  "bits": [
 
59180
  4
59181
  ],
59182
  "bits_prop": [
59183
+ 1
 
59184
  ],
59185
  "scale_bits": 4
59186
  },
59187
  "down_proj": {
59188
  "group_size": {
59189
  "8": 32,
 
59190
  "4": 32
59191
  },
59192
  "bits": [
59193
  8,
 
59194
  4
59195
  ],
59196
  "bits_prop": [
59197
  0.05,
59198
+ 0.95
 
59199
  ],
59200
  "scale_bits": 4
59201
  }
 
59305
  }
59306
  },
59307
  "model.layers.7.self_attn": {
59308
+ "accuracy": 0.9904757384210825,
59309
+ "total_bits": 5413824,
59310
  "q_proj": {
59311
  "group_size": {
59312
+ "6": 128
 
59313
  },
59314
  "bits": [
59315
+ 6
 
59316
  ],
59317
  "bits_prop": [
59318
+ 1
 
59319
  ],
59320
  "scale_bits": 4
59321
  },
59322
  "k_proj": {
59323
  "group_size": {
59324
+ "6": 128
 
59325
  },
59326
  "bits": [
59327
+ 6
 
59328
  ],
59329
  "bits_prop": [
59330
+ 1
 
59331
  ],
59332
  "scale_bits": 4
59333
  },
59334
  "v_proj": {
59335
  "group_size": {
59336
+ "6": 128
59337
  },
59338
  "bits": [
59339
  6
 
59345
  },
59346
  "o_proj": {
59347
  "group_size": {
59348
+ "6": 128
 
59349
  },
59350
  "bits": [
59351
+ 6
 
59352
  ],
59353
  "bits_prop": [
59354
+ 1
 
59355
  ],
59356
  "scale_bits": 4
59357
  }
59358
  },
59359
  "model.layers.7.mlp": {
59360
+ "accuracy": 0.9878600733354688,
59361
+ "total_bits": 14294976,
59362
  "gate_proj": {
59363
  "group_size": {
59364
+ "6": 32,
59365
+ "5": 32
59366
  },
59367
  "bits": [
59368
  6,
 
59376
  },
59377
  "up_proj": {
59378
  "group_size": {
59379
+ "6": 32,
59380
+ "5": 32
59381
  },
59382
  "bits": [
59383
  6,
 
59392
  "down_proj": {
59393
  "group_size": {
59394
  "8": 32,
59395
+ "6": 32,
59396
+ "5": 32
59397
  },
59398
  "bits": [
59399
  8,
 
59583
  }
59584
  },
59585
  "model.layers.9.mlp": {
59586
+ "accuracy": 0.9877331163734198,
59587
+ "total_bits": 14294976,
59588
  "gate_proj": {
59589
  "group_size": {
59590
+ "6": 32,
59591
+ "5": 32
59592
  },
59593
  "bits": [
59594
+ 6,
59595
+ 5
59596
  ],
59597
  "bits_prop": [
59598
+ 0.1,
59599
+ 0.9
59600
  ],
59601
  "scale_bits": 4
59602
  },
59603
  "up_proj": {
59604
  "group_size": {
59605
+ "6": 32,
59606
+ "5": 32
59607
  },
59608
  "bits": [
59609
+ 6,
59610
+ 5
59611
  ],
59612
  "bits_prop": [
59613
+ 0.25,
59614
+ 0.75
59615
  ],
59616
  "scale_bits": 4
59617
  },
59618
  "down_proj": {
59619
  "group_size": {
59620
  "8": 32,
59621
+ "6": 32,
59622
+ "5": 32
59623
  },
59624
  "bits": [
59625
  8,
59626
+ 6,
59627
+ 5
59628
  ],
59629
  "bits_prop": [
59630
  0.05,
59631
+ 0.1,
59632
+ 0.85
59633
  ],
59634
  "scale_bits": 4
59635
  }
59636
  },
59637
  "model.layers.10.self_attn": {
59638
+ "accuracy": 0.9927606964483857,
59639
+ "total_bits": 4808064,
59640
  "q_proj": {
59641
  "group_size": {
59642
+ "6": 32,
59643
+ "5": 32
59644
  },
59645
  "bits": [
59646
+ 6,
59647
+ 5
59648
  ],
59649
  "bits_prop": [
59650
  0.1,
 
59654
  },
59655
  "k_proj": {
59656
  "group_size": {
59657
+ "6": 32,
59658
+ "5": 32
59659
  },
59660
  "bits": [
59661
+ 6,
59662
+ 5
59663
  ],
59664
  "bits_prop": [
59665
  0.1,
 
59669
  },
59670
  "v_proj": {
59671
  "group_size": {
59672
+ "6": 32
59673
  },
59674
  "bits": [
59675
+ 6
59676
  ],
59677
  "bits_prop": [
59678
  1
 
59681
  },
59682
  "o_proj": {
59683
  "group_size": {
59684
+ "6": 32,
59685
+ "5": 32
59686
  },
59687
  "bits": [
59688
+ 6,
59689
+ 5
59690
  ],
59691
  "bits_prop": [
59692
  0.1,
 
59696
  }
59697
  },
59698
  "model.layers.10.mlp": {
59699
+ "accuracy": 0.9865744085982442,
59700
+ "total_bits": 14357424,
59701
  "gate_proj": {
59702
  "group_size": {
59703
+ "6": 128,
59704
+ "5": 128
59705
  },
59706
  "bits": [
59707
  6,
 
59715
  },
59716
  "up_proj": {
59717
  "group_size": {
59718
+ "6": 128,
59719
+ "5": 128
59720
  },
59721
  "bits": [
59722
  6,
 
59731
  "down_proj": {
59732
  "group_size": {
59733
  "8": 32,
59734
+ "6": 128,
59735
+ "5": 128
59736
  },
59737
  "bits": [
59738
  8,
 
59748
  }
59749
  },
59750
  "model.layers.11.self_attn": {
59751
+ "accuracy": 0.9963408834300935,
59752
+ "total_bits": 5717376,
59753
  "q_proj": {
59754
  "group_size": {
59755
+ "6": 32
 
59756
  },
59757
  "bits": [
59758
+ 6
 
59759
  ],
59760
  "bits_prop": [
59761
+ 1
 
59762
  ],
59763
  "scale_bits": 4
59764
  },
59765
  "k_proj": {
59766
  "group_size": {
59767
+ "6": 32
 
59768
  },
59769
  "bits": [
59770
+ 6
 
59771
  ],
59772
  "bits_prop": [
59773
+ 1
 
59774
  ],
59775
  "scale_bits": 4
59776
  },
59777
  "v_proj": {
59778
  "group_size": {
59779
+ "8": 32
59780
  },
59781
  "bits": [
59782
+ 8
59783
  ],
59784
  "bits_prop": [
59785
  1
 
59788
  },
59789
  "o_proj": {
59790
  "group_size": {
59791
+ "6": 32
 
59792
  },
59793
  "bits": [
59794
+ 6
 
59795
  ],
59796
  "bits_prop": [
59797
+ 1
 
59798
  ],
59799
  "scale_bits": 4
59800
  }
59801
  },
59802
  "model.layers.11.mlp": {
59803
+ "accuracy": 0.9859139826148748,
59804
+ "total_bits": 9159600,
59805
  "gate_proj": {
59806
  "group_size": {
59807
+ "4": 128,
59808
+ "3": 128
59809
  },
59810
  "bits": [
59811
  4,
 
59819
  },
59820
  "up_proj": {
59821
  "group_size": {
59822
+ "4": 128,
59823
+ "3": 128
59824
  },
59825
  "bits": [
59826
  4,
 
59835
  "down_proj": {
59836
  "group_size": {
59837
  "8": 32,
59838
+ "4": 128,
59839
+ "3": 128
59840
  },
59841
  "bits": [
59842
  8,
 
59916
  }
59917
  },
59918
  "model.layers.12.mlp": {
59919
+ "accuracy": 0.9815687015652657,
59920
+ "total_bits": 7007328,
59921
  "gate_proj": {
59922
  "group_size": {
59923
+ "3": 64,
59924
+ "2": 64
59925
  },
59926
  "bits": [
59927
+ 3,
59928
+ 2
59929
  ],
59930
  "bits_prop": [
59931
  0.1,
 
59935
  },
59936
  "up_proj": {
59937
  "group_size": {
59938
+ "3": 64,
59939
+ "2": 64
59940
  },
59941
  "bits": [
59942
+ 3,
59943
+ 2
59944
  ],
59945
  "bits_prop": [
59946
+ 0.3,
59947
+ 0.7
59948
  ],
59949
  "scale_bits": 4
59950
  },
59951
  "down_proj": {
59952
  "group_size": {
59953
+ "5": 32,
59954
+ "3": 32
 
59955
  },
59956
  "bits": [
59957
+ 5,
 
59958
  3
59959
  ],
59960
  "bits_prop": [
59961
  0.05,
59962
+ 0.95
 
59963
  ],
59964
  "scale_bits": 4
59965
  }
59966
  },
59967
  "model.layers.13.self_attn": {
59968
+ "accuracy": 0.992874032817781,
59969
+ "total_bits": 3923328,
59970
  "q_proj": {
59971
  "group_size": {
59972
+ "5": 32,
59973
+ "4": 32
59974
  },
59975
  "bits": [
59976
  5,
 
59984
  },
59985
  "k_proj": {
59986
  "group_size": {
59987
+ "5": 32,
59988
+ "4": 32
59989
  },
59990
  "bits": [
59991
  5,
 
59999
  },
60000
  "v_proj": {
60001
  "group_size": {
60002
+ "5": 32
 
60003
  },
60004
  "bits": [
60005
+ 5
 
60006
  ],
60007
  "bits_prop": [
60008
+ 1
 
60009
  ],
60010
  "scale_bits": 4
60011
  },
60012
  "o_proj": {
60013
  "group_size": {
60014
+ "5": 32,
60015
+ "4": 32
60016
  },
60017
  "bits": [
60018
  5,
 
60075
  }
60076
  },
60077
  "model.layers.14.self_attn": {
60078
+ "accuracy": 0.9867758862674236,
60079
+ "total_bits": 3677040,
60080
  "q_proj": {
60081
  "group_size": {
60082
+ "4": 64
 
60083
  },
60084
  "bits": [
60085
+ 4
 
60086
  ],
60087
  "bits_prop": [
60088
+ 1
 
60089
  ],
60090
  "scale_bits": 4
60091
  },
60092
  "k_proj": {
60093
  "group_size": {
60094
+ "4": 64
 
60095
  },
60096
  "bits": [
60097
+ 4
 
60098
  ],
60099
  "bits_prop": [
60100
+ 1
 
60101
  ],
60102
  "scale_bits": 4
60103
  },
60104
  "v_proj": {
60105
  "group_size": {
60106
+ "4": 32
60107
  },
60108
  "bits": [
60109
+ 4
60110
  ],
60111
  "bits_prop": [
60112
  1
 
60115
  },
60116
  "o_proj": {
60117
  "group_size": {
60118
+ "4": 64
 
60119
  },
60120
  "bits": [
60121
+ 4
 
60122
  ],
60123
  "bits_prop": [
60124
+ 1
 
60125
  ],
60126
  "scale_bits": 4
60127
  }
60128
  },
60129
  "model.layers.14.mlp": {
60130
+ "accuracy": 0.9892165157943964,
60131
+ "total_bits": 9097152,
60132
  "gate_proj": {
60133
  "group_size": {
60134
+ "4": 32,
60135
+ "3": 32
60136
  },
60137
  "bits": [
60138
+ 4,
60139
+ 3
60140
  ],
60141
  "bits_prop": [
60142
  0.1,
 
60146
  },
60147
  "up_proj": {
60148
  "group_size": {
60149
+ "4": 32,
60150
+ "3": 32
60151
  },
60152
  "bits": [
60153
+ 4,
60154
+ 3
60155
  ],
60156
  "bits_prop": [
60157
+ 0.25,
60158
+ 0.75
60159
  ],
60160
  "scale_bits": 4
60161
  },
60162
  "down_proj": {
60163
  "group_size": {
60164
+ "8": 32,
60165
+ "4": 32,
60166
  "3": 32
60167
  },
60168
  "bits": [
60169
+ 8,
60170
+ 4,
60171
  3
60172
  ],
60173
  "bits_prop": [
60174
  0.05,
60175
+ 0.1,
60176
+ 0.85
60177
  ],
60178
  "scale_bits": 4
60179
  }
60180
  },
60181
  "model.layers.15.self_attn": {
60182
+ "accuracy": 0.9902772186324,
60183
+ "total_bits": 3923328,
60184
  "q_proj": {
60185
  "group_size": {
60186
+ "5": 32,
60187
+ "4": 32
60188
  },
60189
  "bits": [
60190
  5,
 
60198
  },
60199
  "k_proj": {
60200
  "group_size": {
60201
+ "5": 32,
60202
+ "4": 32
60203
  },
60204
  "bits": [
60205
  5,
 
60213
  },
60214
  "v_proj": {
60215
  "group_size": {
60216
+ "5": 32
 
60217
  },
60218
  "bits": [
60219
+ 5
 
60220
  ],
60221
  "bits_prop": [
60222
+ 1
 
60223
  ],
60224
  "scale_bits": 4
60225
  },
60226
  "o_proj": {
60227
  "group_size": {
60228
+ "5": 32,
60229
+ "4": 32
60230
  },
60231
  "bits": [
60232
  5,
 
60240
  }
60241
  },
60242
  "model.layers.15.mlp": {
60243
+ "accuracy": 0.9878587136045098,
60244
+ "total_bits": 9159600,
60245
  "gate_proj": {
60246
  "group_size": {
60247
+ "4": 128,
60248
+ "3": 128
60249
  },
60250
  "bits": [
60251
+ 4,
60252
+ 3
60253
  ],
60254
  "bits_prop": [
60255
  0.1,
 
60259
  },
60260
  "up_proj": {
60261
  "group_size": {
60262
+ "4": 128,
60263
+ "3": 128
60264
  },
60265
  "bits": [
60266
+ 4,
60267
+ 3
60268
  ],
60269
  "bits_prop": [
60270
+ 0.25,
60271
+ 0.75
60272
  ],
60273
  "scale_bits": 4
60274
  },
60275
  "down_proj": {
60276
  "group_size": {
60277
+ "8": 32,
60278
+ "4": 128,
60279
+ "3": 128
60280
  },
60281
  "bits": [
60282
+ 8,
60283
+ 4,
60284
  3
60285
  ],
60286
  "bits_prop": [
60287
  0.05,
60288
+ 0.1,
60289
+ 0.85
60290
  ],
60291
  "scale_bits": 4
60292
  }
60293
  },
60294
  "model.layers.16.self_attn": {
60295
+ "accuracy": 0.9928999468684196,
60296
+ "total_bits": 3775344,
60297
  "q_proj": {
60298
  "group_size": {
60299
+ "5": 64,
60300
+ "4": 64
60301
  },
60302
  "bits": [
60303
+ 5,
60304
+ 4
60305
  ],
60306
  "bits_prop": [
60307
  0.1,
 
60311
  },
60312
  "k_proj": {
60313
  "group_size": {
60314
+ "5": 64,
60315
+ "4": 64
60316
  },
60317
  "bits": [
60318
+ 5,
60319
+ 4
60320
  ],
60321
  "bits_prop": [
60322
  0.1,
 
60326
  },
60327
  "v_proj": {
60328
  "group_size": {
60329
+ "5": 32,
60330
+ "4": 32
60331
  },
60332
  "bits": [
60333
+ 5,
60334
+ 4
60335
  ],
60336
  "bits_prop": [
60337
+ 0.1,
60338
+ 0.9
60339
  ],
60340
  "scale_bits": 4
60341
  },
60342
  "o_proj": {
60343
  "group_size": {
60344
+ "5": 64,
60345
+ "4": 64
60346
  },
60347
  "bits": [
60348
+ 5,
60349
+ 4
60350
  ],
60351
  "bits_prop": [
60352
  0.1,
 
60356
  }
60357
  },
60358
  "model.layers.16.mlp": {
60359
+ "accuracy": 0.9799803365021944,
60360
+ "total_bits": 7007328,
60361
  "gate_proj": {
60362
  "group_size": {
60363
+ "3": 64,
60364
+ "2": 64
60365
  },
60366
  "bits": [
60367
+ 3,
60368
+ 2
60369
  ],
60370
  "bits_prop": [
60371
  0.1,
 
60375
  },
60376
  "up_proj": {
60377
  "group_size": {
60378
+ "3": 64,
60379
+ "2": 64
60380
  },
60381
  "bits": [
60382
+ 3,
60383
+ 2
60384
  ],
60385
  "bits_prop": [
60386
+ 0.3,
60387
+ 0.7
60388
  ],
60389
  "scale_bits": 4
60390
  },
60391
  "down_proj": {
60392
  "group_size": {
60393
+ "5": 32,
 
60394
  "3": 32
60395
  },
60396
  "bits": [
60397
+ 5,
 
60398
  3
60399
  ],
60400
  "bits_prop": [
60401
  0.05,
60402
+ 0.95
 
60403
  ],
60404
  "scale_bits": 4
60405
  }
60406
  },
60407
  "model.layers.17.self_attn": {
60408
+ "accuracy": 0.9908454418182373,
60409
+ "total_bits": 3866304,
60410
  "q_proj": {
60411
  "group_size": {
60412
  "5": 64,
 
60439
  },
60440
  "v_proj": {
60441
  "group_size": {
60442
+ "5": 64
 
60443
  },
60444
  "bits": [
60445
+ 5
 
60446
  ],
60447
  "bits_prop": [
60448
+ 1
 
60449
  ],
60450
  "scale_bits": 4
60451
  },
 
60466
  }
60467
  },
60468
  "model.layers.17.mlp": {
60469
+ "accuracy": 0.9788082204759121,
60470
+ "total_bits": 7007328,
60471
  "gate_proj": {
60472
  "group_size": {
60473
+ "3": 64,
60474
+ "2": 64
60475
  },
60476
  "bits": [
60477
+ 3,
60478
+ 2
60479
  ],
60480
  "bits_prop": [
60481
  0.1,
 
60485
  },
60486
  "up_proj": {
60487
  "group_size": {
60488
+ "3": 64,
60489
+ "2": 64
60490
  },
60491
  "bits": [
60492
+ 3,
60493
+ 2
60494
  ],
60495
  "bits_prop": [
60496
+ 0.3,
60497
+ 0.7
60498
  ],
60499
  "scale_bits": 4
60500
  },
60501
  "down_proj": {
60502
  "group_size": {
60503
+ "5": 32,
 
60504
  "3": 32
60505
  },
60506
  "bits": [
60507
+ 5,
 
60508
  3
60509
  ],
60510
  "bits_prop": [
60511
  0.05,
60512
+ 0.95
 
60513
  ],
60514
  "scale_bits": 4
60515
  }
 
60692
  }
60693
  },
60694
  "model.layers.19.mlp": {
60695
+ "accuracy": 0.9835127778351307,
60696
+ "total_bits": 9097152,
60697
  "gate_proj": {
60698
  "group_size": {
60699
  "4": 32,
 
60727
  "down_proj": {
60728
  "group_size": {
60729
  "8": 32,
60730
+ "4": 32,
60731
+ "3": 32
60732
  },
60733
  "bits": [
60734
  8,
60735
+ 4,
60736
+ 3
60737
  ],
60738
  "bits_prop": [
60739
  0.05,
60740
+ 0.1,
60741
+ 0.85
60742
  ],
60743
  "scale_bits": 4
60744
  }
60745
  },
60746
  "model.layers.20.self_attn": {
60747
+ "accuracy": 0.9949732795357704,
60748
+ "total_bits": 4808064,
60749
  "q_proj": {
60750
  "group_size": {
60751
+ "6": 32,
60752
+ "5": 32
60753
  },
60754
  "bits": [
60755
+ 6,
60756
+ 5
60757
  ],
60758
  "bits_prop": [
60759
  0.1,
 
60763
  },
60764
  "k_proj": {
60765
  "group_size": {
60766
+ "6": 32,
60767
+ "5": 32
60768
  },
60769
  "bits": [
60770
+ 6,
60771
+ 5
60772
  ],
60773
  "bits_prop": [
60774
  0.1,
 
60778
  },
60779
  "v_proj": {
60780
  "group_size": {
60781
+ "6": 32
60782
  },
60783
  "bits": [
60784
+ 6
60785
  ],
60786
  "bits_prop": [
60787
  1
 
60790
  },
60791
  "o_proj": {
60792
  "group_size": {
60793
+ "6": 32,
60794
+ "5": 32
60795
  },
60796
  "bits": [
60797
+ 6,
60798
+ 5
60799
  ],
60800
  "bits_prop": [
60801
  0.1,
 
60805
  }
60806
  },
60807
  "model.layers.20.mlp": {
60808
+ "accuracy": 0.9870169991627336,
60809
+ "total_bits": 11101728,
60810
  "gate_proj": {
60811
  "group_size": {
60812
+ "4": 128
60813
  },
60814
  "bits": [
60815
  4
 
60834
  "down_proj": {
60835
  "group_size": {
60836
  "8": 32,
60837
+ "4": 128
60838
  },
60839
  "bits": [
60840
  8,
 
60848
  }
60849
  },
60850
  "model.layers.21.self_attn": {
60851
+ "accuracy": 0.9899369664490223,
60852
+ "total_bits": 3775344,
60853
  "q_proj": {
60854
  "group_size": {
60855
  "5": 64,
 
60882
  },
60883
  "v_proj": {
60884
  "group_size": {
60885
+ "5": 32,
60886
+ "4": 32
60887
  },
60888
  "bits": [
60889
+ 5,
60890
+ 4
60891
  ],
60892
  "bits_prop": [
60893
+ 0.1,
60894
+ 0.9
60895
  ],
60896
  "scale_bits": 4
60897
  },
 
60912
  }
60913
  },
60914
  "model.layers.21.mlp": {
60915
+ "accuracy": 0.9894305132329464,
60916
+ "total_bits": 11259840,
60917
  "gate_proj": {
60918
  "group_size": {
60919
+ "4": 32
60920
  },
60921
  "bits": [
60922
  4
 
60941
  "down_proj": {
60942
  "group_size": {
60943
  "8": 32,
60944
+ "4": 32
60945
  },
60946
  "bits": [
60947
  8,
 
60955
  }
60956
  },
60957
  "model.layers.22.self_attn": {
60958
+ "accuracy": 0.9895689664408565,
60959
+ "total_bits": 3775344,
60960
  "q_proj": {
60961
  "group_size": {
60962
+ "5": 64,
60963
+ "4": 64
60964
  },
60965
  "bits": [
60966
  5,
 
60973
  "scale_bits": 4
60974
  },
60975
  "k_proj": {
60976
+ "group_size": {
60977
+ "5": 64,
60978
+ "4": 64
60979
  },
60980
  "bits": [
60981
  5,
 
60989
  },
60990
  "v_proj": {
60991
  "group_size": {
60992
+ "5": 32,
60993
+ "4": 32
60994
  },
60995
  "bits": [
60996
  5,
 
61004
  },
61005
  "o_proj": {
61006
  "group_size": {
61007
+ "5": 64,
61008
+ "4": 64
61009
  },
61010
  "bits": [
61011
  5,
 
61062
  }
61063
  },
61064
  "model.layers.23.self_attn": {
61065
+ "accuracy": 0.9891995070502162,
61066
+ "total_bits": 4808064,
61067
  "q_proj": {
61068
  "group_size": {
61069
+ "6": 32,
61070
+ "5": 32
61071
  },
61072
  "bits": [
61073
+ 6,
61074
+ 5
61075
  ],
61076
  "bits_prop": [
61077
+ 0.1,
61078
+ 0.9
61079
  ],
61080
  "scale_bits": 4
61081
  },
61082
  "k_proj": {
61083
  "group_size": {
61084
+ "6": 32,
61085
+ "5": 32
61086
  },
61087
  "bits": [
61088
+ 6,
61089
+ 5
61090
  ],
61091
  "bits_prop": [
61092
+ 0.1,
61093
+ 0.9
61094
  ],
61095
  "scale_bits": 4
61096
  },
61097
  "v_proj": {
61098
  "group_size": {
61099
+ "6": 32
61100
  },
61101
  "bits": [
61102
+ 6
61103
  ],
61104
  "bits_prop": [
61105
  1
 
61108
  },
61109
  "o_proj": {
61110
  "group_size": {
61111
+ "6": 32,
61112
+ "5": 32
61113
  },
61114
  "bits": [
61115
+ 6,
61116
+ 5
61117
  ],
61118
  "bits_prop": [
61119
+ 0.1,
61120
+ 0.9
61121
  ],
61122
  "scale_bits": 4
61123
  }
 
61166
  }
61167
  },
61168
  "model.layers.24.self_attn": {
61169
+ "accuracy": 0.990554129704833,
61170
+ "total_bits": 5413824,
61171
  "q_proj": {
61172
  "group_size": {
61173
+ "6": 128
61174
  },
61175
  "bits": [
61176
  6
 
61182
  },
61183
  "k_proj": {
61184
  "group_size": {
61185
+ "6": 128
61186
  },
61187
  "bits": [
61188
  6
 
61194
  },
61195
  "v_proj": {
61196
  "group_size": {
61197
+ "6": 128
61198
  },
61199
  "bits": [
61200
+ 6
61201
  ],
61202
  "bits_prop": [
61203
  1
 
61206
  },
61207
  "o_proj": {
61208
  "group_size": {
61209
+ "6": 128
61210
  },
61211
  "bits": [
61212
  6
 
61218
  }
61219
  },
61220
  "model.layers.24.mlp": {
61221
+ "accuracy": 0.9851180734112859,
61222
+ "total_bits": 11758512,
61223
  "gate_proj": {
61224
  "group_size": {
61225
+ "5": 128,
61226
+ "4": 128
61227
  },
61228
  "bits": [
61229
+ 5,
61230
  4
61231
  ],
61232
  "bits_prop": [
61233
+ 0.1,
61234
+ 0.9
61235
  ],
61236
  "scale_bits": 4
61237
  },
61238
  "up_proj": {
61239
  "group_size": {
61240
+ "5": 128,
61241
+ "4": 128
61242
  },
61243
  "bits": [
61244
+ 5,
61245
  4
61246
  ],
61247
  "bits_prop": [
61248
+ 0.25,
61249
+ 0.75
61250
  ],
61251
  "scale_bits": 4
61252
  },
61253
  "down_proj": {
61254
  "group_size": {
61255
  "8": 32,
61256
+ "5": 128,
61257
+ "4": 128
61258
  },
61259
  "bits": [
61260
  8,
61261
+ 5,
61262
  4
61263
  ],
61264
  "bits_prop": [
61265
  0.05,
61266
+ 0.1,
61267
+ 0.85
61268
  ],
61269
  "scale_bits": 4
61270
  }
61271
  },
61272
  "model.layers.25.self_attn": {
61273
+ "accuracy": 0.9908542837947607,
61274
+ "total_bits": 3775344,
61275
  "q_proj": {
61276
  "group_size": {
61277
+ "5": 64,
61278
+ "4": 64
61279
  },
61280
  "bits": [
61281
+ 5,
61282
+ 4
61283
  ],
61284
  "bits_prop": [
61285
  0.1,
 
61289
  },
61290
  "k_proj": {
61291
  "group_size": {
61292
+ "5": 64,
61293
+ "4": 64
61294
  },
61295
  "bits": [
61296
+ 5,
61297
+ 4
61298
  ],
61299
  "bits_prop": [
61300
  0.1,
 
61304
  },
61305
  "v_proj": {
61306
  "group_size": {
61307
+ "5": 32,
61308
+ "4": 32
61309
  },
61310
  "bits": [
61311
+ 5,
61312
+ 4
61313
  ],
61314
  "bits_prop": [
61315
+ 0.1,
61316
+ 0.9
61317
  ],
61318
  "scale_bits": 4
61319
  },
61320
  "o_proj": {
61321
  "group_size": {
61322
+ "5": 64,
61323
+ "4": 64
61324
  },
61325
  "bits": [
61326
+ 5,
61327
+ 4
61328
  ],
61329
  "bits_prop": [
61330
  0.1,
 
61334
  }
61335
  },
61336
  "model.layers.25.mlp": {
61337
+ "accuracy": 0.9851860506460071,
61338
+ "total_bits": 11259840,
61339
  "gate_proj": {
61340
  "group_size": {
61341
+ "4": 32
61342
  },
61343
  "bits": [
61344
  4
 
61363
  "down_proj": {
61364
  "group_size": {
61365
  "8": 32,
61366
+ "4": 32
61367
  },
61368
  "bits": [
61369
  8,
 
61377
  }
61378
  },
61379
  "model.layers.26.self_attn": {
61380
+ "accuracy": 0.9889037329703569,
61381
+ "total_bits": 3775344,
61382
  "q_proj": {
61383
  "group_size": {
61384
+ "5": 64,
61385
+ "4": 64
61386
  },
61387
  "bits": [
61388
  5,
 
61396
  },
61397
  "k_proj": {
61398
  "group_size": {
61399
+ "5": 64,
61400
+ "4": 64
61401
  },
61402
  "bits": [
61403
  5,
 
61411
  },
61412
  "v_proj": {
61413
  "group_size": {
61414
+ "5": 32,
61415
+ "4": 32
61416
  },
61417
  "bits": [
61418
+ 5,
61419
+ 4
61420
  ],
61421
  "bits_prop": [
61422
+ 0.1,
61423
+ 0.9
61424
  ],
61425
  "scale_bits": 4
61426
  },
61427
  "o_proj": {
61428
  "group_size": {
61429
+ "5": 64,
61430
+ "4": 64
61431
  },
61432
  "bits": [
61433
  5,
 
61493
  }
61494
  },
61495
  "model.layers.27.self_attn": {
61496
+ "accuracy": 0.9844594039022923,
61497
+ "total_bits": 3866304,
61498
  "q_proj": {
61499
  "group_size": {
61500
+ "5": 64,
61501
+ "4": 64
61502
  },
61503
  "bits": [
61504
+ 5,
61505
+ 4
61506
  ],
61507
  "bits_prop": [
61508
+ 0.1,
61509
+ 0.9
61510
  ],
61511
  "scale_bits": 4
61512
  },
61513
  "k_proj": {
61514
  "group_size": {
61515
+ "5": 64,
61516
+ "4": 64
61517
  },
61518
  "bits": [
61519
+ 5,
61520
+ 4
61521
  ],
61522
  "bits_prop": [
61523
+ 0.1,
61524
+ 0.9
61525
  ],
61526
  "scale_bits": 4
61527
  },
61528
  "v_proj": {
61529
  "group_size": {
61530
+ "5": 64
61531
  },
61532
  "bits": [
61533
+ 5
61534
  ],
61535
  "bits_prop": [
61536
  1
 
61539
  },
61540
  "o_proj": {
61541
  "group_size": {
61542
+ "5": 64,
61543
+ "4": 64
61544
  },
61545
  "bits": [
61546
+ 5,
61547
+ 4
61548
  ],
61549
  "bits_prop": [
61550
+ 0.1,
61551
+ 0.9
61552
  ],
61553
  "scale_bits": 4
61554
  }
61555
  },
61556
  "model.layers.27.mlp": {
61557
+ "accuracy": 0.9841680098325014,
61558
+ "total_bits": 11696064,
61559
  "gate_proj": {
61560
  "group_size": {
61561
+ "5": 32,
61562
+ "4": 32
61563
  },
61564
  "bits": [
61565
+ 5,
61566
  4
61567
  ],
61568
  "bits_prop": [
61569
+ 0.1,
61570
+ 0.9
61571
  ],
61572
  "scale_bits": 4
61573
  },
61574
  "up_proj": {
61575
  "group_size": {
61576
+ "5": 32,
61577
  "4": 32
61578
  },
61579
  "bits": [
61580
+ 5,
61581
  4
61582
  ],
61583
  "bits_prop": [
61584
+ 0.25,
61585
+ 0.75
61586
  ],
61587
  "scale_bits": 4
61588
  },
61589
  "down_proj": {
61590
  "group_size": {
61591
  "8": 32,
61592
+ "5": 32,
61593
+ "4": 32
61594
  },
61595
  "bits": [
61596
  8,
61597
+ 5,
61598
  4
61599
  ],
61600
  "bits_prop": [
61601
  0.05,
61602
+ 0.1,
61603
+ 0.85
61604
  ],
61605
  "scale_bits": 4
61606
  }
61607
  },
61608
  "model.layers.28.self_attn": {
61609
+ "accuracy": 0.9928587442263961,
61610
+ "total_bits": 4811712,
61611
  "q_proj": {
61612
  "group_size": {
61613
+ "6": 128,
61614
+ "5": 128
61615
  },
61616
  "bits": [
61617
+ 6,
61618
+ 5
61619
  ],
61620
  "bits_prop": [
61621
  0.1,
 
61625
  },
61626
  "k_proj": {
61627
  "group_size": {
61628
+ "6": 128,
61629
+ "5": 128
61630
  },
61631
  "bits": [
61632
+ 6,
61633
+ 5
61634
  ],
61635
  "bits_prop": [
61636
  0.1,
 
61640
  },
61641
  "v_proj": {
61642
  "group_size": {
61643
+ "6": 128
61644
  },
61645
  "bits": [
61646
+ 6
61647
  ],
61648
  "bits_prop": [
61649
  1
 
61652
  },
61653
  "o_proj": {
61654
  "group_size": {
61655
+ "6": 128,
61656
+ "5": 128
61657
  },
61658
  "bits": [
61659
+ 6,
61660
+ 5
61661
  ],
61662
  "bits_prop": [
61663
  0.1,
 
61719
  }
61720
  },
61721
  "model.layers.29.self_attn": {
61722
+ "accuracy": 0.9930798751302063,
61723
+ "total_bits": 4811712,
61724
  "q_proj": {
61725
  "group_size": {
61726
+ "6": 128,
61727
+ "5": 128
61728
  },
61729
  "bits": [
61730
+ 6,
61731
+ 5
61732
  ],
61733
  "bits_prop": [
61734
  0.1,
 
61738
  },
61739
  "k_proj": {
61740
  "group_size": {
61741
+ "6": 128,
61742
+ "5": 128
61743
  },
61744
  "bits": [
61745
+ 6,
61746
+ 5
61747
  ],
61748
  "bits_prop": [
61749
  0.1,
 
61753
  },
61754
  "v_proj": {
61755
  "group_size": {
61756
+ "6": 128
61757
  },
61758
  "bits": [
61759
+ 6
61760
  ],
61761
  "bits_prop": [
61762
  1
 
61765
  },
61766
  "o_proj": {
61767
  "group_size": {
61768
+ "6": 128,
61769
+ "5": 128
61770
  },
61771
  "bits": [
61772
+ 6,
61773
+ 5
61774
  ],
61775
  "bits_prop": [
61776
  0.1,
measurement.json CHANGED
@@ -57957,7 +57957,7 @@
57957
  }
57958
  },
57959
  {
57960
- "accuracy": 0.9459031000733376,
57961
  "total_bits": 9097152,
57962
  "gate_proj": {
57963
  "group_size": {
 
57957
  }
57958
  },
57959
  {
57960
+ "accuracy": 0.945903766900301,
57961
  "total_bits": 9097152,
57962
  "gate_proj": {
57963
  "group_size": {
out_tensor/lm_head.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b9f25323c6818fe4c887f59b19d5d0a524da5582ec86f0c4ad16c1090a450d49
3
  size 22932166
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b00e9a9ad58653c8b753d64bc96345482aeb377528d8a49a22664021233fe51
3
  size 22932166
out_tensor/model.layers.1.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:02ad4e4e1f4f763e77c127e40f1693f16c14cca4be9c58dd7ccca3347f1b0814
3
- size 603194
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:122e426acb718ec8500b571a642bc6798b45e486dab0e383596af8cd42bc9b6c
3
+ size 502304
out_tensor/model.layers.1.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d06165f78e527bb210e5eee44e071d9430b7639696472c6392ce88b566d04450
3
- size 584222
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:89cbc78f723f7c5f29a38cb9d42ae84c09a7332f9b6984869ff801409e97fd4a
3
+ size 471404
out_tensor/model.layers.1.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea571575ee5e595cb8396553c6e3823d05fc4fd02bfa2513b3f02cf12ca140de
3
- size 608790
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:acbc58437663f0da9c0a7310111436b0c35546cfedffce7f76c5c44e967ff095
3
+ size 489828
out_tensor/model.layers.1.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:29e1460798449ac06fb55339ac68dc005180bb1e2f75753de43447afb915accd
3
- size 61492
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1443175687f038577cbe5062f7fa9b9954b66111bde7a407efb50391040dc1a
3
+ size 61702
out_tensor/model.layers.1.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3dc6bbf89dd16c69723149bad79cfd9a560d033d735bbc8fe0ce90aeda1cc630
3
- size 178620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:43634971f137ca8d8fd563c564f2bcc0842893c013811019932debc28b3d00a0
3
+ size 179398
out_tensor/model.layers.1.self_attn.q_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:51b716b59934a4053506e87aca5b33105304201e0f835bef0e0904c9d5f106ab
3
- size 178620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bdcbbb552d31497058166738d8966175d730c296fb7c454cb10b975e8164fca9
3
+ size 179398
out_tensor/model.layers.1.self_attn.v_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c857c9f7b172f986804e76230a088bf642460d86325fdfb4e029eb7f7b184298
3
- size 73780
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ba2e26e21b021c7497dc7ba363a1a70c8a303c7c9737a5a6c4ba4d4eb26fe28a
3
+ size 60574
out_tensor/model.layers.10.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:50837f6e35d08ea49d8280137939435aa67119cad78902921b95eec0fd10d00d
3
- size 605992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5c358e71b6a5272eec2d6a732b7b380d5ee7eedf455156b908e3be611d17c4c2
3
+ size 603202
out_tensor/model.layers.10.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8d570eb4c8e9c8baad90b81355046c1b0fc9be0b004a62d92e465f8bf71e3f83
3
- size 582004
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa1b81408ba9e82a77eef6916698df21c2dab7d12994f61609052e81e32478ad
3
+ size 584222
out_tensor/model.layers.10.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1fb027439776695a9e6244289942ee15e30a7c01b0205e531a9cd2fbc6bc7496
3
- size 600428
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c97dc4ce1700dd6567bb94db216c3fad52f90b46c5c46a4699f3175c226cc40e
3
+ size 608790
out_tensor/model.layers.10.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c67e9afe12af7ed7dae627e29dd48df01310ea6ff8c7bbf4ee3b2e95dd009fff
3
- size 61500
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b108b75252b492881465e9a18807ef9db35f89219728b81082ec7d052f014ac3
3
+ size 75324
out_tensor/model.layers.10.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:26431808f025de4cc2b7b7a156a6ea6d0d1c39d40601270e9e0501154b9a021c
3
- size 178620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91e1efa7d78ad6b225042ae5c222f822d4fddfe77618d0cba3142c8868aa16c3
3
+ size 220092
out_tensor/model.layers.10.self_attn.q_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4cd75411c531fd27b80d9ced008d05969fa466531b1915964cf94aad8c53aaa
3
- size 178620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55391444fa5d42d80e7a36478be427cf84f47fb2d7f3e057cbd3a4bb25a03708
3
+ size 220092
out_tensor/model.layers.10.self_attn.v_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:939daad1a3f92ccd64caf7b238101d972b57eea3544c0904304cce26ce00a38a
3
- size 73788
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2991ae45c537c35acbc7c90b65a56beeb96ed7ee04e26d7b6351f7ac4ae91487
3
+ size 87612
out_tensor/model.layers.11.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:451ad264ef1abf7cd249e3814b0bd464b9da2acc06aeefa8226b12f3fc7dc8c3
3
- size 398632
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a4774d491065db90b70e619e961fb32e7daa1e3a49ca79d7f48b7fb9a192b1f
3
+ size 395842
out_tensor/model.layers.11.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:73836289c08ad5b7ee854f34a984c4e4f0f58d04d854a129964e5c0dfdd341fb
3
- size 360820
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67db48f5754d31acf835016b4104cf11cb24377d2180e3583f73c2f82b0acad0
3
+ size 363038
out_tensor/model.layers.11.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef0dd1fc4c03f438225d6cd536e864010c8f4e0296e7ca93029af181adeb5d94
3
- size 379244
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cda2c19e5ad0b12b1209c3366676220530ffbe3632341ce55e2a5af4e991a265
3
+ size 387606
out_tensor/model.layers.11.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1c1d18f378a37564c5b0098530918ec2f54f90afd95218b28e5b32e97596ce98
3
- size 75526
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:67f6a5bba9db03f095850355740399329783f04bcc7f9156e0d7dac0d81d932e
3
+ size 87612
out_tensor/model.layers.11.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ecb1c12c2829afcf24c21cb36ed16e88a5d6a329c08ca24460e21dbd3a5f6e7c
3
- size 220878
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0194a099426cb0bca691fd745deef804210e613abf061d8ca92b5845be6a389b
3
+ size 256956
out_tensor/model.layers.11.self_attn.q_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9d01296dfe6166f3c07eaefd7d66984c0ffe86d16a8f9b509424fe79aca94abc
3
- size 220878
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f8076ee358670d24c330e78f9afce938888b911505d8ee3460dcfc0c7a14f16a
3
+ size 256956
out_tensor/model.layers.11.self_attn.v_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:af0467497cae04fe932dfb50ee1ec25c09da1dc81b3783ee4ddb494262f50a74
3
- size 86286
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:31af5f20948529d81afd585e97530db8c60144fed28093fb8f5db5d407b57d85
3
+ size 115260
out_tensor/model.layers.12.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:09c46c474e5a6ed96bc46dd789b1b519ae1e33ef39c0e5aabe658d246f0e634d
3
- size 395842
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ed46b77c0bc2988c313c255091c65acc8cc59a8d05ec50b9aa5f63ed58e73ac2
3
+ size 366376
out_tensor/model.layers.12.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c88ed3d13efa984b39f8349c7afec650d477753c51c5569660587a8524bc7f6a
3
- size 363038
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b493b4a727dba8ac8b8230f0eaefce9a4dbfda010828a0154d5914b46d6e521f
3
+ size 243254
out_tensor/model.layers.12.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64081e06e568ab961269178f3fdb5e5902b77c620c8d1aa1f6050d823cb33cf7
3
- size 387606
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:871fb605c1b775b7e2d02d90b176e00b020fc2ca32ea0876ae7495d415b7a4f4
3
+ size 267822
out_tensor/model.layers.12.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6ce77fbfc6e0a78017e6c92b2fea5f1ed0ab406dc96253d4f8df55646c4152c1
3
  size 47878
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:53f10533ac283494d2c2bbe892ca4dc7d5073c1c12b4568e1814de94dd1ff96d
3
  size 47878
out_tensor/model.layers.12.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:6e26feb6c8194733fa00c64aaf1bdf79ec675ba1cbbabbca062b7c2ce3c19b86
3
  size 137934
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99755f5f45d5d503853951ce3d3d2edebdc7b2493d6acd71db123a63c4467c1d
3
  size 137934
out_tensor/model.layers.12.self_attn.q_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7c1f9c2f0219bb4cac4882d3bfafdbe596c86a9e727e38430a659d633036d992
3
  size 137934
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3eaa8f4fa56d69177cb934fc01aa651b981285d39c125bef8fa66674608fab8c
3
  size 137934
out_tensor/model.layers.12.self_attn.v_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0558a2168156d0ef887cb2920f34e8bdefb0f19f1240e04d812bb69f98c69d10
3
  size 47878
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6f48ea439ba50a51c0d91c08de3fd6e9b76ec47a4dd59f30863dae0072375463
3
  size 47878
out_tensor/model.layers.13.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4f9d6a02997ac31f7efae7345a8855ad97e6daa012afe68f9cc90b76a9ba1f0a
3
  size 366376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:737d63835c5aca628c5a7e321483998228baa6785801a116eee92bcf91ebb13a
3
  size 366376
out_tensor/model.layers.13.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7a2e32c5203c5542d1e805df1ac494eda460881fcb9c9bab7ae8c57a2619612d
3
  size 243254
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ff3791d92088b0ca1c4301061f204b8e3770c9f1ced43c9ddb517a1ab1d743ed
3
  size 243254
out_tensor/model.layers.13.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8b0abe3026c6b5c698b0a5851b2a2f7ea1a38f2bab8f31a3d1f155aaae5a1ddf
3
  size 267822
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:09ebef69a690bf0c60ff790aea58aa59d30a1381c677366978519de7ff004b41
3
  size 267822
out_tensor/model.layers.13.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bf7caeedcd3dfbef48c58148881d22b4c8838538fedd695400f39fc6889cf2f7
3
- size 60574
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fae7e013ba28b81c3855c1081d39965240b57e748411513785d6a1da13213455
3
+ size 61500
out_tensor/model.layers.13.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e3d417acf01df3853d33b28a41c5f5d554fec37d748ade70e345d3228971c6e6
3
- size 175974
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:da1189fdc505da691e74d795e5cfd7414f9f130272c37a90268a03507ef54dd3
3
+ size 178620
out_tensor/model.layers.13.self_attn.q_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:15f653b12c7c60874128cd064fff6130be570cbb0c99eea6a5b10c0bae956c5f
3
- size 175974
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f9acee0cc9d7b56d3fe6170865b58a18d13234897678522dccd47ed0268364f6
3
+ size 178620
out_tensor/model.layers.13.self_attn.v_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:66df92a77b88377e3f735de1e31caf9420e745871f54246742aa21fe58a54fa8
3
- size 61500
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:028771ce4245f1bbd1cc1e491e8b347294794dbad7b6e886d2c42d555109b6bf
3
+ size 73788
out_tensor/model.layers.14.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2dc8ae11973b7a63b21477589ff3784464324998ad8286f36504fd0c1291566a
3
- size 366376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:75f2ef701d3132beb3dd02868cbd0d78324bffb99630e317b8fe2ce36881af39
3
+ size 398632
out_tensor/model.layers.14.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:05a4577ce30134c555f82d98264f51ab9a78be27f77b71e5273f3c4173a85474
3
- size 243254
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b1b5396c1ababf6b1dcfa47dee5201903a5c3fce4e0bd371a0ae8af2ca0346bc
3
+ size 360820
out_tensor/model.layers.14.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c1fb7a98964b9c971e2fac9720f04387b2e3e6ea235e4360c304a6537551063a
3
- size 267822
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c92070124ca1ad26746aa06bd01b33d417b168299fc53de3d9a78dcc612b3e7
3
+ size 379244
out_tensor/model.layers.14.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d40df8f718ffe2ead747218791673adb7d771a2f59100893c2cf23c891fcc90d
3
- size 75324
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ceddc15d97327cc721f0d0a618f706eca97d5681e842c6cae7f9044aa69337db
3
+ size 59038
out_tensor/model.layers.14.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:734712ed7e83535b37fe888c623bd12aedc657c16324039552c5226713bbfdf9
3
- size 220092
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8390a3cd72112867844755fa9873aba89a3eadf1b1c943f0cac917df14840b6e
3
+ size 171366
out_tensor/model.layers.14.self_attn.q_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:524f6d68da8992a2294782b46657814703725ea988ecbd950e6f20929d743712
3
- size 220092
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:88ef8213b80f02857314bbdbf8f9179d0673359b17ab3e8ae939833010cae44b
3
+ size 171366
out_tensor/model.layers.14.self_attn.v_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:13ebf7107da18ef3d913b69d3c95419231f2c7d878de621a5ce2798ca7da34e6
3
- size 87612
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f1b8da419b6b51187373d5a25600ea36e54ff0df8de806549707493dd782bfaa
3
+ size 59964
out_tensor/model.layers.15.mlp.down_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ee330925e438adaa98122fc78baf585a4e8feb8ad8dcc7f438093efb8c478d1b
3
- size 366376
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ebe548bd61bcd705769456e9f78ae96eeb56c02de700526e1d93821565ba4724
3
+ size 395842
out_tensor/model.layers.15.mlp.gate_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0cf3f2e20a4aced3b35b4719a6c350a2f10840267a92f3f3300d915f3b5e9122
3
- size 243254
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1e54257e0c2e223ab062386de204e58f4c32e80b325c92261f8d29bd4c67d316
3
+ size 363038
out_tensor/model.layers.15.mlp.up_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ef1dc805d261ad85a118c1800680c5604b0df83435e914e576da180776318a84
3
- size 267822
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46fe76eb5fefb7aca451179e64ea7526fabdc94c128ce48cee5b3fd69309507b
3
+ size 387606
out_tensor/model.layers.15.self_attn.k_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a53a804b8c637fdc92ee5bc3afdc25a27e2289ed0b853fb0bd7ce1eaa17be5f4
3
- size 61702
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7b78c9cbd3001fbfea4bd679ec48ee931583b2e11daf1f7b3624781e9d8b77e
3
+ size 61500
out_tensor/model.layers.15.self_attn.o_proj.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a69f8474ca0b85c02299ac727ed67883e4ed1d5730ce00933aecaaa133a90a07
3
- size 179406
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:80a3fffdb056e6da72b19011e017ef18bd310c5e613bfafd17e3819c9da94ddf
3
+ size 178620