acbdkk commited on
Commit
3e9f4e1
·
verified ·
1 Parent(s): 0926af9

Trained with Unsloth

Browse files

Upload model trained with Unsloth 2x faster

config.json CHANGED
@@ -1,23 +1,22 @@
1
  {
2
- "_name_or_path": "unsloth/llama-3-8b-bnb-4bit",
3
  "architectures": [
4
- "LlamaForCausalLM"
5
  ],
6
- "attention_bias": false,
7
  "attention_dropout": 0.0,
8
- "bos_token_id": 128000,
9
- "eos_token_id": 128001,
10
  "hidden_act": "silu",
11
- "hidden_size": 4096,
12
  "initializer_range": 0.02,
13
- "intermediate_size": 14336,
14
- "max_position_embeddings": 8192,
15
- "mlp_bias": false,
16
- "model_type": "llama",
17
- "num_attention_heads": 32,
18
- "num_hidden_layers": 32,
19
- "num_key_value_heads": 8,
20
- "pretraining_tp": 1,
21
  "quantization_config": {
22
  "bnb_4bit_compute_dtype": "float16",
23
  "bnb_4bit_quant_type": "nf4",
@@ -30,13 +29,15 @@
30
  "load_in_8bit": false,
31
  "quant_method": "bitsandbytes"
32
  },
33
- "rms_norm_eps": 1e-05,
34
  "rope_scaling": null,
35
- "rope_theta": 500000.0,
 
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "float16",
38
- "transformers_version": "4.41.2",
39
- "unsloth_version": "2024.6",
40
  "use_cache": true,
41
- "vocab_size": 128256
 
42
  }
 
1
  {
2
+ "_name_or_path": "unsloth/qwen2-7b-bnb-4bit",
3
  "architectures": [
4
+ "Qwen2ForCausalLM"
5
  ],
 
6
  "attention_dropout": 0.0,
7
+ "bos_token_id": 151643,
8
+ "eos_token_id": 151643,
9
  "hidden_act": "silu",
10
+ "hidden_size": 3584,
11
  "initializer_range": 0.02,
12
+ "intermediate_size": 18944,
13
+ "max_position_embeddings": 131072,
14
+ "max_window_layers": 28,
15
+ "model_type": "qwen2",
16
+ "num_attention_heads": 28,
17
+ "num_hidden_layers": 28,
18
+ "num_key_value_heads": 4,
19
+ "pad_token_id": 151646,
20
  "quantization_config": {
21
  "bnb_4bit_compute_dtype": "float16",
22
  "bnb_4bit_quant_type": "nf4",
 
29
  "load_in_8bit": false,
30
  "quant_method": "bitsandbytes"
31
  },
32
+ "rms_norm_eps": 1e-06,
33
  "rope_scaling": null,
34
+ "rope_theta": 1000000.0,
35
+ "sliding_window": 131072,
36
  "tie_word_embeddings": false,
37
  "torch_dtype": "float16",
38
+ "transformers_version": "4.42.3",
39
+ "unsloth_version": "2024.7",
40
  "use_cache": true,
41
+ "use_sliding_window": false,
42
+ "vocab_size": 152064
43
  }
generation_config.json CHANGED
@@ -1,9 +1,6 @@
1
  {
2
- "bos_token_id": 128000,
3
- "do_sample": true,
4
- "eos_token_id": 128001,
5
- "max_length": 4096,
6
- "temperature": 0.6,
7
- "top_p": 0.9,
8
- "transformers_version": "4.41.2"
9
  }
 
1
  {
2
+ "bos_token_id": 151643,
3
+ "eos_token_id": 151643,
4
+ "max_new_tokens": 2048,
5
+ "transformers_version": "4.42.3"
 
 
 
6
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f34bf30f73bf3e1da0d0986786b1c1c360b8bafd84e08847947f7abbb13657f3
3
- size 4652072861
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5883841ab0093b81c8255c984b2b10b04d4f665110027622675bb78f97eaaa7e
3
+ size 4457259441
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e4f15e64381b74dcb7857c03cbfb62806534f7582ab885c738980a12e5b78989
3
- size 1050673280
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a17b24a36727400faebe507c4efd3192abb501c59965c969027a6676c70902e5
3
+ size 1089994880
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 5702577741
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
@@ -25,6 +25,7 @@
25
  "model.layers.0.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
26
  "model.layers.0.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
27
  "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
28
  "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
29
  "model.layers.0.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
30
  "model.layers.0.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -37,12 +38,14 @@
37
  "model.layers.0.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
38
  "model.layers.0.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
39
  "model.layers.0.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
40
  "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
41
  "model.layers.0.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
42
  "model.layers.0.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
43
  "model.layers.0.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
44
  "model.layers.0.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
45
  "model.layers.0.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
46
  "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
47
  "model.layers.0.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
48
  "model.layers.0.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -69,6 +72,7 @@
69
  "model.layers.1.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
70
  "model.layers.1.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
71
  "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
72
  "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
73
  "model.layers.1.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
74
  "model.layers.1.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -81,12 +85,14 @@
81
  "model.layers.1.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
82
  "model.layers.1.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
83
  "model.layers.1.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
84
  "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
85
  "model.layers.1.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
86
  "model.layers.1.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
87
  "model.layers.1.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
88
  "model.layers.1.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
89
  "model.layers.1.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
90
  "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
91
  "model.layers.1.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
92
  "model.layers.1.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -113,6 +119,7 @@
113
  "model.layers.10.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
114
  "model.layers.10.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
115
  "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
116
  "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
117
  "model.layers.10.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
118
  "model.layers.10.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -125,12 +132,14 @@
125
  "model.layers.10.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
126
  "model.layers.10.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
127
  "model.layers.10.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
128
  "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
129
  "model.layers.10.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
130
  "model.layers.10.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
131
  "model.layers.10.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
132
  "model.layers.10.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
133
  "model.layers.10.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
134
  "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
135
  "model.layers.10.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
136
  "model.layers.10.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -157,6 +166,7 @@
157
  "model.layers.11.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
158
  "model.layers.11.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
159
  "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
160
  "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
161
  "model.layers.11.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
162
  "model.layers.11.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -169,12 +179,14 @@
169
  "model.layers.11.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
170
  "model.layers.11.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
171
  "model.layers.11.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
172
  "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
173
  "model.layers.11.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
174
  "model.layers.11.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
175
  "model.layers.11.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
176
  "model.layers.11.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
177
  "model.layers.11.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
178
  "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
179
  "model.layers.11.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
180
  "model.layers.11.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -201,6 +213,7 @@
201
  "model.layers.12.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
202
  "model.layers.12.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
203
  "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
204
  "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
205
  "model.layers.12.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
206
  "model.layers.12.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -213,12 +226,14 @@
213
  "model.layers.12.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
214
  "model.layers.12.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
215
  "model.layers.12.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
216
  "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
217
  "model.layers.12.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
218
  "model.layers.12.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
219
  "model.layers.12.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
220
  "model.layers.12.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
221
  "model.layers.12.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
222
  "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
223
  "model.layers.12.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
224
  "model.layers.12.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -245,6 +260,7 @@
245
  "model.layers.13.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
246
  "model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
247
  "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
248
  "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
249
  "model.layers.13.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
250
  "model.layers.13.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -257,12 +273,14 @@
257
  "model.layers.13.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
258
  "model.layers.13.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
259
  "model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
260
  "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
261
  "model.layers.13.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
262
  "model.layers.13.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
263
  "model.layers.13.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
264
  "model.layers.13.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
265
  "model.layers.13.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
266
  "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
267
  "model.layers.13.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
268
  "model.layers.13.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -289,6 +307,7 @@
289
  "model.layers.14.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
290
  "model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
291
  "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
292
  "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
293
  "model.layers.14.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
294
  "model.layers.14.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -301,12 +320,14 @@
301
  "model.layers.14.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
302
  "model.layers.14.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
303
  "model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
304
  "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
305
  "model.layers.14.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
306
  "model.layers.14.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
307
  "model.layers.14.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
308
  "model.layers.14.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
309
  "model.layers.14.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
310
  "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
311
  "model.layers.14.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
312
  "model.layers.14.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -333,6 +354,7 @@
333
  "model.layers.15.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
334
  "model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
335
  "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
336
  "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
337
  "model.layers.15.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
338
  "model.layers.15.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -345,12 +367,14 @@
345
  "model.layers.15.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
346
  "model.layers.15.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
347
  "model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
348
  "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
349
  "model.layers.15.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
350
  "model.layers.15.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
351
  "model.layers.15.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
352
  "model.layers.15.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
353
  "model.layers.15.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
354
  "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
355
  "model.layers.15.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
356
  "model.layers.15.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -377,6 +401,7 @@
377
  "model.layers.16.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
378
  "model.layers.16.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
379
  "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
380
  "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
381
  "model.layers.16.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
382
  "model.layers.16.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -389,12 +414,14 @@
389
  "model.layers.16.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
390
  "model.layers.16.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
391
  "model.layers.16.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
392
  "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
393
  "model.layers.16.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
394
  "model.layers.16.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
395
  "model.layers.16.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
396
  "model.layers.16.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
397
  "model.layers.16.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
398
  "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
399
  "model.layers.16.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
400
  "model.layers.16.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -421,6 +448,7 @@
421
  "model.layers.17.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
422
  "model.layers.17.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
423
  "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
424
  "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
425
  "model.layers.17.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
426
  "model.layers.17.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -433,12 +461,14 @@
433
  "model.layers.17.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
434
  "model.layers.17.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
435
  "model.layers.17.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
436
  "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
437
  "model.layers.17.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
438
  "model.layers.17.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
439
  "model.layers.17.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
440
  "model.layers.17.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
441
  "model.layers.17.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
442
  "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
443
  "model.layers.17.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
444
  "model.layers.17.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -465,6 +495,7 @@
465
  "model.layers.18.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
466
  "model.layers.18.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
467
  "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
468
  "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
469
  "model.layers.18.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
470
  "model.layers.18.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -477,12 +508,14 @@
477
  "model.layers.18.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
478
  "model.layers.18.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
479
  "model.layers.18.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
480
  "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
481
  "model.layers.18.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
482
  "model.layers.18.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
483
  "model.layers.18.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
484
  "model.layers.18.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
485
  "model.layers.18.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
486
  "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
487
  "model.layers.18.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
488
  "model.layers.18.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -509,6 +542,7 @@
509
  "model.layers.19.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
510
  "model.layers.19.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
511
  "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
512
  "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
513
  "model.layers.19.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
514
  "model.layers.19.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -521,12 +555,14 @@
521
  "model.layers.19.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
522
  "model.layers.19.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
523
  "model.layers.19.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
524
  "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
525
  "model.layers.19.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
526
  "model.layers.19.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
527
  "model.layers.19.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
528
  "model.layers.19.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
529
  "model.layers.19.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
530
  "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
531
  "model.layers.19.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
532
  "model.layers.19.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -553,6 +589,7 @@
553
  "model.layers.2.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
554
  "model.layers.2.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
555
  "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
556
  "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
557
  "model.layers.2.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
558
  "model.layers.2.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -565,12 +602,14 @@
565
  "model.layers.2.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
566
  "model.layers.2.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
567
  "model.layers.2.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
568
  "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
569
  "model.layers.2.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
570
  "model.layers.2.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
571
  "model.layers.2.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
572
  "model.layers.2.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
573
  "model.layers.2.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
574
  "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
575
  "model.layers.2.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
576
  "model.layers.2.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -597,6 +636,7 @@
597
  "model.layers.20.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
598
  "model.layers.20.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
599
  "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
600
  "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
601
  "model.layers.20.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
602
  "model.layers.20.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -609,12 +649,14 @@
609
  "model.layers.20.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
610
  "model.layers.20.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
611
  "model.layers.20.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
612
  "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
613
  "model.layers.20.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
614
  "model.layers.20.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
615
  "model.layers.20.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
616
  "model.layers.20.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
617
  "model.layers.20.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
618
  "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
619
  "model.layers.20.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
620
  "model.layers.20.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -641,6 +683,7 @@
641
  "model.layers.21.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
642
  "model.layers.21.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
643
  "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
644
  "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
645
  "model.layers.21.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
646
  "model.layers.21.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -653,12 +696,14 @@
653
  "model.layers.21.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
654
  "model.layers.21.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
655
  "model.layers.21.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
656
  "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
657
  "model.layers.21.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
658
  "model.layers.21.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
659
  "model.layers.21.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
660
  "model.layers.21.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
661
  "model.layers.21.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
662
  "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
663
  "model.layers.21.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
664
  "model.layers.21.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -685,6 +730,7 @@
685
  "model.layers.22.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
686
  "model.layers.22.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
687
  "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
688
  "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
689
  "model.layers.22.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
690
  "model.layers.22.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -697,12 +743,14 @@
697
  "model.layers.22.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
698
  "model.layers.22.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
699
  "model.layers.22.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
700
  "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
701
  "model.layers.22.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
702
  "model.layers.22.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
703
  "model.layers.22.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
704
  "model.layers.22.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
705
  "model.layers.22.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
706
  "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
707
  "model.layers.22.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
708
  "model.layers.22.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -729,6 +777,7 @@
729
  "model.layers.23.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
730
  "model.layers.23.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
731
  "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
732
  "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
733
  "model.layers.23.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
734
  "model.layers.23.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -741,12 +790,14 @@
741
  "model.layers.23.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
742
  "model.layers.23.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
743
  "model.layers.23.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
744
  "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
745
  "model.layers.23.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
746
  "model.layers.23.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
747
  "model.layers.23.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
748
  "model.layers.23.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
749
  "model.layers.23.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
750
  "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
751
  "model.layers.23.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
752
  "model.layers.23.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -773,6 +824,7 @@
773
  "model.layers.24.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
774
  "model.layers.24.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
775
  "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
776
  "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
777
  "model.layers.24.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
778
  "model.layers.24.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -785,12 +837,14 @@
785
  "model.layers.24.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
786
  "model.layers.24.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
787
  "model.layers.24.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
788
  "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
789
  "model.layers.24.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
790
  "model.layers.24.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
791
  "model.layers.24.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
792
  "model.layers.24.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
793
  "model.layers.24.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
794
  "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
795
  "model.layers.24.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
796
  "model.layers.24.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -817,6 +871,7 @@
817
  "model.layers.25.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
818
  "model.layers.25.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
819
  "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
820
  "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
821
  "model.layers.25.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
822
  "model.layers.25.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -829,12 +884,14 @@
829
  "model.layers.25.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
830
  "model.layers.25.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
831
  "model.layers.25.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
832
  "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
833
  "model.layers.25.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
834
  "model.layers.25.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
835
  "model.layers.25.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
836
  "model.layers.25.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
837
  "model.layers.25.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
838
  "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
839
  "model.layers.25.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
840
  "model.layers.25.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -861,6 +918,7 @@
861
  "model.layers.26.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
862
  "model.layers.26.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
863
  "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
864
  "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
865
  "model.layers.26.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
866
  "model.layers.26.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -873,12 +931,14 @@
873
  "model.layers.26.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
874
  "model.layers.26.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
875
  "model.layers.26.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
876
  "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
877
  "model.layers.26.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
878
  "model.layers.26.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
879
  "model.layers.26.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
880
  "model.layers.26.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
881
  "model.layers.26.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
882
  "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
883
  "model.layers.26.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
884
  "model.layers.26.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -905,6 +965,7 @@
905
  "model.layers.27.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
906
  "model.layers.27.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
907
  "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
908
  "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
909
  "model.layers.27.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
910
  "model.layers.27.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -917,106 +978,20 @@
917
  "model.layers.27.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
918
  "model.layers.27.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
919
  "model.layers.27.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
920
  "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
921
  "model.layers.27.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
922
  "model.layers.27.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
923
  "model.layers.27.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
924
  "model.layers.27.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
925
  "model.layers.27.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
926
  "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
927
  "model.layers.27.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
928
  "model.layers.27.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
929
  "model.layers.27.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
930
  "model.layers.27.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
931
  "model.layers.27.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
932
- "model.layers.28.input_layernorm.weight": "model-00001-of-00002.safetensors",
933
- "model.layers.28.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
934
- "model.layers.28.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
935
- "model.layers.28.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
936
- "model.layers.28.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
937
- "model.layers.28.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
938
- "model.layers.28.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
939
- "model.layers.28.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
940
- "model.layers.28.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
941
- "model.layers.28.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
942
- "model.layers.28.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
943
- "model.layers.28.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
944
- "model.layers.28.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
945
- "model.layers.28.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
946
- "model.layers.28.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
947
- "model.layers.28.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
948
- "model.layers.28.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
949
- "model.layers.28.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
950
- "model.layers.28.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
951
- "model.layers.28.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
952
- "model.layers.28.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
953
- "model.layers.28.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
954
- "model.layers.28.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
955
- "model.layers.28.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
956
- "model.layers.28.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
957
- "model.layers.28.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
958
- "model.layers.28.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
959
- "model.layers.28.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
960
- "model.layers.28.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
961
- "model.layers.28.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
962
- "model.layers.28.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
963
- "model.layers.28.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
964
- "model.layers.28.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
965
- "model.layers.28.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
966
- "model.layers.28.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
967
- "model.layers.28.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
968
- "model.layers.28.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
969
- "model.layers.28.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
970
- "model.layers.28.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
971
- "model.layers.28.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
972
- "model.layers.28.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
973
- "model.layers.28.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
974
- "model.layers.28.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
975
- "model.layers.28.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
976
- "model.layers.29.input_layernorm.weight": "model-00001-of-00002.safetensors",
977
- "model.layers.29.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
978
- "model.layers.29.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
979
- "model.layers.29.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
980
- "model.layers.29.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
981
- "model.layers.29.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
982
- "model.layers.29.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
983
- "model.layers.29.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
984
- "model.layers.29.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
985
- "model.layers.29.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
986
- "model.layers.29.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
987
- "model.layers.29.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
988
- "model.layers.29.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
989
- "model.layers.29.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
990
- "model.layers.29.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
991
- "model.layers.29.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
992
- "model.layers.29.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
993
- "model.layers.29.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
994
- "model.layers.29.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
995
- "model.layers.29.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
996
- "model.layers.29.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
997
- "model.layers.29.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
998
- "model.layers.29.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
999
- "model.layers.29.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1000
- "model.layers.29.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1001
- "model.layers.29.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1002
- "model.layers.29.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1003
- "model.layers.29.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1004
- "model.layers.29.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1005
- "model.layers.29.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1006
- "model.layers.29.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1007
- "model.layers.29.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1008
- "model.layers.29.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1009
- "model.layers.29.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1010
- "model.layers.29.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1011
- "model.layers.29.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1012
- "model.layers.29.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1013
- "model.layers.29.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1014
- "model.layers.29.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1015
- "model.layers.29.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1016
- "model.layers.29.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1017
- "model.layers.29.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1018
- "model.layers.29.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1019
- "model.layers.29.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1020
  "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
1021
  "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1022
  "model.layers.3.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -1037,6 +1012,7 @@
1037
  "model.layers.3.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1038
  "model.layers.3.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1039
  "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1040
  "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1041
  "model.layers.3.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1042
  "model.layers.3.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1049,106 +1025,20 @@
1049
  "model.layers.3.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1050
  "model.layers.3.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1051
  "model.layers.3.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1052
  "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1053
  "model.layers.3.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1054
  "model.layers.3.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1055
  "model.layers.3.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1056
  "model.layers.3.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1057
  "model.layers.3.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1058
  "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1059
  "model.layers.3.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1060
  "model.layers.3.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1061
  "model.layers.3.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1062
  "model.layers.3.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1063
  "model.layers.3.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1064
- "model.layers.30.input_layernorm.weight": "model-00001-of-00002.safetensors",
1065
- "model.layers.30.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1066
- "model.layers.30.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1067
- "model.layers.30.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1068
- "model.layers.30.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1069
- "model.layers.30.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1070
- "model.layers.30.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1071
- "model.layers.30.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1072
- "model.layers.30.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1073
- "model.layers.30.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1074
- "model.layers.30.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1075
- "model.layers.30.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1076
- "model.layers.30.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1077
- "model.layers.30.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1078
- "model.layers.30.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1079
- "model.layers.30.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1080
- "model.layers.30.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1081
- "model.layers.30.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1082
- "model.layers.30.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1083
- "model.layers.30.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1084
- "model.layers.30.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1085
- "model.layers.30.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1086
- "model.layers.30.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1087
- "model.layers.30.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1088
- "model.layers.30.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1089
- "model.layers.30.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1090
- "model.layers.30.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1091
- "model.layers.30.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1092
- "model.layers.30.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1093
- "model.layers.30.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1094
- "model.layers.30.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1095
- "model.layers.30.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1096
- "model.layers.30.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1097
- "model.layers.30.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1098
- "model.layers.30.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1099
- "model.layers.30.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1100
- "model.layers.30.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1101
- "model.layers.30.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1102
- "model.layers.30.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1103
- "model.layers.30.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1104
- "model.layers.30.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1105
- "model.layers.30.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1106
- "model.layers.30.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1107
- "model.layers.30.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1108
- "model.layers.31.input_layernorm.weight": "model-00001-of-00002.safetensors",
1109
- "model.layers.31.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1110
- "model.layers.31.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1111
- "model.layers.31.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1112
- "model.layers.31.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1113
- "model.layers.31.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1114
- "model.layers.31.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1115
- "model.layers.31.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1116
- "model.layers.31.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1117
- "model.layers.31.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1118
- "model.layers.31.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1119
- "model.layers.31.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1120
- "model.layers.31.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1121
- "model.layers.31.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1122
- "model.layers.31.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1123
- "model.layers.31.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1124
- "model.layers.31.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1125
- "model.layers.31.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1126
- "model.layers.31.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1127
- "model.layers.31.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1128
- "model.layers.31.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1129
- "model.layers.31.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1130
- "model.layers.31.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1131
- "model.layers.31.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1132
- "model.layers.31.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1133
- "model.layers.31.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1134
- "model.layers.31.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1135
- "model.layers.31.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1136
- "model.layers.31.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1137
- "model.layers.31.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1138
- "model.layers.31.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1139
- "model.layers.31.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1140
- "model.layers.31.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1141
- "model.layers.31.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1142
- "model.layers.31.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1143
- "model.layers.31.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1144
- "model.layers.31.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1145
- "model.layers.31.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1146
- "model.layers.31.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1147
- "model.layers.31.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1148
- "model.layers.31.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1149
- "model.layers.31.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1150
- "model.layers.31.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1151
- "model.layers.31.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1152
  "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
1153
  "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1154
  "model.layers.4.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -1169,6 +1059,7 @@
1169
  "model.layers.4.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1170
  "model.layers.4.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1171
  "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1172
  "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1173
  "model.layers.4.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1174
  "model.layers.4.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1181,12 +1072,14 @@
1181
  "model.layers.4.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1182
  "model.layers.4.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1183
  "model.layers.4.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1184
  "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1185
  "model.layers.4.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1186
  "model.layers.4.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1187
  "model.layers.4.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1188
  "model.layers.4.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1189
  "model.layers.4.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1190
  "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1191
  "model.layers.4.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1192
  "model.layers.4.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1213,6 +1106,7 @@
1213
  "model.layers.5.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1214
  "model.layers.5.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1215
  "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1216
  "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1217
  "model.layers.5.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1218
  "model.layers.5.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1225,12 +1119,14 @@
1225
  "model.layers.5.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1226
  "model.layers.5.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1227
  "model.layers.5.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1228
  "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1229
  "model.layers.5.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1230
  "model.layers.5.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1231
  "model.layers.5.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1232
  "model.layers.5.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1233
  "model.layers.5.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1234
  "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1235
  "model.layers.5.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1236
  "model.layers.5.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1257,6 +1153,7 @@
1257
  "model.layers.6.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1258
  "model.layers.6.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1259
  "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1260
  "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1261
  "model.layers.6.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1262
  "model.layers.6.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1269,12 +1166,14 @@
1269
  "model.layers.6.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1270
  "model.layers.6.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1271
  "model.layers.6.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1272
  "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1273
  "model.layers.6.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1274
  "model.layers.6.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1275
  "model.layers.6.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1276
  "model.layers.6.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1277
  "model.layers.6.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1278
  "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1279
  "model.layers.6.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1280
  "model.layers.6.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1301,6 +1200,7 @@
1301
  "model.layers.7.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1302
  "model.layers.7.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1303
  "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1304
  "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1305
  "model.layers.7.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1306
  "model.layers.7.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1313,12 +1213,14 @@
1313
  "model.layers.7.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1314
  "model.layers.7.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1315
  "model.layers.7.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1316
  "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1317
  "model.layers.7.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1318
  "model.layers.7.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1319
  "model.layers.7.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1320
  "model.layers.7.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1321
  "model.layers.7.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1322
  "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1323
  "model.layers.7.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1324
  "model.layers.7.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1345,6 +1247,7 @@
1345
  "model.layers.8.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1346
  "model.layers.8.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1347
  "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1348
  "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1349
  "model.layers.8.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1350
  "model.layers.8.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1357,12 +1260,14 @@
1357
  "model.layers.8.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1358
  "model.layers.8.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1359
  "model.layers.8.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1360
  "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1361
  "model.layers.8.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1362
  "model.layers.8.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1363
  "model.layers.8.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1364
  "model.layers.8.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1365
  "model.layers.8.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1366
  "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1367
  "model.layers.8.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1368
  "model.layers.8.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1389,6 +1294,7 @@
1389
  "model.layers.9.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1390
  "model.layers.9.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1391
  "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
 
1392
  "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1393
  "model.layers.9.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1394
  "model.layers.9.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1401,12 +1307,14 @@
1401
  "model.layers.9.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1402
  "model.layers.9.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1403
  "model.layers.9.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1404
  "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1405
  "model.layers.9.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1406
  "model.layers.9.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1407
  "model.layers.9.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1408
  "model.layers.9.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1409
  "model.layers.9.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
1410
  "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1411
  "model.layers.9.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1412
  "model.layers.9.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1
  {
2
  "metadata": {
3
+ "total_size": 5547097905
4
  },
5
  "weight_map": {
6
  "lm_head.weight": "model-00002-of-00002.safetensors",
 
25
  "model.layers.0.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
26
  "model.layers.0.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
27
  "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
28
+ "model.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
29
  "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
30
  "model.layers.0.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
31
  "model.layers.0.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
38
  "model.layers.0.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
39
  "model.layers.0.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
40
  "model.layers.0.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
41
+ "model.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
42
  "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
43
  "model.layers.0.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
44
  "model.layers.0.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
45
  "model.layers.0.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
46
  "model.layers.0.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
47
  "model.layers.0.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
48
+ "model.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
49
  "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
50
  "model.layers.0.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
51
  "model.layers.0.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
72
  "model.layers.1.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
73
  "model.layers.1.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
74
  "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
75
+ "model.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
76
  "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
77
  "model.layers.1.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
78
  "model.layers.1.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
85
  "model.layers.1.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
86
  "model.layers.1.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
87
  "model.layers.1.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
88
+ "model.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
89
  "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
90
  "model.layers.1.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
91
  "model.layers.1.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
92
  "model.layers.1.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
93
  "model.layers.1.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
94
  "model.layers.1.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
95
+ "model.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
96
  "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
97
  "model.layers.1.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
98
  "model.layers.1.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
119
  "model.layers.10.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
120
  "model.layers.10.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
121
  "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
122
+ "model.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
123
  "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
124
  "model.layers.10.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
125
  "model.layers.10.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
132
  "model.layers.10.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
133
  "model.layers.10.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
134
  "model.layers.10.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
135
+ "model.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
136
  "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
137
  "model.layers.10.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
138
  "model.layers.10.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
139
  "model.layers.10.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
140
  "model.layers.10.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
141
  "model.layers.10.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
142
+ "model.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
143
  "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
144
  "model.layers.10.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
145
  "model.layers.10.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
166
  "model.layers.11.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
167
  "model.layers.11.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
168
  "model.layers.11.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
169
+ "model.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
170
  "model.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
171
  "model.layers.11.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
172
  "model.layers.11.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
179
  "model.layers.11.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
180
  "model.layers.11.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
181
  "model.layers.11.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
182
+ "model.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
183
  "model.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
184
  "model.layers.11.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
185
  "model.layers.11.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
186
  "model.layers.11.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
187
  "model.layers.11.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
188
  "model.layers.11.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
189
+ "model.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
190
  "model.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
191
  "model.layers.11.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
192
  "model.layers.11.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
213
  "model.layers.12.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
214
  "model.layers.12.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
215
  "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
216
+ "model.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
217
  "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
218
  "model.layers.12.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
219
  "model.layers.12.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
226
  "model.layers.12.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
227
  "model.layers.12.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
228
  "model.layers.12.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
229
+ "model.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
230
  "model.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
231
  "model.layers.12.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
232
  "model.layers.12.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
233
  "model.layers.12.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
234
  "model.layers.12.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
235
  "model.layers.12.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
236
+ "model.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
237
  "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
238
  "model.layers.12.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
239
  "model.layers.12.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
260
  "model.layers.13.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
261
  "model.layers.13.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
262
  "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
263
+ "model.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
264
  "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
265
  "model.layers.13.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
266
  "model.layers.13.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
273
  "model.layers.13.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
274
  "model.layers.13.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
275
  "model.layers.13.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
276
+ "model.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
277
  "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
278
  "model.layers.13.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
279
  "model.layers.13.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
280
  "model.layers.13.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
281
  "model.layers.13.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
282
  "model.layers.13.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
283
+ "model.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
284
  "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
285
  "model.layers.13.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
286
  "model.layers.13.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
307
  "model.layers.14.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
308
  "model.layers.14.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
309
  "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
310
+ "model.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
311
  "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
312
  "model.layers.14.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
313
  "model.layers.14.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
320
  "model.layers.14.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
321
  "model.layers.14.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
322
  "model.layers.14.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
323
+ "model.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
324
  "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
325
  "model.layers.14.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
326
  "model.layers.14.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
327
  "model.layers.14.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
328
  "model.layers.14.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
329
  "model.layers.14.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
330
+ "model.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
331
  "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
332
  "model.layers.14.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
333
  "model.layers.14.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
354
  "model.layers.15.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
355
  "model.layers.15.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
356
  "model.layers.15.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
357
+ "model.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
358
  "model.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
359
  "model.layers.15.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
360
  "model.layers.15.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
367
  "model.layers.15.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
368
  "model.layers.15.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
369
  "model.layers.15.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
370
+ "model.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
371
  "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
372
  "model.layers.15.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
373
  "model.layers.15.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
374
  "model.layers.15.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
375
  "model.layers.15.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
376
  "model.layers.15.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
377
+ "model.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
378
  "model.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
379
  "model.layers.15.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
380
  "model.layers.15.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
401
  "model.layers.16.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
402
  "model.layers.16.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
403
  "model.layers.16.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
404
+ "model.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
405
  "model.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
406
  "model.layers.16.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
407
  "model.layers.16.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
414
  "model.layers.16.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
415
  "model.layers.16.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
416
  "model.layers.16.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
417
+ "model.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
418
  "model.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
419
  "model.layers.16.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
420
  "model.layers.16.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
421
  "model.layers.16.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
422
  "model.layers.16.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
423
  "model.layers.16.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
424
+ "model.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
425
  "model.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
426
  "model.layers.16.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
427
  "model.layers.16.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
448
  "model.layers.17.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
449
  "model.layers.17.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
450
  "model.layers.17.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
451
+ "model.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
452
  "model.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
453
  "model.layers.17.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
454
  "model.layers.17.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
461
  "model.layers.17.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
462
  "model.layers.17.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
463
  "model.layers.17.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
464
+ "model.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
465
  "model.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
466
  "model.layers.17.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
467
  "model.layers.17.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
468
  "model.layers.17.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
469
  "model.layers.17.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
470
  "model.layers.17.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
471
+ "model.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
472
  "model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
473
  "model.layers.17.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
474
  "model.layers.17.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
495
  "model.layers.18.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
496
  "model.layers.18.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
497
  "model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
498
+ "model.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
499
  "model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
500
  "model.layers.18.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
501
  "model.layers.18.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
508
  "model.layers.18.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
509
  "model.layers.18.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
510
  "model.layers.18.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
511
+ "model.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
512
  "model.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
513
  "model.layers.18.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
514
  "model.layers.18.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
515
  "model.layers.18.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
516
  "model.layers.18.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
517
  "model.layers.18.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
518
+ "model.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
519
  "model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
520
  "model.layers.18.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
521
  "model.layers.18.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
542
  "model.layers.19.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
543
  "model.layers.19.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
544
  "model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
545
+ "model.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
546
  "model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
547
  "model.layers.19.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
548
  "model.layers.19.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
555
  "model.layers.19.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
556
  "model.layers.19.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
557
  "model.layers.19.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
558
+ "model.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
559
  "model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
560
  "model.layers.19.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
561
  "model.layers.19.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
562
  "model.layers.19.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
563
  "model.layers.19.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
564
  "model.layers.19.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
565
+ "model.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
566
  "model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
567
  "model.layers.19.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
568
  "model.layers.19.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
589
  "model.layers.2.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
590
  "model.layers.2.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
591
  "model.layers.2.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
592
+ "model.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
593
  "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
594
  "model.layers.2.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
595
  "model.layers.2.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
602
  "model.layers.2.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
603
  "model.layers.2.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
604
  "model.layers.2.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
605
+ "model.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
606
  "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
607
  "model.layers.2.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
608
  "model.layers.2.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
609
  "model.layers.2.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
610
  "model.layers.2.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
611
  "model.layers.2.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
612
+ "model.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
613
  "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
614
  "model.layers.2.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
615
  "model.layers.2.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
636
  "model.layers.20.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
637
  "model.layers.20.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
638
  "model.layers.20.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
639
+ "model.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
640
  "model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
641
  "model.layers.20.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
642
  "model.layers.20.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
649
  "model.layers.20.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
650
  "model.layers.20.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
651
  "model.layers.20.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
652
+ "model.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
653
  "model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
654
  "model.layers.20.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
655
  "model.layers.20.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
656
  "model.layers.20.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
657
  "model.layers.20.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
658
  "model.layers.20.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
659
+ "model.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
660
  "model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
661
  "model.layers.20.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
662
  "model.layers.20.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
683
  "model.layers.21.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
684
  "model.layers.21.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
685
  "model.layers.21.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
686
+ "model.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
687
  "model.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
688
  "model.layers.21.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
689
  "model.layers.21.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
696
  "model.layers.21.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
697
  "model.layers.21.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
698
  "model.layers.21.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
699
+ "model.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
700
  "model.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
701
  "model.layers.21.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
702
  "model.layers.21.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
703
  "model.layers.21.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
704
  "model.layers.21.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
705
  "model.layers.21.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
706
+ "model.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
707
  "model.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
708
  "model.layers.21.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
709
  "model.layers.21.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
730
  "model.layers.22.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
731
  "model.layers.22.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
732
  "model.layers.22.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
733
+ "model.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
734
  "model.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
735
  "model.layers.22.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
736
  "model.layers.22.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
743
  "model.layers.22.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
744
  "model.layers.22.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
745
  "model.layers.22.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
746
+ "model.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
747
  "model.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
748
  "model.layers.22.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
749
  "model.layers.22.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
750
  "model.layers.22.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
751
  "model.layers.22.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
752
  "model.layers.22.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
753
+ "model.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
754
  "model.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
755
  "model.layers.22.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
756
  "model.layers.22.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
777
  "model.layers.23.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
778
  "model.layers.23.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
779
  "model.layers.23.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
780
+ "model.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
781
  "model.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
782
  "model.layers.23.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
783
  "model.layers.23.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
790
  "model.layers.23.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
791
  "model.layers.23.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
792
  "model.layers.23.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
793
+ "model.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
794
  "model.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
795
  "model.layers.23.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
796
  "model.layers.23.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
797
  "model.layers.23.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
798
  "model.layers.23.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
799
  "model.layers.23.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
800
+ "model.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
801
  "model.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
802
  "model.layers.23.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
803
  "model.layers.23.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
824
  "model.layers.24.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
825
  "model.layers.24.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
826
  "model.layers.24.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
827
+ "model.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
828
  "model.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
829
  "model.layers.24.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
830
  "model.layers.24.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
837
  "model.layers.24.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
838
  "model.layers.24.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
839
  "model.layers.24.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
840
+ "model.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
841
  "model.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
842
  "model.layers.24.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
843
  "model.layers.24.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
844
  "model.layers.24.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
845
  "model.layers.24.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
846
  "model.layers.24.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
847
+ "model.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
848
  "model.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
849
  "model.layers.24.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
850
  "model.layers.24.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
871
  "model.layers.25.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
872
  "model.layers.25.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
873
  "model.layers.25.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
874
+ "model.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
875
  "model.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
876
  "model.layers.25.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
877
  "model.layers.25.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
884
  "model.layers.25.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
885
  "model.layers.25.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
886
  "model.layers.25.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
887
+ "model.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
888
  "model.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
889
  "model.layers.25.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
890
  "model.layers.25.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
891
  "model.layers.25.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
892
  "model.layers.25.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
893
  "model.layers.25.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
894
+ "model.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
895
  "model.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
896
  "model.layers.25.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
897
  "model.layers.25.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
918
  "model.layers.26.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
919
  "model.layers.26.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
920
  "model.layers.26.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
921
+ "model.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
922
  "model.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
923
  "model.layers.26.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
924
  "model.layers.26.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
931
  "model.layers.26.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
932
  "model.layers.26.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
933
  "model.layers.26.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
934
+ "model.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
935
  "model.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
936
  "model.layers.26.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
937
  "model.layers.26.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
938
  "model.layers.26.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
939
  "model.layers.26.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
940
  "model.layers.26.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
941
+ "model.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
942
  "model.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
943
  "model.layers.26.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
944
  "model.layers.26.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
965
  "model.layers.27.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
966
  "model.layers.27.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
967
  "model.layers.27.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
968
+ "model.layers.27.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
969
  "model.layers.27.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
970
  "model.layers.27.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
971
  "model.layers.27.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
978
  "model.layers.27.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
979
  "model.layers.27.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
980
  "model.layers.27.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
981
+ "model.layers.27.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
982
  "model.layers.27.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
983
  "model.layers.27.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
984
  "model.layers.27.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
985
  "model.layers.27.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
986
  "model.layers.27.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
987
  "model.layers.27.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
988
+ "model.layers.27.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
989
  "model.layers.27.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
990
  "model.layers.27.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
991
  "model.layers.27.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
992
  "model.layers.27.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
993
  "model.layers.27.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
994
  "model.layers.27.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
995
  "model.layers.3.input_layernorm.weight": "model-00001-of-00002.safetensors",
996
  "model.layers.3.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
997
  "model.layers.3.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
1012
  "model.layers.3.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1013
  "model.layers.3.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1014
  "model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1015
+ "model.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1016
  "model.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1017
  "model.layers.3.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1018
  "model.layers.3.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1025
  "model.layers.3.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1026
  "model.layers.3.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1027
  "model.layers.3.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1028
+ "model.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1029
  "model.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1030
  "model.layers.3.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1031
  "model.layers.3.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1032
  "model.layers.3.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1033
  "model.layers.3.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1034
  "model.layers.3.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1035
+ "model.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1036
  "model.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1037
  "model.layers.3.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1038
  "model.layers.3.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1039
  "model.layers.3.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1040
  "model.layers.3.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1041
  "model.layers.3.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1042
  "model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
1043
  "model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1044
  "model.layers.4.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
1059
  "model.layers.4.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1060
  "model.layers.4.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1061
  "model.layers.4.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1062
+ "model.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1063
  "model.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1064
  "model.layers.4.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1065
  "model.layers.4.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1072
  "model.layers.4.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1073
  "model.layers.4.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1074
  "model.layers.4.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1075
+ "model.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1076
  "model.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1077
  "model.layers.4.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1078
  "model.layers.4.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1079
  "model.layers.4.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1080
  "model.layers.4.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1081
  "model.layers.4.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1082
+ "model.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1083
  "model.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1084
  "model.layers.4.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1085
  "model.layers.4.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1106
  "model.layers.5.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1107
  "model.layers.5.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1108
  "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1109
+ "model.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1110
  "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1111
  "model.layers.5.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1112
  "model.layers.5.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1119
  "model.layers.5.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1120
  "model.layers.5.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1121
  "model.layers.5.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1122
+ "model.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1123
  "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1124
  "model.layers.5.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1125
  "model.layers.5.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1126
  "model.layers.5.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1127
  "model.layers.5.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1128
  "model.layers.5.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1129
+ "model.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1130
  "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1131
  "model.layers.5.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1132
  "model.layers.5.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1153
  "model.layers.6.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1154
  "model.layers.6.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1155
  "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1156
+ "model.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1157
  "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1158
  "model.layers.6.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1159
  "model.layers.6.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1166
  "model.layers.6.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1167
  "model.layers.6.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1168
  "model.layers.6.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1169
+ "model.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1170
  "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1171
  "model.layers.6.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1172
  "model.layers.6.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1173
  "model.layers.6.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1174
  "model.layers.6.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1175
  "model.layers.6.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1176
+ "model.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1177
  "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1178
  "model.layers.6.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1179
  "model.layers.6.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1200
  "model.layers.7.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1201
  "model.layers.7.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1202
  "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1203
+ "model.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1204
  "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1205
  "model.layers.7.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1206
  "model.layers.7.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1213
  "model.layers.7.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1214
  "model.layers.7.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1215
  "model.layers.7.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1216
+ "model.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1217
  "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1218
  "model.layers.7.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1219
  "model.layers.7.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1220
  "model.layers.7.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1221
  "model.layers.7.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1222
  "model.layers.7.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1223
+ "model.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1224
  "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1225
  "model.layers.7.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1226
  "model.layers.7.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1247
  "model.layers.8.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1248
  "model.layers.8.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1249
  "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1250
+ "model.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1251
  "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1252
  "model.layers.8.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1253
  "model.layers.8.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1260
  "model.layers.8.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1261
  "model.layers.8.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1262
  "model.layers.8.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1263
+ "model.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1264
  "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1265
  "model.layers.8.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1266
  "model.layers.8.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1267
  "model.layers.8.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1268
  "model.layers.8.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1269
  "model.layers.8.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1270
+ "model.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1271
  "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1272
  "model.layers.8.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1273
  "model.layers.8.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1294
  "model.layers.9.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1295
  "model.layers.9.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1296
  "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1297
+ "model.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1298
  "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1299
  "model.layers.9.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1300
  "model.layers.9.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1307
  "model.layers.9.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1308
  "model.layers.9.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1309
  "model.layers.9.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1310
+ "model.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1311
  "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1312
  "model.layers.9.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1313
  "model.layers.9.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1314
  "model.layers.9.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1315
  "model.layers.9.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1316
  "model.layers.9.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1317
+ "model.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1318
  "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1319
  "model.layers.9.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1320
  "model.layers.9.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",