Undi95 commited on
Commit
c141ccd
1 Parent(s): 6dc0874

Upload folder using huggingface_hub

Browse files
.ipynb_checkpoints/README-checkpoint.md ADDED
@@ -0,0 +1,437 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - 152334H/miqu-1-70b-sf
4
+ - sophosympatheia/Midnight-Miqu-70B-v1.0
5
+ - alchemonaut/BoreanGale-70B
6
+ - NeverSleep/MiquMaid-v1-70B
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+
12
+ ---
13
+ # test
14
+
15
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
+
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using the passthrough merge method.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
+ * [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
26
+ * [sophosympatheia/Midnight-Miqu-70B-v1.0](https://huggingface.co/sophosympatheia/Midnight-Miqu-70B-v1.0)
27
+ * [alchemonaut/BoreanGale-70B](https://huggingface.co/alchemonaut/BoreanGale-70B)
28
+ * [NeverSleep/MiquMaid-v1-70B](https://huggingface.co/NeverSleep/MiquMaid-v1-70B)
29
+
30
+ ### Configuration
31
+
32
+ The following YAML configuration was used to produce this model:
33
+
34
+ ```yaml
35
+ merge_method: passthrough
36
+ slices:
37
+ - sources:
38
+ - layer_range: [0, 1]
39
+ model:
40
+ model:
41
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
42
+ - sources:
43
+ - layer_range: [1, 2]
44
+ model:
45
+ model:
46
+ path: NeverSleep/MiquMaid-v1-70B
47
+ - sources:
48
+ - layer_range: [2, 3]
49
+ model:
50
+ model:
51
+ path: 152334H/miqu-1-70b-sf
52
+ - sources:
53
+ - layer_range: [3, 4]
54
+ model:
55
+ model:
56
+ path: 152334H/miqu-1-70b-sf
57
+ - sources:
58
+ - layer_range: [4, 5]
59
+ model:
60
+ model:
61
+ path: NeverSleep/MiquMaid-v1-70B
62
+ - sources:
63
+ - layer_range: [5, 6]
64
+ model:
65
+ model:
66
+ path: 152334H/miqu-1-70b-sf
67
+ - sources:
68
+ - layer_range: [6, 7]
69
+ model:
70
+ model:
71
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
72
+ - sources:
73
+ - layer_range: [7, 8]
74
+ model:
75
+ model:
76
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
77
+ - sources:
78
+ - layer_range: [8, 9]
79
+ model:
80
+ model:
81
+ path: 152334H/miqu-1-70b-sf
82
+ - sources:
83
+ - layer_range: [9, 10]
84
+ model:
85
+ model:
86
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
87
+ - sources:
88
+ - layer_range: [10, 11]
89
+ model:
90
+ model:
91
+ path: alchemonaut/BoreanGale-70B
92
+ - sources:
93
+ - layer_range: [11, 12]
94
+ model:
95
+ model:
96
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
97
+ - sources:
98
+ - layer_range: [12, 13]
99
+ model:
100
+ model:
101
+ path: 152334H/miqu-1-70b-sf
102
+ - sources:
103
+ - layer_range: [13, 14]
104
+ model:
105
+ model:
106
+ path: alchemonaut/BoreanGale-70B
107
+ - sources:
108
+ - layer_range: [14, 15]
109
+ model:
110
+ model:
111
+ path: 152334H/miqu-1-70b-sf
112
+ - sources:
113
+ - layer_range: [15, 16]
114
+ model:
115
+ model:
116
+ path: alchemonaut/BoreanGale-70B
117
+ - sources:
118
+ - layer_range: [16, 17]
119
+ model:
120
+ model:
121
+ path: NeverSleep/MiquMaid-v1-70B
122
+ - sources:
123
+ - layer_range: [17, 18]
124
+ model:
125
+ model:
126
+ path: 152334H/miqu-1-70b-sf
127
+ - sources:
128
+ - layer_range: [18, 19]
129
+ model:
130
+ model:
131
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
132
+ - sources:
133
+ - layer_range: [19, 20]
134
+ model:
135
+ model:
136
+ path: 152334H/miqu-1-70b-sf
137
+ - sources:
138
+ - layer_range: [20, 21]
139
+ model:
140
+ model:
141
+ path: alchemonaut/BoreanGale-70B
142
+ - sources:
143
+ - layer_range: [21, 22]
144
+ model:
145
+ model:
146
+ path: 152334H/miqu-1-70b-sf
147
+ - sources:
148
+ - layer_range: [22, 23]
149
+ model:
150
+ model:
151
+ path: 152334H/miqu-1-70b-sf
152
+ - sources:
153
+ - layer_range: [23, 24]
154
+ model:
155
+ model:
156
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
157
+ - sources:
158
+ - layer_range: [24, 25]
159
+ model:
160
+ model:
161
+ path: alchemonaut/BoreanGale-70B
162
+ - sources:
163
+ - layer_range: [25, 26]
164
+ model:
165
+ model:
166
+ path: NeverSleep/MiquMaid-v1-70B
167
+ - sources:
168
+ - layer_range: [26, 27]
169
+ model:
170
+ model:
171
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
172
+ - sources:
173
+ - layer_range: [27, 28]
174
+ model:
175
+ model:
176
+ path: 152334H/miqu-1-70b-sf
177
+ - sources:
178
+ - layer_range: [28, 29]
179
+ model:
180
+ model:
181
+ path: 152334H/miqu-1-70b-sf
182
+ - sources:
183
+ - layer_range: [29, 30]
184
+ model:
185
+ model:
186
+ path: NeverSleep/MiquMaid-v1-70B
187
+ - sources:
188
+ - layer_range: [30, 31]
189
+ model:
190
+ model:
191
+ path: NeverSleep/MiquMaid-v1-70B
192
+ - sources:
193
+ - layer_range: [31, 32]
194
+ model:
195
+ model:
196
+ path: NeverSleep/MiquMaid-v1-70B
197
+ - sources:
198
+ - layer_range: [32, 33]
199
+ model:
200
+ model:
201
+ path: alchemonaut/BoreanGale-70B
202
+ - sources:
203
+ - layer_range: [33, 34]
204
+ model:
205
+ model:
206
+ path: alchemonaut/BoreanGale-70B
207
+ - sources:
208
+ - layer_range: [34, 35]
209
+ model:
210
+ model:
211
+ path: alchemonaut/BoreanGale-70B
212
+ - sources:
213
+ - layer_range: [35, 36]
214
+ model:
215
+ model:
216
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
217
+ - sources:
218
+ - layer_range: [36, 37]
219
+ model:
220
+ model:
221
+ path: alchemonaut/BoreanGale-70B
222
+ - sources:
223
+ - layer_range: [37, 38]
224
+ model:
225
+ model:
226
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
227
+ - sources:
228
+ - layer_range: [38, 39]
229
+ model:
230
+ model:
231
+ path: 152334H/miqu-1-70b-sf
232
+ - sources:
233
+ - layer_range: [39, 40]
234
+ model:
235
+ model:
236
+ path: 152334H/miqu-1-70b-sf
237
+ - sources:
238
+ - layer_range: [40, 41]
239
+ model:
240
+ model:
241
+ path: NeverSleep/MiquMaid-v1-70B
242
+ - sources:
243
+ - layer_range: [41, 42]
244
+ model:
245
+ model:
246
+ path: 152334H/miqu-1-70b-sf
247
+ - sources:
248
+ - layer_range: [42, 43]
249
+ model:
250
+ model:
251
+ path: alchemonaut/BoreanGale-70B
252
+ - sources:
253
+ - layer_range: [43, 44]
254
+ model:
255
+ model:
256
+ path: NeverSleep/MiquMaid-v1-70B
257
+ - sources:
258
+ - layer_range: [44, 45]
259
+ model:
260
+ model:
261
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
262
+ - sources:
263
+ - layer_range: [45, 46]
264
+ model:
265
+ model:
266
+ path: 152334H/miqu-1-70b-sf
267
+ - sources:
268
+ - layer_range: [46, 47]
269
+ model:
270
+ model:
271
+ path: alchemonaut/BoreanGale-70B
272
+ - sources:
273
+ - layer_range: [47, 48]
274
+ model:
275
+ model:
276
+ path: 152334H/miqu-1-70b-sf
277
+ - sources:
278
+ - layer_range: [48, 49]
279
+ model:
280
+ model:
281
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
282
+ - sources:
283
+ - layer_range: [49, 50]
284
+ model:
285
+ model:
286
+ path: alchemonaut/BoreanGale-70B
287
+ - sources:
288
+ - layer_range: [50, 51]
289
+ model:
290
+ model:
291
+ path: alchemonaut/BoreanGale-70B
292
+ - sources:
293
+ - layer_range: [51, 52]
294
+ model:
295
+ model:
296
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
297
+ - sources:
298
+ - layer_range: [52, 53]
299
+ model:
300
+ model:
301
+ path: 152334H/miqu-1-70b-sf
302
+ - sources:
303
+ - layer_range: [53, 54]
304
+ model:
305
+ model:
306
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
307
+ - sources:
308
+ - layer_range: [54, 55]
309
+ model:
310
+ model:
311
+ path: 152334H/miqu-1-70b-sf
312
+ - sources:
313
+ - layer_range: [55, 56]
314
+ model:
315
+ model:
316
+ path: alchemonaut/BoreanGale-70B
317
+ - sources:
318
+ - layer_range: [56, 57]
319
+ model:
320
+ model:
321
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
322
+ - sources:
323
+ - layer_range: [57, 58]
324
+ model:
325
+ model:
326
+ path: NeverSleep/MiquMaid-v1-70B
327
+ - sources:
328
+ - layer_range: [58, 59]
329
+ model:
330
+ model:
331
+ path: alchemonaut/BoreanGale-70B
332
+ - sources:
333
+ - layer_range: [59, 60]
334
+ model:
335
+ model:
336
+ path: alchemonaut/BoreanGale-70B
337
+ - sources:
338
+ - layer_range: [60, 61]
339
+ model:
340
+ model:
341
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
342
+ - sources:
343
+ - layer_range: [61, 62]
344
+ model:
345
+ model:
346
+ path: alchemonaut/BoreanGale-70B
347
+ - sources:
348
+ - layer_range: [62, 63]
349
+ model:
350
+ model:
351
+ path: NeverSleep/MiquMaid-v1-70B
352
+ - sources:
353
+ - layer_range: [63, 64]
354
+ model:
355
+ model:
356
+ path: 152334H/miqu-1-70b-sf
357
+ - sources:
358
+ - layer_range: [64, 65]
359
+ model:
360
+ model:
361
+ path: 152334H/miqu-1-70b-sf
362
+ - sources:
363
+ - layer_range: [65, 66]
364
+ model:
365
+ model:
366
+ path: NeverSleep/MiquMaid-v1-70B
367
+ - sources:
368
+ - layer_range: [66, 67]
369
+ model:
370
+ model:
371
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
372
+ - sources:
373
+ - layer_range: [67, 68]
374
+ model:
375
+ model:
376
+ path: alchemonaut/BoreanGale-70B
377
+ - sources:
378
+ - layer_range: [68, 69]
379
+ model:
380
+ model:
381
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
382
+ - sources:
383
+ - layer_range: [69, 70]
384
+ model:
385
+ model:
386
+ path: 152334H/miqu-1-70b-sf
387
+ - sources:
388
+ - layer_range: [70, 71]
389
+ model:
390
+ model:
391
+ path: alchemonaut/BoreanGale-70B
392
+ - sources:
393
+ - layer_range: [71, 72]
394
+ model:
395
+ model:
396
+ path: alchemonaut/BoreanGale-70B
397
+ - sources:
398
+ - layer_range: [72, 73]
399
+ model:
400
+ model:
401
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
402
+ - sources:
403
+ - layer_range: [73, 74]
404
+ model:
405
+ model:
406
+ path: NeverSleep/MiquMaid-v1-70B
407
+ - sources:
408
+ - layer_range: [74, 75]
409
+ model:
410
+ model:
411
+ path: NeverSleep/MiquMaid-v1-70B
412
+ - sources:
413
+ - layer_range: [75, 76]
414
+ model:
415
+ model:
416
+ path: alchemonaut/BoreanGale-70B
417
+ - sources:
418
+ - layer_range: [76, 77]
419
+ model:
420
+ model:
421
+ path: 152334H/miqu-1-70b-sf
422
+ - sources:
423
+ - layer_range: [77, 78]
424
+ model:
425
+ model:
426
+ path: alchemonaut/BoreanGale-70B
427
+ - sources:
428
+ - layer_range: [78, 79]
429
+ model:
430
+ model:
431
+ path: alchemonaut/BoreanGale-70B
432
+ - sources:
433
+ - layer_range: [79, 80]
434
+ model:
435
+ model:
436
+ path: 152334H/miqu-1-70b-sf
437
+ ```
.ipynb_checkpoints/config-checkpoint.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "152334H/miqu-1-70b-sf",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 8192,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 28672,
14
+ "max_position_embeddings": 32764,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 64,
17
+ "num_hidden_layers": 80,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 0,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000,
24
+ "tie_word_embeddings": false,
25
+ "torch_dtype": "float16",
26
+ "transformers_version": "4.38.2",
27
+ "use_cache": true,
28
+ "vocab_size": 32000
29
+ }
.ipynb_checkpoints/mergekit_config-checkpoint.yml ADDED
@@ -0,0 +1,402 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ merge_method: passthrough
2
+ slices:
3
+ - sources:
4
+ - layer_range: [0, 1]
5
+ model:
6
+ model:
7
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
8
+ - sources:
9
+ - layer_range: [1, 2]
10
+ model:
11
+ model:
12
+ path: NeverSleep/MiquMaid-v1-70B
13
+ - sources:
14
+ - layer_range: [2, 3]
15
+ model:
16
+ model:
17
+ path: 152334H/miqu-1-70b-sf
18
+ - sources:
19
+ - layer_range: [3, 4]
20
+ model:
21
+ model:
22
+ path: 152334H/miqu-1-70b-sf
23
+ - sources:
24
+ - layer_range: [4, 5]
25
+ model:
26
+ model:
27
+ path: NeverSleep/MiquMaid-v1-70B
28
+ - sources:
29
+ - layer_range: [5, 6]
30
+ model:
31
+ model:
32
+ path: 152334H/miqu-1-70b-sf
33
+ - sources:
34
+ - layer_range: [6, 7]
35
+ model:
36
+ model:
37
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
38
+ - sources:
39
+ - layer_range: [7, 8]
40
+ model:
41
+ model:
42
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
43
+ - sources:
44
+ - layer_range: [8, 9]
45
+ model:
46
+ model:
47
+ path: 152334H/miqu-1-70b-sf
48
+ - sources:
49
+ - layer_range: [9, 10]
50
+ model:
51
+ model:
52
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
53
+ - sources:
54
+ - layer_range: [10, 11]
55
+ model:
56
+ model:
57
+ path: alchemonaut/BoreanGale-70B
58
+ - sources:
59
+ - layer_range: [11, 12]
60
+ model:
61
+ model:
62
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
63
+ - sources:
64
+ - layer_range: [12, 13]
65
+ model:
66
+ model:
67
+ path: 152334H/miqu-1-70b-sf
68
+ - sources:
69
+ - layer_range: [13, 14]
70
+ model:
71
+ model:
72
+ path: alchemonaut/BoreanGale-70B
73
+ - sources:
74
+ - layer_range: [14, 15]
75
+ model:
76
+ model:
77
+ path: 152334H/miqu-1-70b-sf
78
+ - sources:
79
+ - layer_range: [15, 16]
80
+ model:
81
+ model:
82
+ path: alchemonaut/BoreanGale-70B
83
+ - sources:
84
+ - layer_range: [16, 17]
85
+ model:
86
+ model:
87
+ path: NeverSleep/MiquMaid-v1-70B
88
+ - sources:
89
+ - layer_range: [17, 18]
90
+ model:
91
+ model:
92
+ path: 152334H/miqu-1-70b-sf
93
+ - sources:
94
+ - layer_range: [18, 19]
95
+ model:
96
+ model:
97
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
98
+ - sources:
99
+ - layer_range: [19, 20]
100
+ model:
101
+ model:
102
+ path: 152334H/miqu-1-70b-sf
103
+ - sources:
104
+ - layer_range: [20, 21]
105
+ model:
106
+ model:
107
+ path: alchemonaut/BoreanGale-70B
108
+ - sources:
109
+ - layer_range: [21, 22]
110
+ model:
111
+ model:
112
+ path: 152334H/miqu-1-70b-sf
113
+ - sources:
114
+ - layer_range: [22, 23]
115
+ model:
116
+ model:
117
+ path: 152334H/miqu-1-70b-sf
118
+ - sources:
119
+ - layer_range: [23, 24]
120
+ model:
121
+ model:
122
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
123
+ - sources:
124
+ - layer_range: [24, 25]
125
+ model:
126
+ model:
127
+ path: alchemonaut/BoreanGale-70B
128
+ - sources:
129
+ - layer_range: [25, 26]
130
+ model:
131
+ model:
132
+ path: NeverSleep/MiquMaid-v1-70B
133
+ - sources:
134
+ - layer_range: [26, 27]
135
+ model:
136
+ model:
137
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
138
+ - sources:
139
+ - layer_range: [27, 28]
140
+ model:
141
+ model:
142
+ path: 152334H/miqu-1-70b-sf
143
+ - sources:
144
+ - layer_range: [28, 29]
145
+ model:
146
+ model:
147
+ path: 152334H/miqu-1-70b-sf
148
+ - sources:
149
+ - layer_range: [29, 30]
150
+ model:
151
+ model:
152
+ path: NeverSleep/MiquMaid-v1-70B
153
+ - sources:
154
+ - layer_range: [30, 31]
155
+ model:
156
+ model:
157
+ path: NeverSleep/MiquMaid-v1-70B
158
+ - sources:
159
+ - layer_range: [31, 32]
160
+ model:
161
+ model:
162
+ path: NeverSleep/MiquMaid-v1-70B
163
+ - sources:
164
+ - layer_range: [32, 33]
165
+ model:
166
+ model:
167
+ path: alchemonaut/BoreanGale-70B
168
+ - sources:
169
+ - layer_range: [33, 34]
170
+ model:
171
+ model:
172
+ path: alchemonaut/BoreanGale-70B
173
+ - sources:
174
+ - layer_range: [34, 35]
175
+ model:
176
+ model:
177
+ path: alchemonaut/BoreanGale-70B
178
+ - sources:
179
+ - layer_range: [35, 36]
180
+ model:
181
+ model:
182
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
183
+ - sources:
184
+ - layer_range: [36, 37]
185
+ model:
186
+ model:
187
+ path: alchemonaut/BoreanGale-70B
188
+ - sources:
189
+ - layer_range: [37, 38]
190
+ model:
191
+ model:
192
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
193
+ - sources:
194
+ - layer_range: [38, 39]
195
+ model:
196
+ model:
197
+ path: 152334H/miqu-1-70b-sf
198
+ - sources:
199
+ - layer_range: [39, 40]
200
+ model:
201
+ model:
202
+ path: 152334H/miqu-1-70b-sf
203
+ - sources:
204
+ - layer_range: [40, 41]
205
+ model:
206
+ model:
207
+ path: NeverSleep/MiquMaid-v1-70B
208
+ - sources:
209
+ - layer_range: [41, 42]
210
+ model:
211
+ model:
212
+ path: 152334H/miqu-1-70b-sf
213
+ - sources:
214
+ - layer_range: [42, 43]
215
+ model:
216
+ model:
217
+ path: alchemonaut/BoreanGale-70B
218
+ - sources:
219
+ - layer_range: [43, 44]
220
+ model:
221
+ model:
222
+ path: NeverSleep/MiquMaid-v1-70B
223
+ - sources:
224
+ - layer_range: [44, 45]
225
+ model:
226
+ model:
227
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
228
+ - sources:
229
+ - layer_range: [45, 46]
230
+ model:
231
+ model:
232
+ path: 152334H/miqu-1-70b-sf
233
+ - sources:
234
+ - layer_range: [46, 47]
235
+ model:
236
+ model:
237
+ path: alchemonaut/BoreanGale-70B
238
+ - sources:
239
+ - layer_range: [47, 48]
240
+ model:
241
+ model:
242
+ path: 152334H/miqu-1-70b-sf
243
+ - sources:
244
+ - layer_range: [48, 49]
245
+ model:
246
+ model:
247
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
248
+ - sources:
249
+ - layer_range: [49, 50]
250
+ model:
251
+ model:
252
+ path: alchemonaut/BoreanGale-70B
253
+ - sources:
254
+ - layer_range: [50, 51]
255
+ model:
256
+ model:
257
+ path: alchemonaut/BoreanGale-70B
258
+ - sources:
259
+ - layer_range: [51, 52]
260
+ model:
261
+ model:
262
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
263
+ - sources:
264
+ - layer_range: [52, 53]
265
+ model:
266
+ model:
267
+ path: 152334H/miqu-1-70b-sf
268
+ - sources:
269
+ - layer_range: [53, 54]
270
+ model:
271
+ model:
272
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
273
+ - sources:
274
+ - layer_range: [54, 55]
275
+ model:
276
+ model:
277
+ path: 152334H/miqu-1-70b-sf
278
+ - sources:
279
+ - layer_range: [55, 56]
280
+ model:
281
+ model:
282
+ path: alchemonaut/BoreanGale-70B
283
+ - sources:
284
+ - layer_range: [56, 57]
285
+ model:
286
+ model:
287
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
288
+ - sources:
289
+ - layer_range: [57, 58]
290
+ model:
291
+ model:
292
+ path: NeverSleep/MiquMaid-v1-70B
293
+ - sources:
294
+ - layer_range: [58, 59]
295
+ model:
296
+ model:
297
+ path: alchemonaut/BoreanGale-70B
298
+ - sources:
299
+ - layer_range: [59, 60]
300
+ model:
301
+ model:
302
+ path: alchemonaut/BoreanGale-70B
303
+ - sources:
304
+ - layer_range: [60, 61]
305
+ model:
306
+ model:
307
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
308
+ - sources:
309
+ - layer_range: [61, 62]
310
+ model:
311
+ model:
312
+ path: alchemonaut/BoreanGale-70B
313
+ - sources:
314
+ - layer_range: [62, 63]
315
+ model:
316
+ model:
317
+ path: NeverSleep/MiquMaid-v1-70B
318
+ - sources:
319
+ - layer_range: [63, 64]
320
+ model:
321
+ model:
322
+ path: 152334H/miqu-1-70b-sf
323
+ - sources:
324
+ - layer_range: [64, 65]
325
+ model:
326
+ model:
327
+ path: 152334H/miqu-1-70b-sf
328
+ - sources:
329
+ - layer_range: [65, 66]
330
+ model:
331
+ model:
332
+ path: NeverSleep/MiquMaid-v1-70B
333
+ - sources:
334
+ - layer_range: [66, 67]
335
+ model:
336
+ model:
337
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
338
+ - sources:
339
+ - layer_range: [67, 68]
340
+ model:
341
+ model:
342
+ path: alchemonaut/BoreanGale-70B
343
+ - sources:
344
+ - layer_range: [68, 69]
345
+ model:
346
+ model:
347
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
348
+ - sources:
349
+ - layer_range: [69, 70]
350
+ model:
351
+ model:
352
+ path: 152334H/miqu-1-70b-sf
353
+ - sources:
354
+ - layer_range: [70, 71]
355
+ model:
356
+ model:
357
+ path: alchemonaut/BoreanGale-70B
358
+ - sources:
359
+ - layer_range: [71, 72]
360
+ model:
361
+ model:
362
+ path: alchemonaut/BoreanGale-70B
363
+ - sources:
364
+ - layer_range: [72, 73]
365
+ model:
366
+ model:
367
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
368
+ - sources:
369
+ - layer_range: [73, 74]
370
+ model:
371
+ model:
372
+ path: NeverSleep/MiquMaid-v1-70B
373
+ - sources:
374
+ - layer_range: [74, 75]
375
+ model:
376
+ model:
377
+ path: NeverSleep/MiquMaid-v1-70B
378
+ - sources:
379
+ - layer_range: [75, 76]
380
+ model:
381
+ model:
382
+ path: alchemonaut/BoreanGale-70B
383
+ - sources:
384
+ - layer_range: [76, 77]
385
+ model:
386
+ model:
387
+ path: 152334H/miqu-1-70b-sf
388
+ - sources:
389
+ - layer_range: [77, 78]
390
+ model:
391
+ model:
392
+ path: alchemonaut/BoreanGale-70B
393
+ - sources:
394
+ - layer_range: [78, 79]
395
+ model:
396
+ model:
397
+ path: alchemonaut/BoreanGale-70B
398
+ - sources:
399
+ - layer_range: [79, 80]
400
+ model:
401
+ model:
402
+ path: 152334H/miqu-1-70b-sf
README.md ADDED
@@ -0,0 +1,437 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model:
3
+ - 152334H/miqu-1-70b-sf
4
+ - sophosympatheia/Midnight-Miqu-70B-v1.0
5
+ - alchemonaut/BoreanGale-70B
6
+ - NeverSleep/MiquMaid-v1-70B
7
+ library_name: transformers
8
+ tags:
9
+ - mergekit
10
+ - merge
11
+
12
+ ---
13
+ # test
14
+
15
+ This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
16
+
17
+ ## Merge Details
18
+ ### Merge Method
19
+
20
+ This model was merged using the passthrough merge method.
21
+
22
+ ### Models Merged
23
+
24
+ The following models were included in the merge:
25
+ * [152334H/miqu-1-70b-sf](https://huggingface.co/152334H/miqu-1-70b-sf)
26
+ * [sophosympatheia/Midnight-Miqu-70B-v1.0](https://huggingface.co/sophosympatheia/Midnight-Miqu-70B-v1.0)
27
+ * [alchemonaut/BoreanGale-70B](https://huggingface.co/alchemonaut/BoreanGale-70B)
28
+ * [NeverSleep/MiquMaid-v1-70B](https://huggingface.co/NeverSleep/MiquMaid-v1-70B)
29
+
30
+ ### Configuration
31
+
32
+ The following YAML configuration was used to produce this model:
33
+
34
+ ```yaml
35
+ merge_method: passthrough
36
+ slices:
37
+ - sources:
38
+ - layer_range: [0, 1]
39
+ model:
40
+ model:
41
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
42
+ - sources:
43
+ - layer_range: [1, 2]
44
+ model:
45
+ model:
46
+ path: NeverSleep/MiquMaid-v1-70B
47
+ - sources:
48
+ - layer_range: [2, 3]
49
+ model:
50
+ model:
51
+ path: 152334H/miqu-1-70b-sf
52
+ - sources:
53
+ - layer_range: [3, 4]
54
+ model:
55
+ model:
56
+ path: 152334H/miqu-1-70b-sf
57
+ - sources:
58
+ - layer_range: [4, 5]
59
+ model:
60
+ model:
61
+ path: NeverSleep/MiquMaid-v1-70B
62
+ - sources:
63
+ - layer_range: [5, 6]
64
+ model:
65
+ model:
66
+ path: 152334H/miqu-1-70b-sf
67
+ - sources:
68
+ - layer_range: [6, 7]
69
+ model:
70
+ model:
71
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
72
+ - sources:
73
+ - layer_range: [7, 8]
74
+ model:
75
+ model:
76
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
77
+ - sources:
78
+ - layer_range: [8, 9]
79
+ model:
80
+ model:
81
+ path: 152334H/miqu-1-70b-sf
82
+ - sources:
83
+ - layer_range: [9, 10]
84
+ model:
85
+ model:
86
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
87
+ - sources:
88
+ - layer_range: [10, 11]
89
+ model:
90
+ model:
91
+ path: alchemonaut/BoreanGale-70B
92
+ - sources:
93
+ - layer_range: [11, 12]
94
+ model:
95
+ model:
96
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
97
+ - sources:
98
+ - layer_range: [12, 13]
99
+ model:
100
+ model:
101
+ path: 152334H/miqu-1-70b-sf
102
+ - sources:
103
+ - layer_range: [13, 14]
104
+ model:
105
+ model:
106
+ path: alchemonaut/BoreanGale-70B
107
+ - sources:
108
+ - layer_range: [14, 15]
109
+ model:
110
+ model:
111
+ path: 152334H/miqu-1-70b-sf
112
+ - sources:
113
+ - layer_range: [15, 16]
114
+ model:
115
+ model:
116
+ path: alchemonaut/BoreanGale-70B
117
+ - sources:
118
+ - layer_range: [16, 17]
119
+ model:
120
+ model:
121
+ path: NeverSleep/MiquMaid-v1-70B
122
+ - sources:
123
+ - layer_range: [17, 18]
124
+ model:
125
+ model:
126
+ path: 152334H/miqu-1-70b-sf
127
+ - sources:
128
+ - layer_range: [18, 19]
129
+ model:
130
+ model:
131
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
132
+ - sources:
133
+ - layer_range: [19, 20]
134
+ model:
135
+ model:
136
+ path: 152334H/miqu-1-70b-sf
137
+ - sources:
138
+ - layer_range: [20, 21]
139
+ model:
140
+ model:
141
+ path: alchemonaut/BoreanGale-70B
142
+ - sources:
143
+ - layer_range: [21, 22]
144
+ model:
145
+ model:
146
+ path: 152334H/miqu-1-70b-sf
147
+ - sources:
148
+ - layer_range: [22, 23]
149
+ model:
150
+ model:
151
+ path: 152334H/miqu-1-70b-sf
152
+ - sources:
153
+ - layer_range: [23, 24]
154
+ model:
155
+ model:
156
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
157
+ - sources:
158
+ - layer_range: [24, 25]
159
+ model:
160
+ model:
161
+ path: alchemonaut/BoreanGale-70B
162
+ - sources:
163
+ - layer_range: [25, 26]
164
+ model:
165
+ model:
166
+ path: NeverSleep/MiquMaid-v1-70B
167
+ - sources:
168
+ - layer_range: [26, 27]
169
+ model:
170
+ model:
171
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
172
+ - sources:
173
+ - layer_range: [27, 28]
174
+ model:
175
+ model:
176
+ path: 152334H/miqu-1-70b-sf
177
+ - sources:
178
+ - layer_range: [28, 29]
179
+ model:
180
+ model:
181
+ path: 152334H/miqu-1-70b-sf
182
+ - sources:
183
+ - layer_range: [29, 30]
184
+ model:
185
+ model:
186
+ path: NeverSleep/MiquMaid-v1-70B
187
+ - sources:
188
+ - layer_range: [30, 31]
189
+ model:
190
+ model:
191
+ path: NeverSleep/MiquMaid-v1-70B
192
+ - sources:
193
+ - layer_range: [31, 32]
194
+ model:
195
+ model:
196
+ path: NeverSleep/MiquMaid-v1-70B
197
+ - sources:
198
+ - layer_range: [32, 33]
199
+ model:
200
+ model:
201
+ path: alchemonaut/BoreanGale-70B
202
+ - sources:
203
+ - layer_range: [33, 34]
204
+ model:
205
+ model:
206
+ path: alchemonaut/BoreanGale-70B
207
+ - sources:
208
+ - layer_range: [34, 35]
209
+ model:
210
+ model:
211
+ path: alchemonaut/BoreanGale-70B
212
+ - sources:
213
+ - layer_range: [35, 36]
214
+ model:
215
+ model:
216
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
217
+ - sources:
218
+ - layer_range: [36, 37]
219
+ model:
220
+ model:
221
+ path: alchemonaut/BoreanGale-70B
222
+ - sources:
223
+ - layer_range: [37, 38]
224
+ model:
225
+ model:
226
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
227
+ - sources:
228
+ - layer_range: [38, 39]
229
+ model:
230
+ model:
231
+ path: 152334H/miqu-1-70b-sf
232
+ - sources:
233
+ - layer_range: [39, 40]
234
+ model:
235
+ model:
236
+ path: 152334H/miqu-1-70b-sf
237
+ - sources:
238
+ - layer_range: [40, 41]
239
+ model:
240
+ model:
241
+ path: NeverSleep/MiquMaid-v1-70B
242
+ - sources:
243
+ - layer_range: [41, 42]
244
+ model:
245
+ model:
246
+ path: 152334H/miqu-1-70b-sf
247
+ - sources:
248
+ - layer_range: [42, 43]
249
+ model:
250
+ model:
251
+ path: alchemonaut/BoreanGale-70B
252
+ - sources:
253
+ - layer_range: [43, 44]
254
+ model:
255
+ model:
256
+ path: NeverSleep/MiquMaid-v1-70B
257
+ - sources:
258
+ - layer_range: [44, 45]
259
+ model:
260
+ model:
261
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
262
+ - sources:
263
+ - layer_range: [45, 46]
264
+ model:
265
+ model:
266
+ path: 152334H/miqu-1-70b-sf
267
+ - sources:
268
+ - layer_range: [46, 47]
269
+ model:
270
+ model:
271
+ path: alchemonaut/BoreanGale-70B
272
+ - sources:
273
+ - layer_range: [47, 48]
274
+ model:
275
+ model:
276
+ path: 152334H/miqu-1-70b-sf
277
+ - sources:
278
+ - layer_range: [48, 49]
279
+ model:
280
+ model:
281
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
282
+ - sources:
283
+ - layer_range: [49, 50]
284
+ model:
285
+ model:
286
+ path: alchemonaut/BoreanGale-70B
287
+ - sources:
288
+ - layer_range: [50, 51]
289
+ model:
290
+ model:
291
+ path: alchemonaut/BoreanGale-70B
292
+ - sources:
293
+ - layer_range: [51, 52]
294
+ model:
295
+ model:
296
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
297
+ - sources:
298
+ - layer_range: [52, 53]
299
+ model:
300
+ model:
301
+ path: 152334H/miqu-1-70b-sf
302
+ - sources:
303
+ - layer_range: [53, 54]
304
+ model:
305
+ model:
306
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
307
+ - sources:
308
+ - layer_range: [54, 55]
309
+ model:
310
+ model:
311
+ path: 152334H/miqu-1-70b-sf
312
+ - sources:
313
+ - layer_range: [55, 56]
314
+ model:
315
+ model:
316
+ path: alchemonaut/BoreanGale-70B
317
+ - sources:
318
+ - layer_range: [56, 57]
319
+ model:
320
+ model:
321
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
322
+ - sources:
323
+ - layer_range: [57, 58]
324
+ model:
325
+ model:
326
+ path: NeverSleep/MiquMaid-v1-70B
327
+ - sources:
328
+ - layer_range: [58, 59]
329
+ model:
330
+ model:
331
+ path: alchemonaut/BoreanGale-70B
332
+ - sources:
333
+ - layer_range: [59, 60]
334
+ model:
335
+ model:
336
+ path: alchemonaut/BoreanGale-70B
337
+ - sources:
338
+ - layer_range: [60, 61]
339
+ model:
340
+ model:
341
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
342
+ - sources:
343
+ - layer_range: [61, 62]
344
+ model:
345
+ model:
346
+ path: alchemonaut/BoreanGale-70B
347
+ - sources:
348
+ - layer_range: [62, 63]
349
+ model:
350
+ model:
351
+ path: NeverSleep/MiquMaid-v1-70B
352
+ - sources:
353
+ - layer_range: [63, 64]
354
+ model:
355
+ model:
356
+ path: 152334H/miqu-1-70b-sf
357
+ - sources:
358
+ - layer_range: [64, 65]
359
+ model:
360
+ model:
361
+ path: 152334H/miqu-1-70b-sf
362
+ - sources:
363
+ - layer_range: [65, 66]
364
+ model:
365
+ model:
366
+ path: NeverSleep/MiquMaid-v1-70B
367
+ - sources:
368
+ - layer_range: [66, 67]
369
+ model:
370
+ model:
371
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
372
+ - sources:
373
+ - layer_range: [67, 68]
374
+ model:
375
+ model:
376
+ path: alchemonaut/BoreanGale-70B
377
+ - sources:
378
+ - layer_range: [68, 69]
379
+ model:
380
+ model:
381
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
382
+ - sources:
383
+ - layer_range: [69, 70]
384
+ model:
385
+ model:
386
+ path: 152334H/miqu-1-70b-sf
387
+ - sources:
388
+ - layer_range: [70, 71]
389
+ model:
390
+ model:
391
+ path: alchemonaut/BoreanGale-70B
392
+ - sources:
393
+ - layer_range: [71, 72]
394
+ model:
395
+ model:
396
+ path: alchemonaut/BoreanGale-70B
397
+ - sources:
398
+ - layer_range: [72, 73]
399
+ model:
400
+ model:
401
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
402
+ - sources:
403
+ - layer_range: [73, 74]
404
+ model:
405
+ model:
406
+ path: NeverSleep/MiquMaid-v1-70B
407
+ - sources:
408
+ - layer_range: [74, 75]
409
+ model:
410
+ model:
411
+ path: NeverSleep/MiquMaid-v1-70B
412
+ - sources:
413
+ - layer_range: [75, 76]
414
+ model:
415
+ model:
416
+ path: alchemonaut/BoreanGale-70B
417
+ - sources:
418
+ - layer_range: [76, 77]
419
+ model:
420
+ model:
421
+ path: 152334H/miqu-1-70b-sf
422
+ - sources:
423
+ - layer_range: [77, 78]
424
+ model:
425
+ model:
426
+ path: alchemonaut/BoreanGale-70B
427
+ - sources:
428
+ - layer_range: [78, 79]
429
+ model:
430
+ model:
431
+ path: alchemonaut/BoreanGale-70B
432
+ - sources:
433
+ - layer_range: [79, 80]
434
+ model:
435
+ model:
436
+ path: 152334H/miqu-1-70b-sf
437
+ ```
config.json ADDED
@@ -0,0 +1,29 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "152334H/miqu-1-70b-sf",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "bos_token_id": 1,
9
+ "eos_token_id": 2,
10
+ "hidden_act": "silu",
11
+ "hidden_size": 8192,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 28672,
14
+ "max_position_embeddings": 32764,
15
+ "model_type": "llama",
16
+ "num_attention_heads": 64,
17
+ "num_hidden_layers": 80,
18
+ "num_key_value_heads": 8,
19
+ "pad_token_id": 0,
20
+ "pretraining_tp": 1,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_scaling": null,
23
+ "rope_theta": 1000000,
24
+ "tie_word_embeddings": false,
25
+ "torch_dtype": "float16",
26
+ "transformers_version": "4.38.2",
27
+ "use_cache": true,
28
+ "vocab_size": 32000
29
+ }
mergekit_config.yml ADDED
@@ -0,0 +1,402 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ merge_method: passthrough
2
+ slices:
3
+ - sources:
4
+ - layer_range: [0, 1]
5
+ model:
6
+ model:
7
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
8
+ - sources:
9
+ - layer_range: [1, 2]
10
+ model:
11
+ model:
12
+ path: NeverSleep/MiquMaid-v1-70B
13
+ - sources:
14
+ - layer_range: [2, 3]
15
+ model:
16
+ model:
17
+ path: 152334H/miqu-1-70b-sf
18
+ - sources:
19
+ - layer_range: [3, 4]
20
+ model:
21
+ model:
22
+ path: 152334H/miqu-1-70b-sf
23
+ - sources:
24
+ - layer_range: [4, 5]
25
+ model:
26
+ model:
27
+ path: NeverSleep/MiquMaid-v1-70B
28
+ - sources:
29
+ - layer_range: [5, 6]
30
+ model:
31
+ model:
32
+ path: 152334H/miqu-1-70b-sf
33
+ - sources:
34
+ - layer_range: [6, 7]
35
+ model:
36
+ model:
37
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
38
+ - sources:
39
+ - layer_range: [7, 8]
40
+ model:
41
+ model:
42
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
43
+ - sources:
44
+ - layer_range: [8, 9]
45
+ model:
46
+ model:
47
+ path: 152334H/miqu-1-70b-sf
48
+ - sources:
49
+ - layer_range: [9, 10]
50
+ model:
51
+ model:
52
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
53
+ - sources:
54
+ - layer_range: [10, 11]
55
+ model:
56
+ model:
57
+ path: alchemonaut/BoreanGale-70B
58
+ - sources:
59
+ - layer_range: [11, 12]
60
+ model:
61
+ model:
62
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
63
+ - sources:
64
+ - layer_range: [12, 13]
65
+ model:
66
+ model:
67
+ path: 152334H/miqu-1-70b-sf
68
+ - sources:
69
+ - layer_range: [13, 14]
70
+ model:
71
+ model:
72
+ path: alchemonaut/BoreanGale-70B
73
+ - sources:
74
+ - layer_range: [14, 15]
75
+ model:
76
+ model:
77
+ path: 152334H/miqu-1-70b-sf
78
+ - sources:
79
+ - layer_range: [15, 16]
80
+ model:
81
+ model:
82
+ path: alchemonaut/BoreanGale-70B
83
+ - sources:
84
+ - layer_range: [16, 17]
85
+ model:
86
+ model:
87
+ path: NeverSleep/MiquMaid-v1-70B
88
+ - sources:
89
+ - layer_range: [17, 18]
90
+ model:
91
+ model:
92
+ path: 152334H/miqu-1-70b-sf
93
+ - sources:
94
+ - layer_range: [18, 19]
95
+ model:
96
+ model:
97
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
98
+ - sources:
99
+ - layer_range: [19, 20]
100
+ model:
101
+ model:
102
+ path: 152334H/miqu-1-70b-sf
103
+ - sources:
104
+ - layer_range: [20, 21]
105
+ model:
106
+ model:
107
+ path: alchemonaut/BoreanGale-70B
108
+ - sources:
109
+ - layer_range: [21, 22]
110
+ model:
111
+ model:
112
+ path: 152334H/miqu-1-70b-sf
113
+ - sources:
114
+ - layer_range: [22, 23]
115
+ model:
116
+ model:
117
+ path: 152334H/miqu-1-70b-sf
118
+ - sources:
119
+ - layer_range: [23, 24]
120
+ model:
121
+ model:
122
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
123
+ - sources:
124
+ - layer_range: [24, 25]
125
+ model:
126
+ model:
127
+ path: alchemonaut/BoreanGale-70B
128
+ - sources:
129
+ - layer_range: [25, 26]
130
+ model:
131
+ model:
132
+ path: NeverSleep/MiquMaid-v1-70B
133
+ - sources:
134
+ - layer_range: [26, 27]
135
+ model:
136
+ model:
137
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
138
+ - sources:
139
+ - layer_range: [27, 28]
140
+ model:
141
+ model:
142
+ path: 152334H/miqu-1-70b-sf
143
+ - sources:
144
+ - layer_range: [28, 29]
145
+ model:
146
+ model:
147
+ path: 152334H/miqu-1-70b-sf
148
+ - sources:
149
+ - layer_range: [29, 30]
150
+ model:
151
+ model:
152
+ path: NeverSleep/MiquMaid-v1-70B
153
+ - sources:
154
+ - layer_range: [30, 31]
155
+ model:
156
+ model:
157
+ path: NeverSleep/MiquMaid-v1-70B
158
+ - sources:
159
+ - layer_range: [31, 32]
160
+ model:
161
+ model:
162
+ path: NeverSleep/MiquMaid-v1-70B
163
+ - sources:
164
+ - layer_range: [32, 33]
165
+ model:
166
+ model:
167
+ path: alchemonaut/BoreanGale-70B
168
+ - sources:
169
+ - layer_range: [33, 34]
170
+ model:
171
+ model:
172
+ path: alchemonaut/BoreanGale-70B
173
+ - sources:
174
+ - layer_range: [34, 35]
175
+ model:
176
+ model:
177
+ path: alchemonaut/BoreanGale-70B
178
+ - sources:
179
+ - layer_range: [35, 36]
180
+ model:
181
+ model:
182
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
183
+ - sources:
184
+ - layer_range: [36, 37]
185
+ model:
186
+ model:
187
+ path: alchemonaut/BoreanGale-70B
188
+ - sources:
189
+ - layer_range: [37, 38]
190
+ model:
191
+ model:
192
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
193
+ - sources:
194
+ - layer_range: [38, 39]
195
+ model:
196
+ model:
197
+ path: 152334H/miqu-1-70b-sf
198
+ - sources:
199
+ - layer_range: [39, 40]
200
+ model:
201
+ model:
202
+ path: 152334H/miqu-1-70b-sf
203
+ - sources:
204
+ - layer_range: [40, 41]
205
+ model:
206
+ model:
207
+ path: NeverSleep/MiquMaid-v1-70B
208
+ - sources:
209
+ - layer_range: [41, 42]
210
+ model:
211
+ model:
212
+ path: 152334H/miqu-1-70b-sf
213
+ - sources:
214
+ - layer_range: [42, 43]
215
+ model:
216
+ model:
217
+ path: alchemonaut/BoreanGale-70B
218
+ - sources:
219
+ - layer_range: [43, 44]
220
+ model:
221
+ model:
222
+ path: NeverSleep/MiquMaid-v1-70B
223
+ - sources:
224
+ - layer_range: [44, 45]
225
+ model:
226
+ model:
227
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
228
+ - sources:
229
+ - layer_range: [45, 46]
230
+ model:
231
+ model:
232
+ path: 152334H/miqu-1-70b-sf
233
+ - sources:
234
+ - layer_range: [46, 47]
235
+ model:
236
+ model:
237
+ path: alchemonaut/BoreanGale-70B
238
+ - sources:
239
+ - layer_range: [47, 48]
240
+ model:
241
+ model:
242
+ path: 152334H/miqu-1-70b-sf
243
+ - sources:
244
+ - layer_range: [48, 49]
245
+ model:
246
+ model:
247
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
248
+ - sources:
249
+ - layer_range: [49, 50]
250
+ model:
251
+ model:
252
+ path: alchemonaut/BoreanGale-70B
253
+ - sources:
254
+ - layer_range: [50, 51]
255
+ model:
256
+ model:
257
+ path: alchemonaut/BoreanGale-70B
258
+ - sources:
259
+ - layer_range: [51, 52]
260
+ model:
261
+ model:
262
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
263
+ - sources:
264
+ - layer_range: [52, 53]
265
+ model:
266
+ model:
267
+ path: 152334H/miqu-1-70b-sf
268
+ - sources:
269
+ - layer_range: [53, 54]
270
+ model:
271
+ model:
272
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
273
+ - sources:
274
+ - layer_range: [54, 55]
275
+ model:
276
+ model:
277
+ path: 152334H/miqu-1-70b-sf
278
+ - sources:
279
+ - layer_range: [55, 56]
280
+ model:
281
+ model:
282
+ path: alchemonaut/BoreanGale-70B
283
+ - sources:
284
+ - layer_range: [56, 57]
285
+ model:
286
+ model:
287
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
288
+ - sources:
289
+ - layer_range: [57, 58]
290
+ model:
291
+ model:
292
+ path: NeverSleep/MiquMaid-v1-70B
293
+ - sources:
294
+ - layer_range: [58, 59]
295
+ model:
296
+ model:
297
+ path: alchemonaut/BoreanGale-70B
298
+ - sources:
299
+ - layer_range: [59, 60]
300
+ model:
301
+ model:
302
+ path: alchemonaut/BoreanGale-70B
303
+ - sources:
304
+ - layer_range: [60, 61]
305
+ model:
306
+ model:
307
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
308
+ - sources:
309
+ - layer_range: [61, 62]
310
+ model:
311
+ model:
312
+ path: alchemonaut/BoreanGale-70B
313
+ - sources:
314
+ - layer_range: [62, 63]
315
+ model:
316
+ model:
317
+ path: NeverSleep/MiquMaid-v1-70B
318
+ - sources:
319
+ - layer_range: [63, 64]
320
+ model:
321
+ model:
322
+ path: 152334H/miqu-1-70b-sf
323
+ - sources:
324
+ - layer_range: [64, 65]
325
+ model:
326
+ model:
327
+ path: 152334H/miqu-1-70b-sf
328
+ - sources:
329
+ - layer_range: [65, 66]
330
+ model:
331
+ model:
332
+ path: NeverSleep/MiquMaid-v1-70B
333
+ - sources:
334
+ - layer_range: [66, 67]
335
+ model:
336
+ model:
337
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
338
+ - sources:
339
+ - layer_range: [67, 68]
340
+ model:
341
+ model:
342
+ path: alchemonaut/BoreanGale-70B
343
+ - sources:
344
+ - layer_range: [68, 69]
345
+ model:
346
+ model:
347
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
348
+ - sources:
349
+ - layer_range: [69, 70]
350
+ model:
351
+ model:
352
+ path: 152334H/miqu-1-70b-sf
353
+ - sources:
354
+ - layer_range: [70, 71]
355
+ model:
356
+ model:
357
+ path: alchemonaut/BoreanGale-70B
358
+ - sources:
359
+ - layer_range: [71, 72]
360
+ model:
361
+ model:
362
+ path: alchemonaut/BoreanGale-70B
363
+ - sources:
364
+ - layer_range: [72, 73]
365
+ model:
366
+ model:
367
+ path: sophosympatheia/Midnight-Miqu-70B-v1.0
368
+ - sources:
369
+ - layer_range: [73, 74]
370
+ model:
371
+ model:
372
+ path: NeverSleep/MiquMaid-v1-70B
373
+ - sources:
374
+ - layer_range: [74, 75]
375
+ model:
376
+ model:
377
+ path: NeverSleep/MiquMaid-v1-70B
378
+ - sources:
379
+ - layer_range: [75, 76]
380
+ model:
381
+ model:
382
+ path: alchemonaut/BoreanGale-70B
383
+ - sources:
384
+ - layer_range: [76, 77]
385
+ model:
386
+ model:
387
+ path: 152334H/miqu-1-70b-sf
388
+ - sources:
389
+ - layer_range: [77, 78]
390
+ model:
391
+ model:
392
+ path: alchemonaut/BoreanGale-70B
393
+ - sources:
394
+ - layer_range: [78, 79]
395
+ model:
396
+ model:
397
+ path: alchemonaut/BoreanGale-70B
398
+ - sources:
399
+ - layer_range: [79, 80]
400
+ model:
401
+ model:
402
+ path: 152334H/miqu-1-70b-sf
model-00001-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:01c4398f86b5554634d9abbaa60eec874348d18dba1168eb2145bce8d3da28c5
3
+ size 10511079200
model-00002-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0be96f41fc887c75cbfa95891f8c65f82c32271cf792ea87925df6d68310cc6
3
+ size 9663895528
model-00003-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:899e9d0c04b8deb8236787b35376e9d1a89b98f8064d4ef775fb7702dcf66262
3
+ size 9965852208
model-00004-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1da59ee17e746cfee6e2b2a43849ac40805d2f96eda6e281aa9ea827c11c50c
3
+ size 9965819928
model-00005-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:579dc97bbdba456b1a89fadc512bc33783de8fc2e70c9ba0f5dcfe820d8c3f62
3
+ size 9932347744
model-00006-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:724f815e24c2460625aa7a879320367380c050f59fa64235813034b4186daa47
3
+ size 9999406888
model-00007-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0f884cb3d998d3666fb6004d9df57def111fb54230fad78f5d90f3bb1b45f3cb
3
+ size 9932298248
model-00008-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:146d0b367ccc3454db0c417ab1bdb4a5eeae64e75d19030082ac9362a248e17d
3
+ size 9630291408
model-00009-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f295984db417c5451fbef586b40254b83bfeee5ced20072acb429647d589921
3
+ size 9965868736
model-00010-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bce877c9d157d29cd613bc8c049e2bcb6af5c72966badd32f0a1043457dfd740
3
+ size 9965885928
model-00011-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46a53abe1e3c1807febbee1474e12590ecbd4cabfe4488f3484b0a79ea1d038d
3
+ size 9798079712
model-00012-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c1b13f92ff4dc54de2a8a90b77c2dba247647f853734b525b258fad0e8304a1
3
+ size 9798129912
model-00013-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7201f8107e67ae0972292fcc23066012e07b8ca7dde62c0e4ed4c9a099e67415
3
+ size 9663846096
model-00014-of-00014.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f7fc7505df08e5a433ed9ddc7fb08cbc5831ca268aaac40f74a83b3ddca6b557
3
+ size 9684866824
model.safetensors.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metadata": {"mergekit_version": "0.0.4.1"}, "weight_map": {"model.layers.0.mlp.down_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00014.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00014.safetensors", "model.embed_tokens.weight": "model-00001-of-00014.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00001-of-00014.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00001-of-00014.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00014.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00014.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00014.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00014.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00001-of-00014.safetensors", "model.layers.10.input_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00001-of-00014.safetensors", "model.layers.9.input_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00001-of-00014.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00014.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00001-of-00014.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00014.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00014.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00002-of-00014.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00002-of-00014.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00002-of-00014.safetensors", "model.layers.7.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00014.safetensors", "model.layers.6.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00014.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00014.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00014.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00014.safetensors", "model.layers.4.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00014.safetensors", "model.layers.3.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00014.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00014.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00002-of-00014.safetensors", "model.layers.15.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00002-of-00014.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00002-of-00014.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00002-of-00014.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.13.input_layernorm.weight": "model-00002-of-00014.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00003-of-00014.safetensors", "model.layers.11.input_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00014.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00014.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00014.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00003-of-00014.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00003-of-00014.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.5.input_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00003-of-00014.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00003-of-00014.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00003-of-00014.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00014.safetensors", "model.layers.20.input_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00003-of-00014.safetensors", "model.layers.18.input_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00003-of-00014.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00003-of-00014.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00014.safetensors", "model.layers.8.input_layernorm.weight": "model-00003-of-00014.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00003-of-00014.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00003-of-00014.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00004-of-00014.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00004-of-00014.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.26.input_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00004-of-00014.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00004-of-00014.safetensors", "model.layers.24.input_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.23.input_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.12.input_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.36.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.36.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.36.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.35.mlp.down_proj.weight": "model-00004-of-00014.safetensors", "model.layers.35.mlp.gate_proj.weight": "model-00004-of-00014.safetensors", "model.layers.33.self_attn.o_proj.weight": "model-00004-of-00014.safetensors", "model.layers.33.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.33.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.33.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.mlp.down_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.mlp.gate_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.mlp.up_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.post_attention_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.32.self_attn.o_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.32.input_layernorm.weight": "model-00004-of-00014.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00004-of-00014.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00004-of-00014.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00004-of-00014.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00004-of-00014.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00004-of-00014.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00004-of-00014.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00014.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00005-of-00014.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00005-of-00014.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00005-of-00014.safetensors", "model.layers.14.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.37.mlp.down_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.mlp.gate_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.37.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.self_attn.v_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.self_attn.k_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.self_attn.q_proj.weight": "model-00005-of-00014.safetensors", "model.layers.37.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.36.mlp.down_proj.weight": "model-00005-of-00014.safetensors", "model.layers.36.mlp.gate_proj.weight": "model-00005-of-00014.safetensors", "model.layers.36.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.36.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.36.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.36.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.35.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.35.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.35.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.35.self_attn.v_proj.weight": "model-00005-of-00014.safetensors", "model.layers.35.self_attn.k_proj.weight": "model-00005-of-00014.safetensors", "model.layers.35.self_attn.q_proj.weight": "model-00005-of-00014.safetensors", "model.layers.35.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.34.mlp.down_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.mlp.gate_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.34.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.self_attn.v_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.self_attn.k_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.self_attn.q_proj.weight": "model-00005-of-00014.safetensors", "model.layers.34.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.33.mlp.down_proj.weight": "model-00005-of-00014.safetensors", "model.layers.33.mlp.gate_proj.weight": "model-00005-of-00014.safetensors", "model.layers.33.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.33.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.33.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00014.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00014.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00014.safetensors", "model.layers.17.input_layernorm.weight": "model-00005-of-00014.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00006-of-00014.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00006-of-00014.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.16.input_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.46.mlp.down_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.mlp.gate_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.mlp.up_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.post_attention_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.46.self_attn.o_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.self_attn.v_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.self_attn.k_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.self_attn.q_proj.weight": "model-00006-of-00014.safetensors", "model.layers.46.input_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.44.mlp.gate_proj.weight": "model-00006-of-00014.safetensors", "model.layers.44.self_attn.o_proj.weight": "model-00006-of-00014.safetensors", "model.layers.44.self_attn.v_proj.weight": "model-00006-of-00014.safetensors", "model.layers.44.self_attn.k_proj.weight": "model-00006-of-00014.safetensors", "model.layers.44.self_attn.q_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.mlp.down_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.mlp.gate_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.mlp.up_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.post_attention_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.42.self_attn.o_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.self_attn.v_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.self_attn.k_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.self_attn.q_proj.weight": "model-00006-of-00014.safetensors", "model.layers.42.input_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00014.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00014.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00014.safetensors", "model.layers.21.input_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00006-of-00014.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00006-of-00014.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00006-of-00014.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.19.input_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.50.self_attn.v_proj.weight": "model-00006-of-00014.safetensors", "model.layers.50.self_attn.k_proj.weight": "model-00006-of-00014.safetensors", "model.layers.50.self_attn.q_proj.weight": "model-00006-of-00014.safetensors", "model.layers.49.mlp.down_proj.weight": "model-00006-of-00014.safetensors", "model.layers.49.mlp.gate_proj.weight": "model-00006-of-00014.safetensors", "model.layers.49.mlp.up_proj.weight": "model-00006-of-00014.safetensors", "model.layers.49.post_attention_layernorm.weight": "model-00006-of-00014.safetensors", "model.layers.49.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.49.self_attn.v_proj.weight": "model-00007-of-00014.safetensors", "model.layers.49.self_attn.k_proj.weight": "model-00007-of-00014.safetensors", "model.layers.49.self_attn.q_proj.weight": "model-00007-of-00014.safetensors", "model.layers.49.input_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.48.mlp.down_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.mlp.gate_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.mlp.up_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.post_attention_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.48.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.self_attn.v_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.self_attn.k_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.self_attn.q_proj.weight": "model-00007-of-00014.safetensors", "model.layers.48.input_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.44.mlp.down_proj.weight": "model-00007-of-00014.safetensors", "model.layers.44.mlp.up_proj.weight": "model-00007-of-00014.safetensors", "model.layers.44.post_attention_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.44.input_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00007-of-00014.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00007-of-00014.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00007-of-00014.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.22.input_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.58.mlp.gate_proj.weight": "model-00007-of-00014.safetensors", "model.layers.58.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.58.self_attn.v_proj.weight": "model-00007-of-00014.safetensors", "model.layers.58.self_attn.k_proj.weight": "model-00007-of-00014.safetensors", "model.layers.58.self_attn.q_proj.weight": "model-00007-of-00014.safetensors", "model.layers.55.mlp.gate_proj.weight": "model-00007-of-00014.safetensors", "model.layers.55.mlp.up_proj.weight": "model-00007-of-00014.safetensors", "model.layers.55.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.55.self_attn.v_proj.weight": "model-00007-of-00014.safetensors", "model.layers.55.self_attn.k_proj.weight": "model-00007-of-00014.safetensors", "model.layers.55.self_attn.q_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.mlp.down_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.mlp.gate_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.mlp.up_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.post_attention_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.53.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.self_attn.v_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.self_attn.k_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.self_attn.q_proj.weight": "model-00007-of-00014.safetensors", "model.layers.53.input_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.51.mlp.down_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.mlp.gate_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.mlp.up_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.post_attention_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.51.self_attn.o_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.self_attn.v_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.self_attn.k_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.self_attn.q_proj.weight": "model-00007-of-00014.safetensors", "model.layers.51.input_layernorm.weight": "model-00007-of-00014.safetensors", "model.layers.50.mlp.down_proj.weight": "model-00008-of-00014.safetensors", "model.layers.50.mlp.gate_proj.weight": "model-00008-of-00014.safetensors", "model.layers.50.mlp.up_proj.weight": "model-00008-of-00014.safetensors", "model.layers.50.post_attention_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.50.self_attn.o_proj.weight": "model-00008-of-00014.safetensors", "model.layers.50.input_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00008-of-00014.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00008-of-00014.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00008-of-00014.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00008-of-00014.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00008-of-00014.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00008-of-00014.safetensors", "model.layers.25.input_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.61.self_attn.o_proj.weight": "model-00008-of-00014.safetensors", "model.layers.61.self_attn.v_proj.weight": "model-00008-of-00014.safetensors", "model.layers.61.self_attn.k_proj.weight": "model-00008-of-00014.safetensors", "model.layers.61.self_attn.q_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.mlp.down_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.mlp.gate_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.mlp.up_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.post_attention_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.60.self_attn.o_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.self_attn.v_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.self_attn.k_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.self_attn.q_proj.weight": "model-00008-of-00014.safetensors", "model.layers.60.input_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.59.mlp.down_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.mlp.gate_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.mlp.up_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.post_attention_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.59.self_attn.o_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.self_attn.v_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.self_attn.k_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.self_attn.q_proj.weight": "model-00008-of-00014.safetensors", "model.layers.59.input_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.58.mlp.down_proj.weight": "model-00008-of-00014.safetensors", "model.layers.58.mlp.up_proj.weight": "model-00008-of-00014.safetensors", "model.layers.58.post_attention_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.58.input_layernorm.weight": "model-00008-of-00014.safetensors", "model.layers.56.mlp.down_proj.weight": "model-00008-of-00014.safetensors", "model.layers.56.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.56.mlp.up_proj.weight": "model-00009-of-00014.safetensors", "model.layers.56.post_attention_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.56.self_attn.o_proj.weight": "model-00009-of-00014.safetensors", "model.layers.56.self_attn.v_proj.weight": "model-00009-of-00014.safetensors", "model.layers.56.self_attn.k_proj.weight": "model-00009-of-00014.safetensors", "model.layers.56.self_attn.q_proj.weight": "model-00009-of-00014.safetensors", "model.layers.56.input_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.55.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.55.post_attention_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.55.input_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00009-of-00014.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00009-of-00014.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00009-of-00014.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00009-of-00014.safetensors", "model.layers.29.input_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00009-of-00014.safetensors", "model.layers.28.input_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.27.input_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.68.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.68.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.mlp.up_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.post_attention_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.66.self_attn.o_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.self_attn.v_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.self_attn.k_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.self_attn.q_proj.weight": "model-00009-of-00014.safetensors", "model.layers.66.input_layernorm.weight": "model-00009-of-00014.safetensors", "model.layers.61.mlp.down_proj.weight": "model-00009-of-00014.safetensors", "model.layers.61.mlp.gate_proj.weight": "model-00009-of-00014.safetensors", "model.layers.61.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.61.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.61.input_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00010-of-00014.safetensors", "model.layers.31.input_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00010-of-00014.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.30.input_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.75.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.75.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.75.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.75.self_attn.q_proj.weight": "model-00010-of-00014.safetensors", "model.layers.72.mlp.gate_proj.weight": "model-00010-of-00014.safetensors", "model.layers.72.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.72.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.72.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.72.self_attn.q_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.mlp.down_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.mlp.gate_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.71.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.self_attn.q_proj.weight": "model-00010-of-00014.safetensors", "model.layers.71.input_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.70.mlp.down_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.mlp.gate_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.70.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.self_attn.q_proj.weight": "model-00010-of-00014.safetensors", "model.layers.70.input_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.68.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.68.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.68.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.68.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.68.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.68.self_attn.q_proj.weight": "model-00010-of-00014.safetensors", "model.layers.68.input_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.67.mlp.down_proj.weight": "model-00010-of-00014.safetensors", "model.layers.67.mlp.gate_proj.weight": "model-00010-of-00014.safetensors", "model.layers.67.mlp.up_proj.weight": "model-00010-of-00014.safetensors", "model.layers.67.post_attention_layernorm.weight": "model-00010-of-00014.safetensors", "model.layers.67.self_attn.o_proj.weight": "model-00010-of-00014.safetensors", "model.layers.67.self_attn.v_proj.weight": "model-00010-of-00014.safetensors", "model.layers.67.self_attn.k_proj.weight": "model-00010-of-00014.safetensors", "model.layers.67.self_attn.q_proj.weight": "model-00011-of-00014.safetensors", "model.layers.67.input_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.78.mlp.gate_proj.weight": "model-00011-of-00014.safetensors", "model.layers.78.post_attention_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.78.self_attn.o_proj.weight": "model-00011-of-00014.safetensors", "model.layers.78.self_attn.v_proj.weight": "model-00011-of-00014.safetensors", "model.layers.78.self_attn.k_proj.weight": "model-00011-of-00014.safetensors", "model.layers.78.self_attn.q_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.mlp.down_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.mlp.gate_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.mlp.up_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.post_attention_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.77.self_attn.o_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.self_attn.v_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.self_attn.k_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.self_attn.q_proj.weight": "model-00011-of-00014.safetensors", "model.layers.77.input_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.75.mlp.down_proj.weight": "model-00011-of-00014.safetensors", "model.layers.75.mlp.gate_proj.weight": "model-00011-of-00014.safetensors", "model.layers.75.mlp.up_proj.weight": "model-00011-of-00014.safetensors", "model.layers.75.post_attention_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.75.input_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.72.mlp.down_proj.weight": "model-00011-of-00014.safetensors", "model.layers.72.mlp.up_proj.weight": "model-00011-of-00014.safetensors", "model.layers.72.post_attention_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.72.input_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.38.mlp.down_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.mlp.gate_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.mlp.up_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.post_attention_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.38.self_attn.o_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.self_attn.v_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.self_attn.k_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.self_attn.q_proj.weight": "model-00011-of-00014.safetensors", "model.layers.38.input_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.78.mlp.down_proj.weight": "model-00011-of-00014.safetensors", "model.layers.78.mlp.up_proj.weight": "model-00011-of-00014.safetensors", "model.layers.78.input_layernorm.weight": "model-00011-of-00014.safetensors", "model.layers.41.mlp.gate_proj.weight": "model-00011-of-00014.safetensors", "model.layers.41.mlp.up_proj.weight": "model-00011-of-00014.safetensors", "model.layers.41.self_attn.o_proj.weight": "model-00011-of-00014.safetensors", "model.layers.41.self_attn.v_proj.weight": "model-00011-of-00014.safetensors", "model.layers.41.self_attn.k_proj.weight": "model-00011-of-00014.safetensors", "model.layers.41.self_attn.q_proj.weight": "model-00011-of-00014.safetensors", "model.layers.40.mlp.down_proj.weight": "model-00011-of-00014.safetensors", "model.layers.40.mlp.gate_proj.weight": "model-00011-of-00014.safetensors", "model.layers.40.mlp.up_proj.weight": "model-00012-of-00014.safetensors", "model.layers.40.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.40.self_attn.o_proj.weight": "model-00012-of-00014.safetensors", "model.layers.40.self_attn.v_proj.weight": "model-00012-of-00014.safetensors", "model.layers.40.self_attn.k_proj.weight": "model-00012-of-00014.safetensors", "model.layers.40.self_attn.q_proj.weight": "model-00012-of-00014.safetensors", "model.layers.40.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.39.mlp.down_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.mlp.gate_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.mlp.up_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.39.self_attn.o_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.self_attn.v_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.self_attn.k_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.self_attn.q_proj.weight": "model-00012-of-00014.safetensors", "model.layers.39.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.43.mlp.down_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.mlp.gate_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.mlp.up_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.43.self_attn.o_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.self_attn.v_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.self_attn.k_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.self_attn.q_proj.weight": "model-00012-of-00014.safetensors", "model.layers.43.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.41.mlp.down_proj.weight": "model-00012-of-00014.safetensors", "model.layers.41.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.41.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.47.self_attn.o_proj.weight": "model-00012-of-00014.safetensors", "model.layers.47.self_attn.v_proj.weight": "model-00012-of-00014.safetensors", "model.layers.47.self_attn.k_proj.weight": "model-00012-of-00014.safetensors", "model.layers.47.self_attn.q_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.mlp.down_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.mlp.gate_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.mlp.up_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.45.self_attn.o_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.self_attn.v_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.self_attn.k_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.self_attn.q_proj.weight": "model-00012-of-00014.safetensors", "model.layers.45.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.47.mlp.down_proj.weight": "model-00012-of-00014.safetensors", "model.layers.47.mlp.gate_proj.weight": "model-00012-of-00014.safetensors", "model.layers.47.mlp.up_proj.weight": "model-00012-of-00014.safetensors", "model.layers.47.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.47.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.52.mlp.down_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.mlp.gate_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.mlp.up_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.post_attention_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.52.self_attn.o_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.self_attn.v_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.self_attn.k_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.self_attn.q_proj.weight": "model-00012-of-00014.safetensors", "model.layers.52.input_layernorm.weight": "model-00012-of-00014.safetensors", "model.layers.54.mlp.down_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.mlp.gate_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.mlp.up_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.post_attention_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.54.self_attn.o_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.self_attn.v_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.self_attn.k_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.self_attn.q_proj.weight": "model-00013-of-00014.safetensors", "model.layers.54.input_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.57.mlp.down_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.mlp.gate_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.mlp.up_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.post_attention_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.57.self_attn.o_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.self_attn.v_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.self_attn.k_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.self_attn.q_proj.weight": "model-00013-of-00014.safetensors", "model.layers.57.input_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.64.self_attn.v_proj.weight": "model-00013-of-00014.safetensors", "model.layers.64.self_attn.k_proj.weight": "model-00013-of-00014.safetensors", "model.layers.64.self_attn.q_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.mlp.down_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.mlp.gate_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.mlp.up_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.post_attention_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.63.self_attn.o_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.self_attn.v_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.self_attn.k_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.self_attn.q_proj.weight": "model-00013-of-00014.safetensors", "model.layers.63.input_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.62.mlp.down_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.mlp.gate_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.mlp.up_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.post_attention_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.62.self_attn.o_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.self_attn.v_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.self_attn.k_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.self_attn.q_proj.weight": "model-00013-of-00014.safetensors", "model.layers.62.input_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.65.mlp.down_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.mlp.gate_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.mlp.up_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.post_attention_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.65.self_attn.o_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.self_attn.v_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.self_attn.k_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.self_attn.q_proj.weight": "model-00013-of-00014.safetensors", "model.layers.65.input_layernorm.weight": "model-00013-of-00014.safetensors", "model.layers.64.mlp.down_proj.weight": "model-00013-of-00014.safetensors", "model.layers.64.mlp.gate_proj.weight": "model-00013-of-00014.safetensors", "model.layers.64.mlp.up_proj.weight": "model-00014-of-00014.safetensors", "model.layers.64.post_attention_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.64.self_attn.o_proj.weight": "model-00014-of-00014.safetensors", "model.layers.64.input_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.69.mlp.gate_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.mlp.up_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.self_attn.o_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.self_attn.v_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.self_attn.k_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.self_attn.q_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.mlp.down_proj.weight": "model-00014-of-00014.safetensors", "model.layers.69.post_attention_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.69.input_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.74.mlp.down_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.mlp.gate_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.mlp.up_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.post_attention_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.74.self_attn.o_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.self_attn.v_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.self_attn.k_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.self_attn.q_proj.weight": "model-00014-of-00014.safetensors", "model.layers.74.input_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.73.mlp.down_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.mlp.gate_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.mlp.up_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.post_attention_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.73.self_attn.o_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.self_attn.v_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.self_attn.k_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.self_attn.q_proj.weight": "model-00014-of-00014.safetensors", "model.layers.73.input_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.76.mlp.down_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.mlp.gate_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.mlp.up_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.post_attention_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.76.self_attn.o_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.self_attn.v_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.self_attn.k_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.self_attn.q_proj.weight": "model-00014-of-00014.safetensors", "model.layers.76.input_layernorm.weight": "model-00014-of-00014.safetensors", "lm_head.weight": "model-00014-of-00014.safetensors", "model.norm.weight": "model-00014-of-00014.safetensors", "model.layers.79.mlp.down_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.mlp.gate_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.mlp.up_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.post_attention_layernorm.weight": "model-00014-of-00014.safetensors", "model.layers.79.self_attn.o_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.self_attn.v_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.self_attn.k_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.self_attn.q_proj.weight": "model-00014-of-00014.safetensors", "model.layers.79.input_layernorm.weight": "model-00014-of-00014.safetensors"}}
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": true,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": true,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": true,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": true,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "bos_token": "<s>",
32
+ "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}",
33
+ "clean_up_tokenization_spaces": false,
34
+ "eos_token": "</s>",
35
+ "legacy": false,
36
+ "model_max_length": 1000000000000000019884624838656,
37
+ "pad_token": "<unk>",
38
+ "sp_model_kwargs": {},
39
+ "spaces_between_special_tokens": false,
40
+ "tokenizer_class": "LlamaTokenizer",
41
+ "unk_token": "<unk>",
42
+ "use_default_system_prompt": false
43
+ }