allknowingroger commited on
Commit
d0a71f0
1 Parent(s): 4407170

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,63 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - merge
4
+ - mergekit
5
+ - lazymergekit
6
+ - allknowingroger/Multimerge-12B-MoE
7
+ - yunconglong/Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B
8
+ base_model:
9
+ - allknowingroger/Multimerge-12B-MoE
10
+ - yunconglong/Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B
11
+ ---
12
+
13
+ # MultiMash2-12B-slerp
14
+
15
+ MultiMash2-12B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
+ * [allknowingroger/Multimerge-12B-MoE](https://huggingface.co/allknowingroger/Multimerge-12B-MoE)
17
+ * [yunconglong/Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B](https://huggingface.co/yunconglong/Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B)
18
+
19
+ ## 🧩 Configuration
20
+
21
+ ```yaml
22
+ slices:
23
+ - sources:
24
+ - model: allknowingroger/Multimerge-12B-MoE
25
+ layer_range: [0, 32]
26
+ - model: yunconglong/Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B
27
+ layer_range: [0, 32]
28
+ merge_method: slerp
29
+ base_model: allknowingroger/Multimerge-12B-MoE
30
+ parameters:
31
+ t:
32
+ - filter: self_attn
33
+ value: [0, 0.5, 0.3, 0.7, 1]
34
+ - filter: mlp
35
+ value: [1, 0.5, 0.7, 0.3, 0]
36
+ - value: 0.5
37
+ dtype: bfloat16
38
+ ```
39
+
40
+ ## 💻 Usage
41
+
42
+ ```python
43
+ !pip install -qU transformers accelerate
44
+
45
+ from transformers import AutoTokenizer
46
+ import transformers
47
+ import torch
48
+
49
+ model = "allknowingroger/MultiMash2-12B-slerp"
50
+ messages = [{"role": "user", "content": "What is a large language model?"}]
51
+
52
+ tokenizer = AutoTokenizer.from_pretrained(model)
53
+ prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
54
+ pipeline = transformers.pipeline(
55
+ "text-generation",
56
+ model=model,
57
+ torch_dtype=torch.float16,
58
+ device_map="auto",
59
+ )
60
+
61
+ outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
62
+ print(outputs[0]["generated_text"])
63
+ ```
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "allknowingroger/Multimerge-12B-MoE",
3
+ "architectures": [
4
+ "MixtralForCausalLM"
5
+ ],
6
+ "attention_dropout": 0.0,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 4096,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 14336,
13
+ "max_position_embeddings": 32768,
14
+ "model_type": "mixtral",
15
+ "num_attention_heads": 32,
16
+ "num_experts_per_tok": 2,
17
+ "num_hidden_layers": 32,
18
+ "num_key_value_heads": 8,
19
+ "num_local_experts": 2,
20
+ "output_router_logits": false,
21
+ "rms_norm_eps": 1e-05,
22
+ "rope_theta": 10000.0,
23
+ "router_aux_loss_coef": 0.001,
24
+ "router_jitter_noise": 0.0,
25
+ "sliding_window": null,
26
+ "tie_word_embeddings": false,
27
+ "torch_dtype": "bfloat16",
28
+ "transformers_version": "4.40.2",
29
+ "use_cache": true,
30
+ "vocab_size": 32000
31
+ }
mergekit_config.yml ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ slices:
3
+ - sources:
4
+ - model: allknowingroger/Multimerge-12B-MoE
5
+ layer_range: [0, 32]
6
+ - model: yunconglong/Truthful_DPO_TomGrc_FusionNet_7Bx2_MoE_13B
7
+ layer_range: [0, 32]
8
+ merge_method: slerp
9
+ base_model: allknowingroger/Multimerge-12B-MoE
10
+ parameters:
11
+ t:
12
+ - filter: self_attn
13
+ value: [0, 0.5, 0.3, 0.7, 1]
14
+ - filter: mlp
15
+ value: [1, 0.5, 0.7, 0.3, 0]
16
+ - value: 0.5
17
+ dtype: bfloat16
model-00001-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bc7698e5c17c1ebc67500187297a392169dcfd5cabe2c20339783bac0e8d2f2a
3
+ size 994050792
model-00002-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:23fd7b68939349a0ae5ede2fc789171d239e5d4e934db9c70d2c90b251f6ba12
3
+ size 906004176
model-00003-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e5df8ac22f47a694c03f82a64ec13fff302f26d96fdac96e55aa2b514c51c72
3
+ size 989923840
model-00004-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fa34e4b5852b311de10572083db404cb11b0c4567d67989162a19572bf137336
3
+ size 906004192
model-00005-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2e38825e539d770dff4dc893a9ce7fdc8b2bccb376180d9c9c8bf491e6566d7b
3
+ size 906004192
model-00006-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f29f8942d5aeb36b969143731a95bfb3222dd02ed69a636761d47191091d1f6b
3
+ size 906004192
model-00007-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36de2f0102bf1d5e0b6d2e83a6cb12a1836470e7700b72cfabe70bea92d672b9
3
+ size 906004192
model-00008-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:607baeed54c148c9e9a83313492ef2a39cd4264d2d8012a1afb15eab466e1a15
3
+ size 906004192
model-00009-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6fbedfc03fac4a4daa13bc7ccc5c9c33dc204c63f8fb97c69727caf02b7932c5
3
+ size 989923848
model-00010-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:afd9f6f43ba4d592982386dde5272fa62e787843ba9d92db27e8af97f690cfd8
3
+ size 906004192
model-00011-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d533bb9917a4313ae862a839c85e88b16fb606e941ada5b276ad8c5c2892d130
3
+ size 906004184
model-00012-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:365cb246b890742c1ec3aa5141a4d47123a4d3eca638cd023b2b0cac13386555
3
+ size 906004176
model-00013-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ddcce3f5d33686cfff4025419b97a047c72a5956af1f63423e70531a2b5a543
3
+ size 906004192
model-00014-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:954ff1c67cbe769a6f97f907ce1039eb6ff895d624bb3663766a206728689318
3
+ size 906004192
model-00015-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:16889ed14be4ad87933cfd8d69092d3ba4d8f5befdcab0d88d0fdec2f7c82252
3
+ size 989923848
model-00016-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:669cbd5db9dae8d2553f348db6f10d6b06c4a215bc8f8bb8fd35cdbaa377bf2b
3
+ size 906004192
model-00017-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0090817bf5062c353dc01cb9e8424d3fc1e77de31ef9dc870e7f1998e7afe43d
3
+ size 906004192
model-00018-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1cc1867eb864a63193d4ba2d64e0d4204919b9e2adcbac8a35595fc327b4fd8d
3
+ size 906004192
model-00019-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9a596d82d453b0f0b83efaed40a77b28fa2c65c326da9cd3da2fa63c3cd71da3
3
+ size 906004192
model-00020-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2c415cea37eb6bfa60cfd22b7aac791caaa596369ca1810bd4044f02be501b9a
3
+ size 906004192
model-00021-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:51ec6b4292d4d19432050dbdceb5112ed51d8d290a2d82aa6b72f2c302f99db7
3
+ size 989923832
model-00022-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:636336bc4477f1d8a849f5e49ea28c0e38eb39441da39fb113a752624fe4b8b7
3
+ size 906004192
model-00023-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:de18707dc7014b67288258403a6c0ffe7e0eb5a4c31f2c3d01f46cf64acc5cda
3
+ size 906004184
model-00024-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:180e4ce6f3fe4984ce52b380556852d0ff6e3ea9cf0428a76977eea1e3c58767
3
+ size 906004176
model-00025-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a1608ce8faf03d114ce10bec6299d0352e54f3544b0d645798679c01b73bedbc
3
+ size 906004176
model-00026-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b5ceb90bcc9ed186b4fdaf6887fcd5c150b976cff7e642d7f78617e34276941
3
+ size 906004176
model-00027-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:def968b63c754b49755aacd6b826d3eeba7d09f0413a7fe9120726fed737345c
3
+ size 989923824
model-00028-of-00028.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c9468e22b38480b740b0b950b19fb0bc46dcdb305622644973a579f3d0d7423d
3
+ size 788571816
model.safetensors.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metadata": {"mergekit_version": "0.0.4.2", "total_size": 25758277632}, "weight_map": {"lm_head.weight": "model-00001-of-00028.safetensors", "model.embed_tokens.weight": "model-00001-of-00028.safetensors", "model.layers.0.block_sparse_moe.experts.0.w1.weight": "model-00001-of-00028.safetensors", "model.layers.0.block_sparse_moe.experts.0.w2.weight": "model-00001-of-00028.safetensors", "model.layers.0.block_sparse_moe.experts.0.w3.weight": "model-00001-of-00028.safetensors", "model.layers.0.block_sparse_moe.experts.1.w1.weight": "model-00001-of-00028.safetensors", "model.layers.0.block_sparse_moe.experts.1.w2.weight": "model-00002-of-00028.safetensors", "model.layers.0.block_sparse_moe.experts.1.w3.weight": "model-00002-of-00028.safetensors", "model.layers.0.block_sparse_moe.gate.weight": "model-00002-of-00028.safetensors", "model.layers.0.input_layernorm.weight": "model-00002-of-00028.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00028.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00028.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00028.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00028.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00028.safetensors", "model.layers.1.block_sparse_moe.experts.0.w1.weight": "model-00002-of-00028.safetensors", "model.layers.1.block_sparse_moe.experts.0.w2.weight": "model-00002-of-00028.safetensors", "model.layers.1.block_sparse_moe.experts.0.w3.weight": "model-00002-of-00028.safetensors", "model.layers.1.block_sparse_moe.experts.1.w1.weight": "model-00002-of-00028.safetensors", "model.layers.1.block_sparse_moe.experts.1.w2.weight": "model-00002-of-00028.safetensors", "model.layers.1.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00028.safetensors", "model.layers.1.block_sparse_moe.gate.weight": "model-00003-of-00028.safetensors", "model.layers.1.input_layernorm.weight": "model-00003-of-00028.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00003-of-00028.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00003-of-00028.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00003-of-00028.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00003-of-00028.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.experts.0.w1.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.experts.0.w2.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.experts.0.w3.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.experts.1.w1.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.experts.1.w2.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.experts.1.w3.weight": "model-00003-of-00028.safetensors", "model.layers.10.block_sparse_moe.gate.weight": "model-00003-of-00028.safetensors", "model.layers.10.input_layernorm.weight": "model-00003-of-00028.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00028.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00028.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00028.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00028.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00003-of-00028.safetensors", "model.layers.11.block_sparse_moe.experts.0.w1.weight": "model-00004-of-00028.safetensors", "model.layers.11.block_sparse_moe.experts.0.w2.weight": "model-00004-of-00028.safetensors", "model.layers.11.block_sparse_moe.experts.0.w3.weight": "model-00004-of-00028.safetensors", "model.layers.11.block_sparse_moe.experts.1.w1.weight": "model-00004-of-00028.safetensors", "model.layers.11.block_sparse_moe.experts.1.w2.weight": "model-00004-of-00028.safetensors", "model.layers.11.block_sparse_moe.experts.1.w3.weight": "model-00004-of-00028.safetensors", "model.layers.11.block_sparse_moe.gate.weight": "model-00004-of-00028.safetensors", "model.layers.11.input_layernorm.weight": "model-00004-of-00028.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00004-of-00028.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00028.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00028.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00028.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00028.safetensors", "model.layers.12.block_sparse_moe.experts.0.w1.weight": "model-00004-of-00028.safetensors", "model.layers.12.block_sparse_moe.experts.0.w2.weight": "model-00005-of-00028.safetensors", "model.layers.12.block_sparse_moe.experts.0.w3.weight": "model-00005-of-00028.safetensors", "model.layers.12.block_sparse_moe.experts.1.w1.weight": "model-00005-of-00028.safetensors", "model.layers.12.block_sparse_moe.experts.1.w2.weight": "model-00005-of-00028.safetensors", "model.layers.12.block_sparse_moe.experts.1.w3.weight": "model-00005-of-00028.safetensors", "model.layers.12.block_sparse_moe.gate.weight": "model-00005-of-00028.safetensors", "model.layers.12.input_layernorm.weight": "model-00005-of-00028.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00028.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00028.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00028.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00028.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00028.safetensors", "model.layers.13.block_sparse_moe.experts.0.w1.weight": "model-00005-of-00028.safetensors", "model.layers.13.block_sparse_moe.experts.0.w2.weight": "model-00005-of-00028.safetensors", "model.layers.13.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00028.safetensors", "model.layers.13.block_sparse_moe.experts.1.w1.weight": "model-00006-of-00028.safetensors", "model.layers.13.block_sparse_moe.experts.1.w2.weight": "model-00006-of-00028.safetensors", "model.layers.13.block_sparse_moe.experts.1.w3.weight": "model-00006-of-00028.safetensors", "model.layers.13.block_sparse_moe.gate.weight": "model-00006-of-00028.safetensors", "model.layers.13.input_layernorm.weight": "model-00006-of-00028.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00006-of-00028.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00006-of-00028.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00006-of-00028.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00006-of-00028.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00006-of-00028.safetensors", "model.layers.14.block_sparse_moe.experts.0.w1.weight": "model-00006-of-00028.safetensors", "model.layers.14.block_sparse_moe.experts.0.w2.weight": "model-00006-of-00028.safetensors", "model.layers.14.block_sparse_moe.experts.0.w3.weight": "model-00006-of-00028.safetensors", "model.layers.14.block_sparse_moe.experts.1.w1.weight": "model-00007-of-00028.safetensors", "model.layers.14.block_sparse_moe.experts.1.w2.weight": "model-00007-of-00028.safetensors", "model.layers.14.block_sparse_moe.experts.1.w3.weight": "model-00007-of-00028.safetensors", "model.layers.14.block_sparse_moe.gate.weight": "model-00007-of-00028.safetensors", "model.layers.14.input_layernorm.weight": "model-00007-of-00028.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00007-of-00028.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00007-of-00028.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00007-of-00028.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00007-of-00028.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00007-of-00028.safetensors", "model.layers.15.block_sparse_moe.experts.0.w1.weight": "model-00007-of-00028.safetensors", "model.layers.15.block_sparse_moe.experts.0.w2.weight": "model-00007-of-00028.safetensors", "model.layers.15.block_sparse_moe.experts.0.w3.weight": "model-00007-of-00028.safetensors", "model.layers.15.block_sparse_moe.experts.1.w1.weight": "model-00007-of-00028.safetensors", "model.layers.15.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00028.safetensors", "model.layers.15.block_sparse_moe.experts.1.w3.weight": "model-00008-of-00028.safetensors", "model.layers.15.block_sparse_moe.gate.weight": "model-00008-of-00028.safetensors", "model.layers.15.input_layernorm.weight": "model-00008-of-00028.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00008-of-00028.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00008-of-00028.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00008-of-00028.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00008-of-00028.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00008-of-00028.safetensors", "model.layers.16.block_sparse_moe.experts.0.w1.weight": "model-00008-of-00028.safetensors", "model.layers.16.block_sparse_moe.experts.0.w2.weight": "model-00008-of-00028.safetensors", "model.layers.16.block_sparse_moe.experts.0.w3.weight": "model-00008-of-00028.safetensors", "model.layers.16.block_sparse_moe.experts.1.w1.weight": "model-00008-of-00028.safetensors", "model.layers.16.block_sparse_moe.experts.1.w2.weight": "model-00008-of-00028.safetensors", "model.layers.16.block_sparse_moe.experts.1.w3.weight": "model-00009-of-00028.safetensors", "model.layers.16.block_sparse_moe.gate.weight": "model-00009-of-00028.safetensors", "model.layers.16.input_layernorm.weight": "model-00009-of-00028.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00009-of-00028.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00009-of-00028.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00009-of-00028.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00009-of-00028.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.experts.0.w1.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.experts.0.w2.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.experts.0.w3.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.experts.1.w1.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.experts.1.w2.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.experts.1.w3.weight": "model-00009-of-00028.safetensors", "model.layers.17.block_sparse_moe.gate.weight": "model-00009-of-00028.safetensors", "model.layers.17.input_layernorm.weight": "model-00009-of-00028.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00009-of-00028.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00009-of-00028.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00009-of-00028.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00009-of-00028.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00009-of-00028.safetensors", "model.layers.18.block_sparse_moe.experts.0.w1.weight": "model-00010-of-00028.safetensors", "model.layers.18.block_sparse_moe.experts.0.w2.weight": "model-00010-of-00028.safetensors", "model.layers.18.block_sparse_moe.experts.0.w3.weight": "model-00010-of-00028.safetensors", "model.layers.18.block_sparse_moe.experts.1.w1.weight": "model-00010-of-00028.safetensors", "model.layers.18.block_sparse_moe.experts.1.w2.weight": "model-00010-of-00028.safetensors", "model.layers.18.block_sparse_moe.experts.1.w3.weight": "model-00010-of-00028.safetensors", "model.layers.18.block_sparse_moe.gate.weight": "model-00010-of-00028.safetensors", "model.layers.18.input_layernorm.weight": "model-00010-of-00028.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00010-of-00028.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00010-of-00028.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00010-of-00028.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00010-of-00028.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00010-of-00028.safetensors", "model.layers.19.block_sparse_moe.experts.0.w1.weight": "model-00010-of-00028.safetensors", "model.layers.19.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00028.safetensors", "model.layers.19.block_sparse_moe.experts.0.w3.weight": "model-00011-of-00028.safetensors", "model.layers.19.block_sparse_moe.experts.1.w1.weight": "model-00011-of-00028.safetensors", "model.layers.19.block_sparse_moe.experts.1.w2.weight": "model-00011-of-00028.safetensors", "model.layers.19.block_sparse_moe.experts.1.w3.weight": "model-00011-of-00028.safetensors", "model.layers.19.block_sparse_moe.gate.weight": "model-00011-of-00028.safetensors", "model.layers.19.input_layernorm.weight": "model-00011-of-00028.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00011-of-00028.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00011-of-00028.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00011-of-00028.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00011-of-00028.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00011-of-00028.safetensors", "model.layers.2.block_sparse_moe.experts.0.w1.weight": "model-00011-of-00028.safetensors", "model.layers.2.block_sparse_moe.experts.0.w2.weight": "model-00011-of-00028.safetensors", "model.layers.2.block_sparse_moe.experts.0.w3.weight": "model-00012-of-00028.safetensors", "model.layers.2.block_sparse_moe.experts.1.w1.weight": "model-00012-of-00028.safetensors", "model.layers.2.block_sparse_moe.experts.1.w2.weight": "model-00012-of-00028.safetensors", "model.layers.2.block_sparse_moe.experts.1.w3.weight": "model-00012-of-00028.safetensors", "model.layers.2.block_sparse_moe.gate.weight": "model-00012-of-00028.safetensors", "model.layers.2.input_layernorm.weight": "model-00012-of-00028.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00012-of-00028.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00012-of-00028.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00012-of-00028.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00012-of-00028.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00012-of-00028.safetensors", "model.layers.20.block_sparse_moe.experts.0.w1.weight": "model-00012-of-00028.safetensors", "model.layers.20.block_sparse_moe.experts.0.w2.weight": "model-00012-of-00028.safetensors", "model.layers.20.block_sparse_moe.experts.0.w3.weight": "model-00012-of-00028.safetensors", "model.layers.20.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00028.safetensors", "model.layers.20.block_sparse_moe.experts.1.w2.weight": "model-00013-of-00028.safetensors", "model.layers.20.block_sparse_moe.experts.1.w3.weight": "model-00013-of-00028.safetensors", "model.layers.20.block_sparse_moe.gate.weight": "model-00013-of-00028.safetensors", "model.layers.20.input_layernorm.weight": "model-00013-of-00028.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00013-of-00028.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00013-of-00028.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00013-of-00028.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00013-of-00028.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00013-of-00028.safetensors", "model.layers.21.block_sparse_moe.experts.0.w1.weight": "model-00013-of-00028.safetensors", "model.layers.21.block_sparse_moe.experts.0.w2.weight": "model-00013-of-00028.safetensors", "model.layers.21.block_sparse_moe.experts.0.w3.weight": "model-00013-of-00028.safetensors", "model.layers.21.block_sparse_moe.experts.1.w1.weight": "model-00013-of-00028.safetensors", "model.layers.21.block_sparse_moe.experts.1.w2.weight": "model-00014-of-00028.safetensors", "model.layers.21.block_sparse_moe.experts.1.w3.weight": "model-00014-of-00028.safetensors", "model.layers.21.block_sparse_moe.gate.weight": "model-00014-of-00028.safetensors", "model.layers.21.input_layernorm.weight": "model-00014-of-00028.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00014-of-00028.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00014-of-00028.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00014-of-00028.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00014-of-00028.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00014-of-00028.safetensors", "model.layers.22.block_sparse_moe.experts.0.w1.weight": "model-00014-of-00028.safetensors", "model.layers.22.block_sparse_moe.experts.0.w2.weight": "model-00014-of-00028.safetensors", "model.layers.22.block_sparse_moe.experts.0.w3.weight": "model-00014-of-00028.safetensors", "model.layers.22.block_sparse_moe.experts.1.w1.weight": "model-00014-of-00028.safetensors", "model.layers.22.block_sparse_moe.experts.1.w2.weight": "model-00014-of-00028.safetensors", "model.layers.22.block_sparse_moe.experts.1.w3.weight": "model-00015-of-00028.safetensors", "model.layers.22.block_sparse_moe.gate.weight": "model-00015-of-00028.safetensors", "model.layers.22.input_layernorm.weight": "model-00015-of-00028.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00015-of-00028.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00015-of-00028.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00015-of-00028.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00015-of-00028.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.experts.0.w1.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.experts.0.w2.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.experts.0.w3.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.experts.1.w1.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.experts.1.w2.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.experts.1.w3.weight": "model-00015-of-00028.safetensors", "model.layers.23.block_sparse_moe.gate.weight": "model-00015-of-00028.safetensors", "model.layers.23.input_layernorm.weight": "model-00015-of-00028.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00015-of-00028.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00015-of-00028.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00015-of-00028.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00015-of-00028.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00015-of-00028.safetensors", "model.layers.24.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00028.safetensors", "model.layers.24.block_sparse_moe.experts.0.w2.weight": "model-00016-of-00028.safetensors", "model.layers.24.block_sparse_moe.experts.0.w3.weight": "model-00016-of-00028.safetensors", "model.layers.24.block_sparse_moe.experts.1.w1.weight": "model-00016-of-00028.safetensors", "model.layers.24.block_sparse_moe.experts.1.w2.weight": "model-00016-of-00028.safetensors", "model.layers.24.block_sparse_moe.experts.1.w3.weight": "model-00016-of-00028.safetensors", "model.layers.24.block_sparse_moe.gate.weight": "model-00016-of-00028.safetensors", "model.layers.24.input_layernorm.weight": "model-00016-of-00028.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00016-of-00028.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00016-of-00028.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00016-of-00028.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00016-of-00028.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00016-of-00028.safetensors", "model.layers.25.block_sparse_moe.experts.0.w1.weight": "model-00016-of-00028.safetensors", "model.layers.25.block_sparse_moe.experts.0.w2.weight": "model-00017-of-00028.safetensors", "model.layers.25.block_sparse_moe.experts.0.w3.weight": "model-00017-of-00028.safetensors", "model.layers.25.block_sparse_moe.experts.1.w1.weight": "model-00017-of-00028.safetensors", "model.layers.25.block_sparse_moe.experts.1.w2.weight": "model-00017-of-00028.safetensors", "model.layers.25.block_sparse_moe.experts.1.w3.weight": "model-00017-of-00028.safetensors", "model.layers.25.block_sparse_moe.gate.weight": "model-00017-of-00028.safetensors", "model.layers.25.input_layernorm.weight": "model-00017-of-00028.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00017-of-00028.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00017-of-00028.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00017-of-00028.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00017-of-00028.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00017-of-00028.safetensors", "model.layers.26.block_sparse_moe.experts.0.w1.weight": "model-00017-of-00028.safetensors", "model.layers.26.block_sparse_moe.experts.0.w2.weight": "model-00017-of-00028.safetensors", "model.layers.26.block_sparse_moe.experts.0.w3.weight": "model-00018-of-00028.safetensors", "model.layers.26.block_sparse_moe.experts.1.w1.weight": "model-00018-of-00028.safetensors", "model.layers.26.block_sparse_moe.experts.1.w2.weight": "model-00018-of-00028.safetensors", "model.layers.26.block_sparse_moe.experts.1.w3.weight": "model-00018-of-00028.safetensors", "model.layers.26.block_sparse_moe.gate.weight": "model-00018-of-00028.safetensors", "model.layers.26.input_layernorm.weight": "model-00018-of-00028.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00018-of-00028.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00018-of-00028.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00018-of-00028.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00018-of-00028.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00018-of-00028.safetensors", "model.layers.27.block_sparse_moe.experts.0.w1.weight": "model-00018-of-00028.safetensors", "model.layers.27.block_sparse_moe.experts.0.w2.weight": "model-00018-of-00028.safetensors", "model.layers.27.block_sparse_moe.experts.0.w3.weight": "model-00018-of-00028.safetensors", "model.layers.27.block_sparse_moe.experts.1.w1.weight": "model-00019-of-00028.safetensors", "model.layers.27.block_sparse_moe.experts.1.w2.weight": "model-00019-of-00028.safetensors", "model.layers.27.block_sparse_moe.experts.1.w3.weight": "model-00019-of-00028.safetensors", "model.layers.27.block_sparse_moe.gate.weight": "model-00019-of-00028.safetensors", "model.layers.27.input_layernorm.weight": "model-00019-of-00028.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00019-of-00028.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00019-of-00028.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00019-of-00028.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00019-of-00028.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00019-of-00028.safetensors", "model.layers.28.block_sparse_moe.experts.0.w1.weight": "model-00019-of-00028.safetensors", "model.layers.28.block_sparse_moe.experts.0.w2.weight": "model-00019-of-00028.safetensors", "model.layers.28.block_sparse_moe.experts.0.w3.weight": "model-00019-of-00028.safetensors", "model.layers.28.block_sparse_moe.experts.1.w1.weight": "model-00019-of-00028.safetensors", "model.layers.28.block_sparse_moe.experts.1.w2.weight": "model-00020-of-00028.safetensors", "model.layers.28.block_sparse_moe.experts.1.w3.weight": "model-00020-of-00028.safetensors", "model.layers.28.block_sparse_moe.gate.weight": "model-00020-of-00028.safetensors", "model.layers.28.input_layernorm.weight": "model-00020-of-00028.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00020-of-00028.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00020-of-00028.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00020-of-00028.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00020-of-00028.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00020-of-00028.safetensors", "model.layers.29.block_sparse_moe.experts.0.w1.weight": "model-00020-of-00028.safetensors", "model.layers.29.block_sparse_moe.experts.0.w2.weight": "model-00020-of-00028.safetensors", "model.layers.29.block_sparse_moe.experts.0.w3.weight": "model-00020-of-00028.safetensors", "model.layers.29.block_sparse_moe.experts.1.w1.weight": "model-00020-of-00028.safetensors", "model.layers.29.block_sparse_moe.experts.1.w2.weight": "model-00020-of-00028.safetensors", "model.layers.29.block_sparse_moe.experts.1.w3.weight": "model-00021-of-00028.safetensors", "model.layers.29.block_sparse_moe.gate.weight": "model-00021-of-00028.safetensors", "model.layers.29.input_layernorm.weight": "model-00021-of-00028.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00021-of-00028.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00021-of-00028.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00021-of-00028.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00021-of-00028.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.experts.0.w1.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.experts.0.w2.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.experts.0.w3.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.experts.1.w1.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.experts.1.w2.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.experts.1.w3.weight": "model-00021-of-00028.safetensors", "model.layers.3.block_sparse_moe.gate.weight": "model-00021-of-00028.safetensors", "model.layers.3.input_layernorm.weight": "model-00021-of-00028.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00021-of-00028.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00021-of-00028.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00021-of-00028.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00021-of-00028.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00021-of-00028.safetensors", "model.layers.30.block_sparse_moe.experts.0.w1.weight": "model-00022-of-00028.safetensors", "model.layers.30.block_sparse_moe.experts.0.w2.weight": "model-00022-of-00028.safetensors", "model.layers.30.block_sparse_moe.experts.0.w3.weight": "model-00022-of-00028.safetensors", "model.layers.30.block_sparse_moe.experts.1.w1.weight": "model-00022-of-00028.safetensors", "model.layers.30.block_sparse_moe.experts.1.w2.weight": "model-00022-of-00028.safetensors", "model.layers.30.block_sparse_moe.experts.1.w3.weight": "model-00022-of-00028.safetensors", "model.layers.30.block_sparse_moe.gate.weight": "model-00022-of-00028.safetensors", "model.layers.30.input_layernorm.weight": "model-00022-of-00028.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00022-of-00028.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00022-of-00028.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00022-of-00028.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00022-of-00028.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00022-of-00028.safetensors", "model.layers.31.block_sparse_moe.experts.0.w1.weight": "model-00022-of-00028.safetensors", "model.layers.31.block_sparse_moe.experts.0.w2.weight": "model-00023-of-00028.safetensors", "model.layers.31.block_sparse_moe.experts.0.w3.weight": "model-00023-of-00028.safetensors", "model.layers.31.block_sparse_moe.experts.1.w1.weight": "model-00023-of-00028.safetensors", "model.layers.31.block_sparse_moe.experts.1.w2.weight": "model-00023-of-00028.safetensors", "model.layers.31.block_sparse_moe.experts.1.w3.weight": "model-00023-of-00028.safetensors", "model.layers.31.block_sparse_moe.gate.weight": "model-00023-of-00028.safetensors", "model.layers.31.input_layernorm.weight": "model-00023-of-00028.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00023-of-00028.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00023-of-00028.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00023-of-00028.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00023-of-00028.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00023-of-00028.safetensors", "model.layers.4.block_sparse_moe.experts.0.w1.weight": "model-00023-of-00028.safetensors", "model.layers.4.block_sparse_moe.experts.0.w2.weight": "model-00023-of-00028.safetensors", "model.layers.4.block_sparse_moe.experts.0.w3.weight": "model-00024-of-00028.safetensors", "model.layers.4.block_sparse_moe.experts.1.w1.weight": "model-00024-of-00028.safetensors", "model.layers.4.block_sparse_moe.experts.1.w2.weight": "model-00024-of-00028.safetensors", "model.layers.4.block_sparse_moe.experts.1.w3.weight": "model-00024-of-00028.safetensors", "model.layers.4.block_sparse_moe.gate.weight": "model-00024-of-00028.safetensors", "model.layers.4.input_layernorm.weight": "model-00024-of-00028.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00024-of-00028.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00024-of-00028.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00024-of-00028.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00024-of-00028.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00024-of-00028.safetensors", "model.layers.5.block_sparse_moe.experts.0.w1.weight": "model-00024-of-00028.safetensors", "model.layers.5.block_sparse_moe.experts.0.w2.weight": "model-00024-of-00028.safetensors", "model.layers.5.block_sparse_moe.experts.0.w3.weight": "model-00024-of-00028.safetensors", "model.layers.5.block_sparse_moe.experts.1.w1.weight": "model-00025-of-00028.safetensors", "model.layers.5.block_sparse_moe.experts.1.w2.weight": "model-00025-of-00028.safetensors", "model.layers.5.block_sparse_moe.experts.1.w3.weight": "model-00025-of-00028.safetensors", "model.layers.5.block_sparse_moe.gate.weight": "model-00025-of-00028.safetensors", "model.layers.5.input_layernorm.weight": "model-00025-of-00028.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00025-of-00028.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00025-of-00028.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00025-of-00028.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00025-of-00028.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00025-of-00028.safetensors", "model.layers.6.block_sparse_moe.experts.0.w1.weight": "model-00025-of-00028.safetensors", "model.layers.6.block_sparse_moe.experts.0.w2.weight": "model-00025-of-00028.safetensors", "model.layers.6.block_sparse_moe.experts.0.w3.weight": "model-00025-of-00028.safetensors", "model.layers.6.block_sparse_moe.experts.1.w1.weight": "model-00025-of-00028.safetensors", "model.layers.6.block_sparse_moe.experts.1.w2.weight": "model-00026-of-00028.safetensors", "model.layers.6.block_sparse_moe.experts.1.w3.weight": "model-00026-of-00028.safetensors", "model.layers.6.block_sparse_moe.gate.weight": "model-00026-of-00028.safetensors", "model.layers.6.input_layernorm.weight": "model-00026-of-00028.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00026-of-00028.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00026-of-00028.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00026-of-00028.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00026-of-00028.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00026-of-00028.safetensors", "model.layers.7.block_sparse_moe.experts.0.w1.weight": "model-00026-of-00028.safetensors", "model.layers.7.block_sparse_moe.experts.0.w2.weight": "model-00026-of-00028.safetensors", "model.layers.7.block_sparse_moe.experts.0.w3.weight": "model-00026-of-00028.safetensors", "model.layers.7.block_sparse_moe.experts.1.w1.weight": "model-00026-of-00028.safetensors", "model.layers.7.block_sparse_moe.experts.1.w2.weight": "model-00026-of-00028.safetensors", "model.layers.7.block_sparse_moe.experts.1.w3.weight": "model-00027-of-00028.safetensors", "model.layers.7.block_sparse_moe.gate.weight": "model-00027-of-00028.safetensors", "model.layers.7.input_layernorm.weight": "model-00027-of-00028.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00027-of-00028.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00027-of-00028.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00027-of-00028.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00027-of-00028.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.experts.0.w1.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.experts.0.w2.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.experts.0.w3.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.experts.1.w1.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.experts.1.w2.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.experts.1.w3.weight": "model-00027-of-00028.safetensors", "model.layers.8.block_sparse_moe.gate.weight": "model-00027-of-00028.safetensors", "model.layers.8.input_layernorm.weight": "model-00027-of-00028.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00027-of-00028.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00027-of-00028.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00027-of-00028.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00027-of-00028.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00027-of-00028.safetensors", "model.layers.9.block_sparse_moe.experts.0.w1.weight": "model-00028-of-00028.safetensors", "model.layers.9.block_sparse_moe.experts.0.w2.weight": "model-00028-of-00028.safetensors", "model.layers.9.block_sparse_moe.experts.0.w3.weight": "model-00028-of-00028.safetensors", "model.layers.9.block_sparse_moe.experts.1.w1.weight": "model-00028-of-00028.safetensors", "model.layers.9.block_sparse_moe.experts.1.w2.weight": "model-00028-of-00028.safetensors", "model.layers.9.block_sparse_moe.experts.1.w3.weight": "model-00028-of-00028.safetensors", "model.layers.9.block_sparse_moe.gate.weight": "model-00028-of-00028.safetensors", "model.layers.9.input_layernorm.weight": "model-00028-of-00028.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00028-of-00028.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00028-of-00028.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00028-of-00028.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00028-of-00028.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00028-of-00028.safetensors", "model.norm.weight": "model-00028-of-00028.safetensors"}}
special_tokens_map.json ADDED
@@ -0,0 +1,30 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "<s>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "unk_token": {
24
+ "content": "<unk>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ }
30
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ }
29
+ },
30
+ "additional_special_tokens": [],
31
+ "bos_token": "<s>",
32
+ "clean_up_tokenization_spaces": false,
33
+ "eos_token": "</s>",
34
+ "legacy": true,
35
+ "max_length": null,
36
+ "model_max_length": 32768,
37
+ "pad_to_multiple_of": null,
38
+ "pad_token": "<s>",
39
+ "pad_token_type_id": 0,
40
+ "padding_side": "left",
41
+ "sp_model_kwargs": {},
42
+ "spaces_between_special_tokens": false,
43
+ "tokenizer_class": "LlamaTokenizer",
44
+ "unk_token": "<unk>",
45
+ "use_default_system_prompt": false
46
+ }