cstr commited on
Commit
75ce5a7
1 Parent(s): c00fcd8

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -3,47 +3,41 @@ tags:
3
  - merge
4
  - mergekit
5
  - lazymergekit
6
- - abhishek/autotrain-llama3-8b-open-hermes-sft
7
- - cognitivecomputations/dolphin-2.9-llama3-8b
8
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
9
  base_model:
10
- - abhishek/autotrain-llama3-8b-open-hermes-sft
11
- - cognitivecomputations/dolphin-2.9-llama3-8b
12
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
13
  ---
14
 
15
  # llama3-discolm-orpo-t2
16
 
17
  llama3-discolm-orpo-t2 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
18
- * [abhishek/autotrain-llama3-8b-open-hermes-sft](https://huggingface.co/abhishek/autotrain-llama3-8b-open-hermes-sft)
19
- * [cognitivecomputations/dolphin-2.9-llama3-8b](https://huggingface.co/cognitivecomputations/dolphin-2.9-llama3-8b)
20
  * [DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental](https://huggingface.co/DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental)
21
 
22
  ## 🧩 Configuration
23
 
24
  ```yaml
25
  models:
26
- - model: abhishek/autotrain-llama3-8b-open-hermes-sft
 
 
27
  parameters:
28
- density: 0.5
29
- weight: 0.4
30
- - model: cognitivecomputations/dolphin-2.9-llama3-8b
31
- parameters:
32
- density: 0.5
33
- weight: 0.3
34
  - model: DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
35
  parameters:
36
- density: 0.6
37
- weight: [0, 0.3, 0.7, 1]
38
- # - filter: mlp
39
- # value: 0.5
40
- # - value: 0.3
41
- merge_method: ties
42
- base_model: mlabonne/OrpoLlama-3-8B
43
  parameters:
44
- normalize: true
45
  int8_mask: true
46
  dtype: bfloat16
 
 
 
47
  ```
48
 
49
  ## 💻 Usage
 
3
  - merge
4
  - mergekit
5
  - lazymergekit
6
+ - Locutusque/llama-3-neural-chat-v1-8b
 
7
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
8
  base_model:
9
+ - Locutusque/llama-3-neural-chat-v1-8b
 
10
  - DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
11
  ---
12
 
13
  # llama3-discolm-orpo-t2
14
 
15
  llama3-discolm-orpo-t2 is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
+ * [Locutusque/llama-3-neural-chat-v1-8b](https://huggingface.co/Locutusque/llama-3-neural-chat-v1-8b)
 
17
  * [DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental](https://huggingface.co/DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental)
18
 
19
  ## 🧩 Configuration
20
 
21
  ```yaml
22
  models:
23
+ - model: Locutusque/Llama-3-Orca-1.0-8B
24
+ # no parameters necessary for base model
25
+ - model: Locutusque/llama-3-neural-chat-v1-8b
26
  parameters:
27
+ density: 0.60
28
+ weight: 0.15
 
 
 
 
29
  - model: DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
30
  parameters:
31
+ density: 0.65
32
+ weight: 0.7
33
+ merge_method: dare_ties
34
+ base_model: Locutusque/Llama-3-Orca-1.0-8B
 
 
 
35
  parameters:
 
36
  int8_mask: true
37
  dtype: bfloat16
38
+ random_seed: 0
39
+ tokenizer_source: base
40
+
41
  ```
42
 
43
  ## 💻 Usage
config.json CHANGED
@@ -1,12 +1,12 @@
1
  {
2
- "_name_or_path": "mlabonne/OrpoLlama-3-8B",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
- "bos_token_id": 128256,
9
- "eos_token_id": 128257,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
@@ -16,7 +16,6 @@
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
19
- "pad_token_id": 128257,
20
  "pretraining_tp": 1,
21
  "rms_norm_eps": 1e-05,
22
  "rope_scaling": null,
@@ -25,5 +24,5 @@
25
  "torch_dtype": "bfloat16",
26
  "transformers_version": "4.39.3",
27
  "use_cache": true,
28
- "vocab_size": 128258
29
  }
 
1
  {
2
+ "_name_or_path": "Locutusque/Llama-3-Orca-1.0-8B",
3
  "architectures": [
4
  "LlamaForCausalLM"
5
  ],
6
  "attention_bias": false,
7
  "attention_dropout": 0.0,
8
+ "bos_token_id": 128000,
9
+ "eos_token_id": 128001,
10
  "hidden_act": "silu",
11
  "hidden_size": 4096,
12
  "initializer_range": 0.02,
 
16
  "num_attention_heads": 32,
17
  "num_hidden_layers": 32,
18
  "num_key_value_heads": 8,
 
19
  "pretraining_tp": 1,
20
  "rms_norm_eps": 1e-05,
21
  "rope_scaling": null,
 
24
  "torch_dtype": "bfloat16",
25
  "transformers_version": "4.39.3",
26
  "use_cache": true,
27
+ "vocab_size": 128256
28
  }
mergekit_config.yml CHANGED
@@ -1,17 +1,20 @@
1
 
2
- slices:
3
- - sources:
4
- - model: mlabonne/OrpoLlama-3-8B
5
- layer_range: [0, 32]
6
- - model: DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
7
- layer_range: [0, 32]
8
- merge_method: slerp
9
- base_model: mlabonne/OrpoLlama-3-8B
 
 
 
 
 
10
  parameters:
11
- t:
12
- - filter: self_attn
13
- value: [1, 0.7, 0.5, 0.3, 0.1]
14
- - filter: mlp
15
- value: [0, 0.3, 0.5, 0.7, 0.9]
16
- - value: 0.5
17
  dtype: bfloat16
 
 
 
 
1
 
2
+ models:
3
+ - model: Locutusque/Llama-3-Orca-1.0-8B
4
+ # no parameters necessary for base model
5
+ - model: Locutusque/llama-3-neural-chat-v1-8b
6
+ parameters:
7
+ density: 0.60
8
+ weight: 0.15
9
+ - model: DiscoResearch/Llama3_DiscoLM_German_8b_v0.1_experimental
10
+ parameters:
11
+ density: 0.65
12
+ weight: 0.7
13
+ merge_method: dare_ties
14
+ base_model: Locutusque/Llama-3-Orca-1.0-8B
15
  parameters:
16
+ int8_mask: true
 
 
 
 
 
17
  dtype: bfloat16
18
+ random_seed: 0
19
+ tokenizer_source: base
20
+
model-00001-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ea7dbe23c8a4216515c732d86d1d1dfbb9d7bcdc99adccf1dafff911d5d22fca
3
- size 1979781432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cc30c136d6d9bbcf0c9914b7532d28a9380a299233a929a0658427516d192920
3
+ size 1973455376
model-00002-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:894dd03966f7d27772fd51b1b5953aa550ed75defc4a47f6231628dd6bf0a9c5
3
- size 1946243944
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:78aa0ce8e2086002087b3b9144451aa84a20d0db8487d6019b6d607e548777f7
3
+ size 1895895336
model-00003-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9e7fd0940e6b6b0ec61a6022e5fca54f0d35e46ac01ff8142dc93f7959a5e24c
3
- size 1923106776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27c585c80dbfc08297f3ed35217684e2cbbb763cf6b1a88401a695b2a4aa29b7
3
+ size 1979798040
model-00004-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0fd865707be249dc2979e496035dcd59c068fcb0075ea9031e60d289abeb989d
3
- size 1946243984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6606b4a1fe0873adeff97a7e1e6e91a56feaddf70d1fc9dacd2edfdd51051972
3
+ size 1946227368
model-00005-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f78e0dc55a781fe4eea66e6465e81c71592e88428bd03db5704ef84207cb7249
3
- size 1979781440
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:921507a21f07fb31ddb57a506ead667f81a3e7851837d5c8fcb392febd46b0b6
3
+ size 1979798064
model-00006-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e1e52389df2bfa658c1da3185dfbc248f28a776117d9985ba3b76180e487006f
3
  size 1946227368
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:94a3cfdcbee86b0c580c0b1eda1a8fed858f28e3bbb1af99b0b721aa8e8f50f5
3
  size 1946227368
model-00007-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:477275a21da2fd43d8dde246b534a8dfff0f1e2730467cb908c26f664d56fd09
3
- size 1979789776
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:823c139db3ae4451a555bdd69076326daf61daf47b29a4cebc477e9a4b99e911
3
+ size 1979806352
model-00008-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:475d44f97f66ded653c2630210d5cf4789b84567eb107f53d47f79541aac926e
3
- size 1191242824
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cdb2ad44c84e8da8c6d56c92a3f8274e56d0b6dcb4fe7e08f967ef1394334055
3
+ size 1308675136
model-00009-of-00009.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e85e224321851428be2d341c3d4ee0894ed483d8db83a0b50cc2abae2ff6244
3
- size 1168138808
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68b033e3fcfecf40639a530f7a6de46f152c9f463d1c5ec56bf25bde00cffb3f
3
+ size 1050673280
model.safetensors.index.json CHANGED
@@ -1 +1 @@
1
- {"metadata": {"mergekit_version": "0.0.4.2", "total_size": 16060522496}, "weight_map": {"model.layers.8.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.8.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.7.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.6.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.5.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.1.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.0.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.embed_tokens.weight": "model-00003-of-00009.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.19.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.18.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.17.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.16.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.13.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.12.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.11.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.10.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.9.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.30.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.29.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.28.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.23.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.22.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.21.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.20.input_layernorm.weight": "model-00008-of-00009.safetensors", "lm_head.weight": "model-00009-of-00009.safetensors", "model.norm.weight": "model-00009-of-00009.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00009-of-00009.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00009-of-00009.safetensors", "model.layers.31.input_layernorm.weight": "model-00009-of-00009.safetensors"}}
 
1
+ {"metadata": {"mergekit_version": "0.0.4.2", "total_size": 16060522496}, "weight_map": {"model.layers.2.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.1.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00009.safetensors", "model.layers.0.input_layernorm.weight": "model-00001-of-00009.safetensors", "model.embed_tokens.weight": "model-00001-of-00009.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.5.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.4.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00002-of-00009.safetensors", "model.layers.3.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00009.safetensors", "model.layers.2.input_layernorm.weight": "model-00002-of-00009.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.10.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.9.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.8.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00009.safetensors", "model.layers.7.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00003-of-00009.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00003-of-00009.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.6.input_layernorm.weight": "model-00003-of-00009.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.14.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.13.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.12.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00004-of-00009.safetensors", "model.layers.11.input_layernorm.weight": "model-00004-of-00009.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.19.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.18.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.17.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00005-of-00009.safetensors", "model.layers.16.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00005-of-00009.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00005-of-00009.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.15.input_layernorm.weight": "model-00005-of-00009.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.23.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.22.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.21.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00006-of-00009.safetensors", "model.layers.20.input_layernorm.weight": "model-00006-of-00009.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.28.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.27.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.26.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00007-of-00009.safetensors", "model.layers.25.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00007-of-00009.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00007-of-00009.safetensors", "model.layers.24.input_layernorm.weight": "model-00007-of-00009.safetensors", "model.norm.weight": "model-00007-of-00009.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.31.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.30.input_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00008-of-00009.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00008-of-00009.safetensors", "model.layers.29.input_layernorm.weight": "model-00008-of-00009.safetensors", "lm_head.weight": "model-00009-of-00009.safetensors"}}
special_tokens_map.json CHANGED
@@ -1,24 +1,20 @@
1
  {
2
- "additional_special_tokens": [
3
- "<|im_start|>",
4
- "<|im_end|>"
5
- ],
6
  "bos_token": {
7
- "content": "<|im_start|>",
8
  "lstrip": false,
9
  "normalized": false,
10
  "rstrip": false,
11
  "single_word": false
12
  },
13
  "eos_token": {
14
- "content": "<|im_end|>",
15
  "lstrip": false,
16
  "normalized": false,
17
  "rstrip": false,
18
  "single_word": false
19
  },
20
  "pad_token": {
21
- "content": "<|im_end|>",
22
  "lstrip": false,
23
  "normalized": false,
24
  "rstrip": false,
 
1
  {
 
 
 
 
2
  "bos_token": {
3
+ "content": "<|begin_of_text|>",
4
  "lstrip": false,
5
  "normalized": false,
6
  "rstrip": false,
7
  "single_word": false
8
  },
9
  "eos_token": {
10
+ "content": "<|end_of_text|>",
11
  "lstrip": false,
12
  "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
  "pad_token": {
17
+ "content": "<|end_of_text|>",
18
  "lstrip": false,
19
  "normalized": false,
20
  "rstrip": false,
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 128000,
@@ -2306,24 +2320,6 @@
2306
  "rstrip": false,
2307
  "normalized": false,
2308
  "special": true
2309
- },
2310
- {
2311
- "id": 128256,
2312
- "content": "<|im_start|>",
2313
- "single_word": false,
2314
- "lstrip": false,
2315
- "rstrip": false,
2316
- "normalized": false,
2317
- "special": true
2318
- },
2319
- {
2320
- "id": 128257,
2321
- "content": "<|im_end|>",
2322
- "single_word": false,
2323
- "lstrip": false,
2324
- "rstrip": false,
2325
- "normalized": false,
2326
- "special": true
2327
  }
2328
  ],
2329
  "normalizer": null,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 512
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 128001,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<|end_of_text|>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 128000,
 
2320
  "rstrip": false,
2321
  "normalized": false,
2322
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2323
  }
2324
  ],
2325
  "normalizer": null,
tokenizer_config.json CHANGED
@@ -2047,37 +2047,24 @@
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
2050
- },
2051
- "128256": {
2052
- "content": "<|im_start|>",
2053
- "lstrip": false,
2054
- "normalized": false,
2055
- "rstrip": false,
2056
- "single_word": false,
2057
- "special": true
2058
- },
2059
- "128257": {
2060
- "content": "<|im_end|>",
2061
- "lstrip": false,
2062
- "normalized": false,
2063
- "rstrip": false,
2064
- "single_word": false,
2065
- "special": true
2066
  }
2067
  },
2068
- "additional_special_tokens": [
2069
- "<|im_start|>",
2070
- "<|im_end|>"
2071
- ],
2072
- "bos_token": "<|im_start|>",
2073
- "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
2074
  "clean_up_tokenization_spaces": true,
2075
- "eos_token": "<|im_end|>",
 
2076
  "model_input_names": [
2077
  "input_ids",
2078
  "attention_mask"
2079
  ],
2080
  "model_max_length": 1000000000000000019884624838656,
2081
- "pad_token": "<|im_end|>",
2082
- "tokenizer_class": "PreTrainedTokenizerFast"
 
 
 
 
 
 
2083
  }
 
2047
  "rstrip": false,
2048
  "single_word": false,
2049
  "special": true
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2050
  }
2051
  },
2052
+ "bos_token": "<|begin_of_text|>",
2053
+ "chat_template": "{%- set ns = namespace(found=false) -%}{%- for message in messages -%}{%- if message['role'] == 'system' -%}{%- set ns.found = true -%}{%- endif -%}{%- endfor -%}{%- for message in messages %}{%- if message['role'] == 'system' -%}{{- '<|im_start|>system\n' + message['content'].rstrip() + '<|im_end|>\n' -}}{%- else -%}{%- if message['role'] == 'user' -%}{{-'<|im_start|>user\n' + message['content'].rstrip() + '<|im_end|>\n'-}}{%- else -%}{{-'<|im_start|>assistant\n' + message['content'] + '<|im_end|>\n' -}}{%- endif -%}{%- endif -%}{%- endfor -%}{%- if add_generation_prompt -%}{{-'<|im_start|>assistant\n'-}}{%- endif -%}",
 
 
 
 
2054
  "clean_up_tokenization_spaces": true,
2055
+ "eos_token": "<|end_of_text|>",
2056
+ "max_length": 512,
2057
  "model_input_names": [
2058
  "input_ids",
2059
  "attention_mask"
2060
  ],
2061
  "model_max_length": 1000000000000000019884624838656,
2062
+ "pad_to_multiple_of": null,
2063
+ "pad_token": "<|end_of_text|>",
2064
+ "pad_token_type_id": 0,
2065
+ "padding_side": "right",
2066
+ "stride": 0,
2067
+ "tokenizer_class": "PreTrainedTokenizerFast",
2068
+ "truncation_side": "right",
2069
+ "truncation_strategy": "longest_first"
2070
  }