Omotayo commited on
Commit
e71bbc4
1 Parent(s): dee7343

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - merge
5
+ - mergekit
6
+ - lazymergekit
7
+ - psmathur/orca_mini_v3_13b
8
+ - garage-bAInd/Platypus2-13B
9
+ ---
10
+
11
+ # ainao-7B-slerp
12
+
13
+ ainao-7B-slerp is a merge of the following models using [mergekit](https://github.com/cg123/mergekit):
14
+ * [psmathur/orca_mini_v3_13b](https://huggingface.co/psmathur/orca_mini_v3_13b)
15
+ * [garage-bAInd/Platypus2-13B](https://huggingface.co/garage-bAInd/Platypus2-13B)
16
+
17
+ ## 🧩 Configuration
18
+
19
+ ```yaml
20
+ slices:
21
+ - sources:
22
+ - model: psmathur/orca_mini_v3_13b
23
+ layer_range: [0, 32]
24
+ - model: garage-bAInd/Platypus2-13B
25
+ layer_range: [0, 32]
26
+ merge_method: slerp
27
+ base_model: psmathur/orca_mini_v3_13b
28
+ parameters:
29
+ t:
30
+ - filter: self_attn
31
+ value: [0, 0.5, 0.3, 0.7, 1]
32
+ - filter: mlp
33
+ value: [1, 0.5, 0.7, 0.3, 0]
34
+ - value: 0.5
35
+ dtype: bfloat16
36
+
37
+ ```
config.json ADDED
@@ -0,0 +1,28 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "psmathur/orca_mini_v3_13b",
3
+ "architectures": [
4
+ "LlamaForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "bos_token_id": 1,
8
+ "eos_token_id": 2,
9
+ "hidden_act": "silu",
10
+ "hidden_size": 5120,
11
+ "initializer_range": 0.02,
12
+ "intermediate_size": 13824,
13
+ "max_position_embeddings": 4096,
14
+ "model_type": "llama",
15
+ "num_attention_heads": 40,
16
+ "num_hidden_layers": 32,
17
+ "num_key_value_heads": 40,
18
+ "pad_token_id": 0,
19
+ "pretraining_tp": 1,
20
+ "rms_norm_eps": 1e-05,
21
+ "rope_scaling": null,
22
+ "rope_theta": 10000.0,
23
+ "tie_word_embeddings": false,
24
+ "torch_dtype": "bfloat16",
25
+ "transformers_version": "4.35.2",
26
+ "use_cache": true,
27
+ "vocab_size": 32000
28
+ }
mergekit_config.yml ADDED
@@ -0,0 +1,18 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
+ slices:
3
+ - sources:
4
+ - model: psmathur/orca_mini_v3_13b
5
+ layer_range: [0, 32]
6
+ - model: garage-bAInd/Platypus2-13B
7
+ layer_range: [0, 32]
8
+ merge_method: slerp
9
+ base_model: psmathur/orca_mini_v3_13b
10
+ parameters:
11
+ t:
12
+ - filter: self_attn
13
+ value: [0, 0.5, 0.3, 0.7, 1]
14
+ - filter: mlp
15
+ value: [1, 0.5, 0.7, 0.3, 0]
16
+ - value: 0.5
17
+ dtype: bfloat16
18
+
model-00001-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b41fdc6318764c03ab0203939caad78a2cbd66243bef77e24797c932cf7bdf7
3
+ size 1955648616
model-00002-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6682b541c1d5a5b6d07b6443b5df60ca4887b0d974a63f3b825ab7b545c8a42c
3
+ size 1955669296
model-00003-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:024a369d7d01daf0fd50c157f95188680480a2603d1931feeab6230ebd57565c
3
+ size 1903230008
model-00004-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:21f305884233cf17a2965386dbd6c1c96019bc7a45947063d05f964f48c490b3
3
+ size 1903230008
model-00005-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b9442000f23609071c910449cb3f452bd6bb5faf75fea774dc388bcdcec2e3e6
3
+ size 1903230008
model-00006-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f3821d79d647c6d6997eabc319bbefc823b3906ac3da5ad4bf7955417c49b771
3
+ size 1984463376
model-00007-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5e2ae50c9283500ff4c4ee1002b1c836317fc90b5052dc29695520226b1b20b5
3
+ size 1955658968
model-00008-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:239bd534e270cc1ebe9b36c06e0b4a153879fbd9542e0ea69c0fdc4c2b85826a
3
+ size 1955669304
model-00009-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8be449b98d724bf633d45464601df349f75675f52fd48b7dfe860a9d0bfde83d
3
+ size 1903230032
model-00010-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:07796758f4d9e58d553bd62bc3c33b4ae1d3dd81db96efd5e2e8a5719b41f4ec
3
+ size 1903230032
model-00011-of-00011.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:493029fac02ffd2cb08d5143e9d5aadb6f8ef6bf9a6915b4d9eacae0af03da43
3
+ size 1633231208
model.safetensors.index.json ADDED
@@ -0,0 +1 @@
 
 
1
+ {"metadata": {"mergekit_version": "0.0.4.1"}, "weight_map": {"model.layers.15.self_attn.k_proj.weight": "model-00001-of-00011.safetensors", "model.layers.15.self_attn.q_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.mlp.down_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.mlp.gate_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.mlp.up_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.post_attention_layernorm.weight": "model-00001-of-00011.safetensors", "model.layers.14.self_attn.o_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.self_attn.v_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.self_attn.k_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.self_attn.q_proj.weight": "model-00001-of-00011.safetensors", "model.layers.14.input_layernorm.weight": "model-00001-of-00011.safetensors", "model.layers.13.mlp.down_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.mlp.gate_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.mlp.up_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.post_attention_layernorm.weight": "model-00001-of-00011.safetensors", "model.layers.13.self_attn.o_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.self_attn.v_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.self_attn.k_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.self_attn.q_proj.weight": "model-00001-of-00011.safetensors", "model.layers.13.input_layernorm.weight": "model-00001-of-00011.safetensors", "model.layers.12.mlp.down_proj.weight": "model-00001-of-00011.safetensors", "model.layers.12.mlp.gate_proj.weight": "model-00001-of-00011.safetensors", "model.layers.12.mlp.up_proj.weight": "model-00001-of-00011.safetensors", "model.layers.12.post_attention_layernorm.weight": "model-00001-of-00011.safetensors", "model.layers.12.self_attn.o_proj.weight": "model-00001-of-00011.safetensors", "model.layers.12.self_attn.v_proj.weight": "model-00001-of-00011.safetensors", "model.layers.12.self_attn.k_proj.weight": "model-00001-of-00011.safetensors", "model.layers.12.self_attn.q_proj.weight": "model-00002-of-00011.safetensors", "model.layers.12.input_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.11.mlp.down_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.mlp.gate_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.mlp.up_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.post_attention_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.11.self_attn.o_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.self_attn.v_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.self_attn.k_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.self_attn.q_proj.weight": "model-00002-of-00011.safetensors", "model.layers.11.input_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.10.mlp.down_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.mlp.gate_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.mlp.up_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.post_attention_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.10.self_attn.o_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.self_attn.v_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.self_attn.k_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.self_attn.q_proj.weight": "model-00002-of-00011.safetensors", "model.layers.10.input_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.9.mlp.down_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.mlp.gate_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.mlp.up_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.post_attention_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.9.self_attn.o_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.self_attn.v_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.self_attn.k_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.self_attn.q_proj.weight": "model-00002-of-00011.safetensors", "model.layers.9.input_layernorm.weight": "model-00002-of-00011.safetensors", "model.layers.8.mlp.down_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.mlp.gate_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.mlp.up_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.post_attention_layernorm.weight": "model-00003-of-00011.safetensors", "model.layers.8.self_attn.o_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.self_attn.v_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.self_attn.k_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.self_attn.q_proj.weight": "model-00003-of-00011.safetensors", "model.layers.8.input_layernorm.weight": "model-00003-of-00011.safetensors", "model.layers.7.mlp.down_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.mlp.gate_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.mlp.up_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.post_attention_layernorm.weight": "model-00003-of-00011.safetensors", "model.layers.7.self_attn.o_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.self_attn.v_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.self_attn.k_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.self_attn.q_proj.weight": "model-00003-of-00011.safetensors", "model.layers.7.input_layernorm.weight": "model-00003-of-00011.safetensors", "model.layers.6.mlp.down_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.mlp.gate_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.mlp.up_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.post_attention_layernorm.weight": "model-00003-of-00011.safetensors", "model.layers.6.self_attn.o_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.self_attn.v_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.self_attn.k_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.self_attn.q_proj.weight": "model-00003-of-00011.safetensors", "model.layers.6.input_layernorm.weight": "model-00003-of-00011.safetensors", "model.layers.5.mlp.down_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.mlp.gate_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.mlp.up_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.post_attention_layernorm.weight": "model-00004-of-00011.safetensors", "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00011.safetensors", "model.layers.5.input_layernorm.weight": "model-00004-of-00011.safetensors", "model.layers.4.mlp.down_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.mlp.up_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00011.safetensors", "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00011.safetensors", "model.layers.4.input_layernorm.weight": "model-00004-of-00011.safetensors", "model.layers.3.mlp.down_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.mlp.gate_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.mlp.up_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00011.safetensors", "model.layers.3.self_attn.o_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.self_attn.v_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.self_attn.k_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.self_attn.q_proj.weight": "model-00004-of-00011.safetensors", "model.layers.3.input_layernorm.weight": "model-00004-of-00011.safetensors", "model.layers.2.mlp.down_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.mlp.gate_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.mlp.up_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.post_attention_layernorm.weight": "model-00005-of-00011.safetensors", "model.layers.2.self_attn.o_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.self_attn.v_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.self_attn.k_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.self_attn.q_proj.weight": "model-00005-of-00011.safetensors", "model.layers.2.input_layernorm.weight": "model-00005-of-00011.safetensors", "model.layers.1.mlp.down_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.mlp.gate_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.mlp.up_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.post_attention_layernorm.weight": "model-00005-of-00011.safetensors", "model.layers.1.self_attn.o_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.self_attn.v_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.self_attn.k_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.self_attn.q_proj.weight": "model-00005-of-00011.safetensors", "model.layers.1.input_layernorm.weight": "model-00005-of-00011.safetensors", "model.layers.0.mlp.down_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.mlp.gate_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.mlp.up_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.post_attention_layernorm.weight": "model-00005-of-00011.safetensors", "model.layers.0.self_attn.o_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.self_attn.v_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.self_attn.k_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.self_attn.q_proj.weight": "model-00005-of-00011.safetensors", "model.layers.0.input_layernorm.weight": "model-00005-of-00011.safetensors", "model.embed_tokens.weight": "model-00006-of-00011.safetensors", "model.layers.30.mlp.gate_proj.weight": "model-00006-of-00011.safetensors", "model.layers.30.mlp.up_proj.weight": "model-00006-of-00011.safetensors", "model.layers.30.self_attn.o_proj.weight": "model-00006-of-00011.safetensors", "model.layers.30.self_attn.v_proj.weight": "model-00006-of-00011.safetensors", "model.layers.30.self_attn.k_proj.weight": "model-00006-of-00011.safetensors", "model.layers.30.self_attn.q_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.mlp.down_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.mlp.gate_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.mlp.up_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.post_attention_layernorm.weight": "model-00006-of-00011.safetensors", "model.layers.29.self_attn.o_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.self_attn.v_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.self_attn.k_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.self_attn.q_proj.weight": "model-00006-of-00011.safetensors", "model.layers.29.input_layernorm.weight": "model-00006-of-00011.safetensors", "model.layers.28.mlp.down_proj.weight": "model-00006-of-00011.safetensors", "model.layers.28.mlp.gate_proj.weight": "model-00006-of-00011.safetensors", "model.layers.28.mlp.up_proj.weight": "model-00006-of-00011.safetensors", "model.layers.28.post_attention_layernorm.weight": "model-00006-of-00011.safetensors", "model.layers.28.self_attn.o_proj.weight": "model-00006-of-00011.safetensors", "model.layers.28.self_attn.v_proj.weight": "model-00006-of-00011.safetensors", "model.layers.28.self_attn.k_proj.weight": "model-00007-of-00011.safetensors", "model.layers.28.self_attn.q_proj.weight": "model-00007-of-00011.safetensors", "model.layers.28.input_layernorm.weight": "model-00007-of-00011.safetensors", "model.layers.27.mlp.down_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.mlp.gate_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.mlp.up_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.post_attention_layernorm.weight": "model-00007-of-00011.safetensors", "model.layers.27.self_attn.o_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.self_attn.v_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.self_attn.k_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.self_attn.q_proj.weight": "model-00007-of-00011.safetensors", "model.layers.27.input_layernorm.weight": "model-00007-of-00011.safetensors", "model.layers.26.mlp.down_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.mlp.gate_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.mlp.up_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.post_attention_layernorm.weight": "model-00007-of-00011.safetensors", "model.layers.26.self_attn.o_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.self_attn.v_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.self_attn.k_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.self_attn.q_proj.weight": "model-00007-of-00011.safetensors", "model.layers.26.input_layernorm.weight": "model-00007-of-00011.safetensors", "model.layers.25.mlp.down_proj.weight": "model-00007-of-00011.safetensors", "model.layers.25.mlp.gate_proj.weight": "model-00007-of-00011.safetensors", "model.layers.25.mlp.up_proj.weight": "model-00007-of-00011.safetensors", "model.layers.25.post_attention_layernorm.weight": "model-00007-of-00011.safetensors", "model.layers.25.self_attn.o_proj.weight": "model-00007-of-00011.safetensors", "model.layers.25.self_attn.v_proj.weight": "model-00007-of-00011.safetensors", "model.layers.25.self_attn.k_proj.weight": "model-00007-of-00011.safetensors", "model.layers.25.self_attn.q_proj.weight": "model-00008-of-00011.safetensors", "model.layers.25.input_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.24.mlp.down_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.mlp.gate_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.mlp.up_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.post_attention_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.24.self_attn.o_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.self_attn.v_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.self_attn.k_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.self_attn.q_proj.weight": "model-00008-of-00011.safetensors", "model.layers.24.input_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.23.mlp.down_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.mlp.gate_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.mlp.up_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.post_attention_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.23.self_attn.o_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.self_attn.v_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.self_attn.k_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.self_attn.q_proj.weight": "model-00008-of-00011.safetensors", "model.layers.23.input_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.22.mlp.down_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.mlp.gate_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.mlp.up_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.post_attention_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.22.self_attn.o_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.self_attn.v_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.self_attn.k_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.self_attn.q_proj.weight": "model-00008-of-00011.safetensors", "model.layers.22.input_layernorm.weight": "model-00008-of-00011.safetensors", "model.layers.21.mlp.down_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.mlp.gate_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.mlp.up_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.post_attention_layernorm.weight": "model-00009-of-00011.safetensors", "model.layers.21.self_attn.o_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.self_attn.v_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.self_attn.k_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.self_attn.q_proj.weight": "model-00009-of-00011.safetensors", "model.layers.21.input_layernorm.weight": "model-00009-of-00011.safetensors", "model.layers.20.mlp.down_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.mlp.gate_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.mlp.up_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.post_attention_layernorm.weight": "model-00009-of-00011.safetensors", "model.layers.20.self_attn.o_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.self_attn.v_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.self_attn.k_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.self_attn.q_proj.weight": "model-00009-of-00011.safetensors", "model.layers.20.input_layernorm.weight": "model-00009-of-00011.safetensors", "model.layers.19.mlp.down_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.mlp.gate_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.mlp.up_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.post_attention_layernorm.weight": "model-00009-of-00011.safetensors", "model.layers.19.self_attn.o_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.self_attn.v_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.self_attn.k_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.self_attn.q_proj.weight": "model-00009-of-00011.safetensors", "model.layers.19.input_layernorm.weight": "model-00009-of-00011.safetensors", "model.layers.18.mlp.down_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.mlp.gate_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.mlp.up_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.post_attention_layernorm.weight": "model-00010-of-00011.safetensors", "model.layers.18.self_attn.o_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.self_attn.v_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.self_attn.k_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.self_attn.q_proj.weight": "model-00010-of-00011.safetensors", "model.layers.18.input_layernorm.weight": "model-00010-of-00011.safetensors", "model.layers.17.mlp.down_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.mlp.gate_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.mlp.up_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.post_attention_layernorm.weight": "model-00010-of-00011.safetensors", "model.layers.17.self_attn.o_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.self_attn.v_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.self_attn.k_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.self_attn.q_proj.weight": "model-00010-of-00011.safetensors", "model.layers.17.input_layernorm.weight": "model-00010-of-00011.safetensors", "model.layers.16.mlp.down_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.mlp.gate_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.mlp.up_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.post_attention_layernorm.weight": "model-00010-of-00011.safetensors", "model.layers.16.self_attn.o_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.self_attn.v_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.self_attn.k_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.self_attn.q_proj.weight": "model-00010-of-00011.safetensors", "model.layers.16.input_layernorm.weight": "model-00010-of-00011.safetensors", "model.layers.15.mlp.down_proj.weight": "model-00011-of-00011.safetensors", "model.layers.15.mlp.gate_proj.weight": "model-00011-of-00011.safetensors", "model.layers.15.mlp.up_proj.weight": "model-00011-of-00011.safetensors", "model.layers.15.post_attention_layernorm.weight": "model-00011-of-00011.safetensors", "model.layers.15.self_attn.o_proj.weight": "model-00011-of-00011.safetensors", "model.layers.15.self_attn.v_proj.weight": "model-00011-of-00011.safetensors", "model.layers.15.input_layernorm.weight": "model-00011-of-00011.safetensors", "lm_head.weight": "model-00011-of-00011.safetensors", "model.norm.weight": "model-00011-of-00011.safetensors", "model.layers.31.mlp.down_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.mlp.gate_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.mlp.up_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.post_attention_layernorm.weight": "model-00011-of-00011.safetensors", "model.layers.31.self_attn.o_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.self_attn.v_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.self_attn.k_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.self_attn.q_proj.weight": "model-00011-of-00011.safetensors", "model.layers.31.input_layernorm.weight": "model-00011-of-00011.safetensors", "model.layers.30.mlp.down_proj.weight": "model-00011-of-00011.safetensors", "model.layers.30.post_attention_layernorm.weight": "model-00011-of-00011.safetensors", "model.layers.30.input_layernorm.weight": "model-00011-of-00011.safetensors"}}
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": {
3
+ "content": "<s>",
4
+ "lstrip": false,
5
+ "normalized": true,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "eos_token": {
10
+ "content": "</s>",
11
+ "lstrip": false,
12
+ "normalized": true,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": true,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<unk>",
5
+ "lstrip": false,
6
+ "normalized": true,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "<s>",
13
+ "lstrip": false,
14
+ "normalized": true,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "</s>",
21
+ "lstrip": false,
22
+ "normalized": true,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ }
27
+ },
28
+ "bos_token": "<s>",
29
+ "clean_up_tokenization_spaces": false,
30
+ "eos_token": "</s>",
31
+ "model_max_length": 4096,
32
+ "pad_token": null,
33
+ "sp_model_kwargs": {},
34
+ "tokenizer_class": "LlamaTokenizer",
35
+ "unk_token": "<unk>",
36
+ "use_default_system_prompt": false
37
+ }