notzero commited on
Commit
c53119a
1 Parent(s): 1c95aac

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -3,30 +3,30 @@ tags:
3
  - merge
4
  - mergekit
5
  - lazymergekit
 
6
  - notzero/model_combined
7
- - EmbeddedLLM/Mistral-7B-Merge-14-v0.5
8
  base_model:
 
9
  - notzero/model_combined
10
- - EmbeddedLLM/Mistral-7B-Merge-14-v0.5
11
  ---
12
 
13
  # test_merge
14
 
15
  test_merge is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
 
16
  * [notzero/model_combined](https://huggingface.co/notzero/model_combined)
17
- * [EmbeddedLLM/Mistral-7B-Merge-14-v0.5](https://huggingface.co/EmbeddedLLM/Mistral-7B-Merge-14-v0.5)
18
 
19
  ## 🧩 Configuration
20
 
21
  ```yaml
22
  slices:
23
  - sources:
24
- - model: notzero/model_combined
25
  layer_range: [0, 32]
26
- - model: EmbeddedLLM/Mistral-7B-Merge-14-v0.5
27
  layer_range: [0, 32]
28
  merge_method: slerp
29
- base_model: notzero/model_combined
30
  parameters:
31
  t:
32
  - filter: self_attn
 
3
  - merge
4
  - mergekit
5
  - lazymergekit
6
+ - OpenPipe/mistral-ft-optimized-1218
7
  - notzero/model_combined
 
8
  base_model:
9
+ - OpenPipe/mistral-ft-optimized-1218
10
  - notzero/model_combined
 
11
  ---
12
 
13
  # test_merge
14
 
15
  test_merge is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
16
+ * [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)
17
  * [notzero/model_combined](https://huggingface.co/notzero/model_combined)
 
18
 
19
  ## 🧩 Configuration
20
 
21
  ```yaml
22
  slices:
23
  - sources:
24
+ - model: OpenPipe/mistral-ft-optimized-1218
25
  layer_range: [0, 32]
26
+ - model: notzero/model_combined
27
  layer_range: [0, 32]
28
  merge_method: slerp
29
+ base_model: OpenPipe/mistral-ft-optimized-1218
30
  parameters:
31
  t:
32
  - filter: self_attn
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "notzero/model_combined",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
@@ -21,7 +21,6 @@
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.38.2",
24
- "unsloth_version": "2024.2",
25
  "use_cache": true,
26
- "vocab_size": 48384
27
  }
 
1
  {
2
+ "_name_or_path": "OpenPipe/mistral-ft-optimized-1218",
3
  "architectures": [
4
  "MistralForCausalLM"
5
  ],
 
21
  "tie_word_embeddings": false,
22
  "torch_dtype": "bfloat16",
23
  "transformers_version": "4.38.2",
 
24
  "use_cache": true,
25
+ "vocab_size": 32000
26
  }
mergekit_config.yml CHANGED
@@ -1,12 +1,12 @@
1
 
2
  slices:
3
  - sources:
4
- - model: notzero/model_combined
5
  layer_range: [0, 32]
6
- - model: EmbeddedLLM/Mistral-7B-Merge-14-v0.5
7
  layer_range: [0, 32]
8
  merge_method: slerp
9
- base_model: notzero/model_combined
10
  parameters:
11
  t:
12
  - filter: self_attn
 
1
 
2
  slices:
3
  - sources:
4
+ - model: OpenPipe/mistral-ft-optimized-1218
5
  layer_range: [0, 32]
6
+ - model: notzero/model_combined
7
  layer_range: [0, 32]
8
  merge_method: slerp
9
+ base_model: OpenPipe/mistral-ft-optimized-1218
10
  parameters:
11
  t:
12
  - filter: self_attn
model-00001-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fbfcd5118db412ed9250a4627cadd0f0d76c8a9f22d8e4daa95198b12a98e489
3
  size 1946227344
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6bf48c6feec4050a9801e21ffd04e397685f0fe6e56bf7e8a3e0fae55dc17f16
3
  size 1946227344
model-00002-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9bdf09e81f65cee8f9c69aea7c37e70878b36075d10eef906dba7b9914bc669c
3
  size 1979781432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8be67e4abb48b92c9157a96b67b1430b581abacba42029dcbf21aeaf97f48d2b
3
  size 1979781432
model-00003-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:313123915390af6824f338bfb14deaddf1247449db5630b2de33a305cc48bc00
3
- size 1889603432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:20770224a8ab3f912e8540ee37f3c491327f57fb3e5e4607b54ec587fe07904c
3
+ size 1889587048
model-00004-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d2c02ec726ce851346cf8c10ae2ba744cf052afb5686acc9c06b06834db9d805
3
  size 1946243984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a8a34b1f103c487fb28cd05fddddb81dc979e7c9e3360b8ba002fc5eb6a4e4a8
3
  size 1946243984
model-00005-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c07af49535816b87fc9709008e6fc8680298404035754c6195b16a360b687adf
3
  size 1979798064
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5dafa07d0358b3d8f0910615e3fd317f2a7ed3b471e325aacd8b79bc08da2282
3
  size 1979798064
model-00006-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2f52c511d718c9586cc1072ca739363124f8c7bada8295f1d4a84e04cb1a4e10
3
- size 1973506120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a161653a4cbe203533cb34531b00b761e78e5d3fbd69910bdac95e2801c00fb0
3
+ size 1973489736
model-00007-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:471550d410f77b9f459418c27af107de3c573cf640ec91420688e886f44fc4e5
3
  size 1895903680
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:665961920f442c78e9fe45e5f2e9ba0b038e4d25c85da6480c0c4246e724c811
3
  size 1895903680
model-00008-of-00008.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:da5f8b0b33650a3dc1ecd669b7763f7e0d7b663c1603eb929a33ccd06a13451a
3
  size 872466704
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e01166f9fe2a17347343e0eb0922cd6381d08558d0216b8221393cf0bd2c864f
3
  size 872466704
special_tokens_map.json CHANGED
@@ -13,13 +13,6 @@
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
- "pad_token": {
17
- "content": "</s>",
18
- "lstrip": false,
19
- "normalized": false,
20
- "rstrip": false,
21
- "single_word": false
22
- },
23
  "unk_token": {
24
  "content": "<unk>",
25
  "lstrip": false,
 
13
  "rstrip": false,
14
  "single_word": false
15
  },
 
 
 
 
 
 
 
16
  "unk_token": {
17
  "content": "<unk>",
18
  "lstrip": false,
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer.model CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:4d88bdadaa2a065aa7c6e18a4b5999ce4c76cec14d9fea882102e7b4931d7ef0
3
- size 779539
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json CHANGED
@@ -29,13 +29,11 @@
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
32
- "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{{ bos_token }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '</s>'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
33
  "clean_up_tokenization_spaces": false,
34
  "eos_token": "</s>",
35
  "legacy": true,
36
- "model_max_length": 32768,
37
- "pad_token": "</s>",
38
- "padding_side": "right",
39
  "sp_model_kwargs": {},
40
  "spaces_between_special_tokens": false,
41
  "tokenizer_class": "LlamaTokenizer",
 
29
  },
30
  "additional_special_tokens": [],
31
  "bos_token": "<s>",
 
32
  "clean_up_tokenization_spaces": false,
33
  "eos_token": "</s>",
34
  "legacy": true,
35
+ "model_max_length": 1000000000000000019884624838656,
36
+ "pad_token": null,
 
37
  "sp_model_kwargs": {},
38
  "spaces_between_special_tokens": false,
39
  "tokenizer_class": "LlamaTokenizer",