Upload folder using huggingface_hub
Browse files- README.md +6 -6
- config.json +2 -3
- mergekit_config.yml +3 -3
- model-00001-of-00008.safetensors +1 -1
- model-00002-of-00008.safetensors +1 -1
- model-00003-of-00008.safetensors +2 -2
- model-00004-of-00008.safetensors +1 -1
- model-00005-of-00008.safetensors +1 -1
- model-00006-of-00008.safetensors +2 -2
- model-00007-of-00008.safetensors +1 -1
- model-00008-of-00008.safetensors +1 -1
- special_tokens_map.json +0 -7
- tokenizer.json +0 -0
- tokenizer.model +2 -2
- tokenizer_config.json +2 -4
README.md
CHANGED
@@ -3,30 +3,30 @@ tags:
|
|
3 |
- merge
|
4 |
- mergekit
|
5 |
- lazymergekit
|
|
|
6 |
- notzero/model_combined
|
7 |
-
- EmbeddedLLM/Mistral-7B-Merge-14-v0.5
|
8 |
base_model:
|
|
|
9 |
- notzero/model_combined
|
10 |
-
- EmbeddedLLM/Mistral-7B-Merge-14-v0.5
|
11 |
---
|
12 |
|
13 |
# test_merge
|
14 |
|
15 |
test_merge is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
|
|
16 |
* [notzero/model_combined](https://huggingface.co/notzero/model_combined)
|
17 |
-
* [EmbeddedLLM/Mistral-7B-Merge-14-v0.5](https://huggingface.co/EmbeddedLLM/Mistral-7B-Merge-14-v0.5)
|
18 |
|
19 |
## 🧩 Configuration
|
20 |
|
21 |
```yaml
|
22 |
slices:
|
23 |
- sources:
|
24 |
-
- model:
|
25 |
layer_range: [0, 32]
|
26 |
-
- model:
|
27 |
layer_range: [0, 32]
|
28 |
merge_method: slerp
|
29 |
-
base_model:
|
30 |
parameters:
|
31 |
t:
|
32 |
- filter: self_attn
|
|
|
3 |
- merge
|
4 |
- mergekit
|
5 |
- lazymergekit
|
6 |
+
- OpenPipe/mistral-ft-optimized-1218
|
7 |
- notzero/model_combined
|
|
|
8 |
base_model:
|
9 |
+
- OpenPipe/mistral-ft-optimized-1218
|
10 |
- notzero/model_combined
|
|
|
11 |
---
|
12 |
|
13 |
# test_merge
|
14 |
|
15 |
test_merge is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
|
16 |
+
* [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)
|
17 |
* [notzero/model_combined](https://huggingface.co/notzero/model_combined)
|
|
|
18 |
|
19 |
## 🧩 Configuration
|
20 |
|
21 |
```yaml
|
22 |
slices:
|
23 |
- sources:
|
24 |
+
- model: OpenPipe/mistral-ft-optimized-1218
|
25 |
layer_range: [0, 32]
|
26 |
+
- model: notzero/model_combined
|
27 |
layer_range: [0, 32]
|
28 |
merge_method: slerp
|
29 |
+
base_model: OpenPipe/mistral-ft-optimized-1218
|
30 |
parameters:
|
31 |
t:
|
32 |
- filter: self_attn
|
config.json
CHANGED
@@ -1,5 +1,5 @@
|
|
1 |
{
|
2 |
-
"_name_or_path": "
|
3 |
"architectures": [
|
4 |
"MistralForCausalLM"
|
5 |
],
|
@@ -21,7 +21,6 @@
|
|
21 |
"tie_word_embeddings": false,
|
22 |
"torch_dtype": "bfloat16",
|
23 |
"transformers_version": "4.38.2",
|
24 |
-
"unsloth_version": "2024.2",
|
25 |
"use_cache": true,
|
26 |
-
"vocab_size":
|
27 |
}
|
|
|
1 |
{
|
2 |
+
"_name_or_path": "OpenPipe/mistral-ft-optimized-1218",
|
3 |
"architectures": [
|
4 |
"MistralForCausalLM"
|
5 |
],
|
|
|
21 |
"tie_word_embeddings": false,
|
22 |
"torch_dtype": "bfloat16",
|
23 |
"transformers_version": "4.38.2",
|
|
|
24 |
"use_cache": true,
|
25 |
+
"vocab_size": 32000
|
26 |
}
|
mergekit_config.yml
CHANGED
@@ -1,12 +1,12 @@
|
|
1 |
|
2 |
slices:
|
3 |
- sources:
|
4 |
-
- model:
|
5 |
layer_range: [0, 32]
|
6 |
-
- model:
|
7 |
layer_range: [0, 32]
|
8 |
merge_method: slerp
|
9 |
-
base_model:
|
10 |
parameters:
|
11 |
t:
|
12 |
- filter: self_attn
|
|
|
1 |
|
2 |
slices:
|
3 |
- sources:
|
4 |
+
- model: OpenPipe/mistral-ft-optimized-1218
|
5 |
layer_range: [0, 32]
|
6 |
+
- model: notzero/model_combined
|
7 |
layer_range: [0, 32]
|
8 |
merge_method: slerp
|
9 |
+
base_model: OpenPipe/mistral-ft-optimized-1218
|
10 |
parameters:
|
11 |
t:
|
12 |
- filter: self_attn
|
model-00001-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1946227344
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6bf48c6feec4050a9801e21ffd04e397685f0fe6e56bf7e8a3e0fae55dc17f16
|
3 |
size 1946227344
|
model-00002-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1979781432
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8be67e4abb48b92c9157a96b67b1430b581abacba42029dcbf21aeaf97f48d2b
|
3 |
size 1979781432
|
model-00003-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:20770224a8ab3f912e8540ee37f3c491327f57fb3e5e4607b54ec587fe07904c
|
3 |
+
size 1889587048
|
model-00004-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1946243984
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a8a34b1f103c487fb28cd05fddddb81dc979e7c9e3360b8ba002fc5eb6a4e4a8
|
3 |
size 1946243984
|
model-00005-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1979798064
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5dafa07d0358b3d8f0910615e3fd317f2a7ed3b471e325aacd8b79bc08da2282
|
3 |
size 1979798064
|
model-00006-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a161653a4cbe203533cb34531b00b761e78e5d3fbd69910bdac95e2801c00fb0
|
3 |
+
size 1973489736
|
model-00007-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 1895903680
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:665961920f442c78e9fe45e5f2e9ba0b038e4d25c85da6480c0c4246e724c811
|
3 |
size 1895903680
|
model-00008-of-00008.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 872466704
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e01166f9fe2a17347343e0eb0922cd6381d08558d0216b8221393cf0bd2c864f
|
3 |
size 872466704
|
special_tokens_map.json
CHANGED
@@ -13,13 +13,6 @@
|
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
},
|
16 |
-
"pad_token": {
|
17 |
-
"content": "</s>",
|
18 |
-
"lstrip": false,
|
19 |
-
"normalized": false,
|
20 |
-
"rstrip": false,
|
21 |
-
"single_word": false
|
22 |
-
},
|
23 |
"unk_token": {
|
24 |
"content": "<unk>",
|
25 |
"lstrip": false,
|
|
|
13 |
"rstrip": false,
|
14 |
"single_word": false
|
15 |
},
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
16 |
"unk_token": {
|
17 |
"content": "<unk>",
|
18 |
"lstrip": false,
|
tokenizer.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
3 |
+
size 493443
|
tokenizer_config.json
CHANGED
@@ -29,13 +29,11 @@
|
|
29 |
},
|
30 |
"additional_special_tokens": [],
|
31 |
"bos_token": "<s>",
|
32 |
-
"chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{{ bos_token }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '</s>'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
33 |
"clean_up_tokenization_spaces": false,
|
34 |
"eos_token": "</s>",
|
35 |
"legacy": true,
|
36 |
-
"model_max_length":
|
37 |
-
"pad_token":
|
38 |
-
"padding_side": "right",
|
39 |
"sp_model_kwargs": {},
|
40 |
"spaces_between_special_tokens": false,
|
41 |
"tokenizer_class": "LlamaTokenizer",
|
|
|
29 |
},
|
30 |
"additional_special_tokens": [],
|
31 |
"bos_token": "<s>",
|
|
|
32 |
"clean_up_tokenization_spaces": false,
|
33 |
"eos_token": "</s>",
|
34 |
"legacy": true,
|
35 |
+
"model_max_length": 1000000000000000019884624838656,
|
36 |
+
"pad_token": null,
|
|
|
37 |
"sp_model_kwargs": {},
|
38 |
"spaces_between_special_tokens": false,
|
39 |
"tokenizer_class": "LlamaTokenizer",
|