Upload folder using huggingface_hub
Browse files- .ipynb_checkpoints/README-checkpoint.md +2 -42
- README.md +2 -42
- image.webp +0 -0
- mergekit_config.yml +4 -21
- model-00001-of-00005.safetensors +1 -1
- model-00002-of-00005.safetensors +1 -1
- model-00003-of-00005.safetensors +1 -1
- model-00004-of-00005.safetensors +1 -1
- model-00005-of-00005.safetensors +1 -1
.ipynb_checkpoints/README-checkpoint.md
CHANGED
@@ -14,51 +14,11 @@ tags:
|
|
14 |
---
|
15 |
# Llamazing-3.1-8B-Instruct
|
16 |
|
|
|
|
|
17 |
### Overview
|
18 |
Llamazing-3.1-8B-Instruct is a state-of-the-art language model resulting from the thoughtful merging of six specialized models. It balances reasoning, creativity, and conversational capabilities to deliver exceptional performance across various applications.
|
19 |
|
20 |
-
### Models Merged
|
21 |
-
The following models contributed to the creation of Llamazing-3.1-8B-Instruct:
|
22 |
-
|
23 |
-
1. [DreadPoor/Zelus-8B-Model_Stock](https://huggingface.co/DreadPoor/Zelus-8B-Model_Stock)
|
24 |
-
2. [HumanLLMs/Human-Like-LLama3-8B-Instruct](https://huggingface.co/HumanLLMs/Human-Like-LLama3-8B-Instruct)
|
25 |
-
3. [sethuiyer/Llamaverse-3.1-8B-Instruct](https://huggingface.co/sethuiyer/Llamaverse-3.1-8B-Instruct)
|
26 |
-
4. [CultriX/Llama3-8B-DPO](https://huggingface.co/CultriX/Llama3-8B-DPO)
|
27 |
-
5. [DevsDoCode/LLama-3-8b-Uncensored](https://huggingface.co/DevsDoCode/LLama-3-8b-Uncensored)
|
28 |
-
6. [HiroseKoichi/Llama-3-8B-Stroganoff](https://huggingface.co/HiroseKoichi/Llama-3-8B-Stroganoff)
|
29 |
-
|
30 |
-
### Configuration
|
31 |
-
Below is the YAML configuration used for merging:
|
32 |
-
|
33 |
-
```yaml
|
34 |
-
base_model: sethuiyer/Llamaverse-3.1-8B-Instruct
|
35 |
-
dtype: bfloat16
|
36 |
-
merge_method: ties
|
37 |
-
models:
|
38 |
-
- model: HumanLLMs/Human-Like-LLama3-8B-Instruct
|
39 |
-
parameters:
|
40 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
41 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
42 |
-
- model: DevsDoCode/LLama-3-8b-Uncensored
|
43 |
-
parameters:
|
44 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
45 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
46 |
-
- model: HiroseKoichi/Llama-3-8B-Stroganoff
|
47 |
-
parameters:
|
48 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
49 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
50 |
-
- model: DreadPoor/Zelus-8B-Model_Stock
|
51 |
-
parameters:
|
52 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
53 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
54 |
-
- model: CultriX/Llama3-8B-DPO
|
55 |
-
parameters:
|
56 |
-
density: [0.1, 0.3, 0.45, 0.3, 0.1]
|
57 |
-
weight: [0.15, 0.35, 0.6, 0.35, 0.15]
|
58 |
-
parameters:
|
59 |
-
int8_mask: true
|
60 |
-
```
|
61 |
-
|
62 |
### Usage
|
63 |
The following Python code demonstrates how to use Llamazing-3.1-8B-Instruct with the Divine Intellect preset:
|
64 |
|
|
|
14 |
---
|
15 |
# Llamazing-3.1-8B-Instruct
|
16 |
|
17 |
+
![img](./image.webp)
|
18 |
+
|
19 |
### Overview
|
20 |
Llamazing-3.1-8B-Instruct is a state-of-the-art language model resulting from the thoughtful merging of six specialized models. It balances reasoning, creativity, and conversational capabilities to deliver exceptional performance across various applications.
|
21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
### Usage
|
23 |
The following Python code demonstrates how to use Llamazing-3.1-8B-Instruct with the Divine Intellect preset:
|
24 |
|
README.md
CHANGED
@@ -14,51 +14,11 @@ tags:
|
|
14 |
---
|
15 |
# Llamazing-3.1-8B-Instruct
|
16 |
|
|
|
|
|
17 |
### Overview
|
18 |
Llamazing-3.1-8B-Instruct is a state-of-the-art language model resulting from the thoughtful merging of six specialized models. It balances reasoning, creativity, and conversational capabilities to deliver exceptional performance across various applications.
|
19 |
|
20 |
-
### Models Merged
|
21 |
-
The following models contributed to the creation of Llamazing-3.1-8B-Instruct:
|
22 |
-
|
23 |
-
1. [DreadPoor/Zelus-8B-Model_Stock](https://huggingface.co/DreadPoor/Zelus-8B-Model_Stock)
|
24 |
-
2. [HumanLLMs/Human-Like-LLama3-8B-Instruct](https://huggingface.co/HumanLLMs/Human-Like-LLama3-8B-Instruct)
|
25 |
-
3. [sethuiyer/Llamaverse-3.1-8B-Instruct](https://huggingface.co/sethuiyer/Llamaverse-3.1-8B-Instruct)
|
26 |
-
4. [CultriX/Llama3-8B-DPO](https://huggingface.co/CultriX/Llama3-8B-DPO)
|
27 |
-
5. [DevsDoCode/LLama-3-8b-Uncensored](https://huggingface.co/DevsDoCode/LLama-3-8b-Uncensored)
|
28 |
-
6. [HiroseKoichi/Llama-3-8B-Stroganoff](https://huggingface.co/HiroseKoichi/Llama-3-8B-Stroganoff)
|
29 |
-
|
30 |
-
### Configuration
|
31 |
-
Below is the YAML configuration used for merging:
|
32 |
-
|
33 |
-
```yaml
|
34 |
-
base_model: sethuiyer/Llamaverse-3.1-8B-Instruct
|
35 |
-
dtype: bfloat16
|
36 |
-
merge_method: ties
|
37 |
-
models:
|
38 |
-
- model: HumanLLMs/Human-Like-LLama3-8B-Instruct
|
39 |
-
parameters:
|
40 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
41 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
42 |
-
- model: DevsDoCode/LLama-3-8b-Uncensored
|
43 |
-
parameters:
|
44 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
45 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
46 |
-
- model: HiroseKoichi/Llama-3-8B-Stroganoff
|
47 |
-
parameters:
|
48 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
49 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
50 |
-
- model: DreadPoor/Zelus-8B-Model_Stock
|
51 |
-
parameters:
|
52 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
53 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
54 |
-
- model: CultriX/Llama3-8B-DPO
|
55 |
-
parameters:
|
56 |
-
density: [0.1, 0.3, 0.45, 0.3, 0.1]
|
57 |
-
weight: [0.15, 0.35, 0.6, 0.35, 0.15]
|
58 |
-
parameters:
|
59 |
-
int8_mask: true
|
60 |
-
```
|
61 |
-
|
62 |
### Usage
|
63 |
The following Python code demonstrates how to use Llamazing-3.1-8B-Instruct with the Divine Intellect preset:
|
64 |
|
|
|
14 |
---
|
15 |
# Llamazing-3.1-8B-Instruct
|
16 |
|
17 |
+
![img](./image.webp)
|
18 |
+
|
19 |
### Overview
|
20 |
Llamazing-3.1-8B-Instruct is a state-of-the-art language model resulting from the thoughtful merging of six specialized models. It balances reasoning, creativity, and conversational capabilities to deliver exceptional performance across various applications.
|
21 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
### Usage
|
23 |
The following Python code demonstrates how to use Llamazing-3.1-8B-Instruct with the Divine Intellect preset:
|
24 |
|
image.webp
ADDED
mergekit_config.yml
CHANGED
@@ -1,27 +1,10 @@
|
|
1 |
|
2 |
base_model: sethuiyer/Llamaverse-3.1-8B-Instruct
|
3 |
dtype: bfloat16
|
4 |
-
merge_method:
|
5 |
models:
|
6 |
-
- model:
|
7 |
-
|
8 |
-
|
9 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
10 |
-
- model: DevsDoCode/LLama-3-8b-Uncensored
|
11 |
-
parameters:
|
12 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
13 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
14 |
-
- model: HiroseKoichi/Llama-3-8B-Stroganoff
|
15 |
-
parameters:
|
16 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
17 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
18 |
-
- model: DreadPoor/Zelus-8B-Model_Stock
|
19 |
-
parameters:
|
20 |
-
density: [0.05, 0.15, 0.2, 0.15, 0.05]
|
21 |
-
weight: [0.05, 0.1, 0.2, 0.1, 0.05]
|
22 |
-
- model: CultriX/Llama3-8B-DPO
|
23 |
-
parameters:
|
24 |
-
density: [0.1, 0.3, 0.45, 0.3, 0.1]
|
25 |
-
weight: [0.15, 0.35, 0.6, 0.35, 0.15]
|
26 |
parameters:
|
27 |
int8_mask: true
|
|
|
1 |
|
2 |
base_model: sethuiyer/Llamaverse-3.1-8B-Instruct
|
3 |
dtype: bfloat16
|
4 |
+
merge_method: model_stock
|
5 |
models:
|
6 |
+
- model: ./merge
|
7 |
+
- model: ./merge_2
|
8 |
+
- model: ./merge_3
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
parameters:
|
10 |
int8_mask: true
|
model-00001-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3963695656
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:66550221793b4fcedf02ef146a5e970474be8244d11aad5f7165c6c5d39f6dec
|
3 |
size 3963695656
|
model-00002-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3926025448
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:a17771e244b4fb050671f6b397418ac433f72e9f4a909510e0c782005abda5b7
|
3 |
size 3926025448
|
model-00003-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3926025456
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d4fbe54647c65269671f863bbc411fcf23193cc8d42dabe2984fe25e3fe67431
|
3 |
size 3926025456
|
model-00004-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 3926025400
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d628e8ce48c34d513a2c89b7e3a21af5bf7f84e1f000c479d98d9746aaa64efc
|
3 |
size 3926025400
|
model-00005-of-00005.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 318784416
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:f4488d54be20a8d81f9eef9dc2fa82e8200212be1039d92958400e21593d5f18
|
3 |
size 318784416
|