|
--- |
|
base_model: |
|
- Azazelle/MN-Halide-12b-v1.0 |
|
- benhaotang/nemo-math-science-philosophy-12B |
|
- FallenMerick/MN-Chunky-Lotus-12B |
|
- FallenMerick/MN-Violet-Lotus-12B |
|
- GalrionSoftworks/Canidori-12B-v1 |
|
- GalrionSoftworks/Pleiades-12B-v1 |
|
- inflatebot/MN-12B-Mag-Mell-R1 |
|
- Nohobby/MN-12B-Siskin-v0.2 |
|
- ThijsL202/MadMix-Unleashed-12B |
|
- Trappu/Abomination-merge-attempt-12B |
|
- VongolaChouko/Starcannon-Unleashed-12B-v1.0 |
|
library_name: transformers |
|
tags: |
|
- mergekit |
|
- merge |
|
- bfloat16 |
|
- safetensors |
|
- 12b |
|
- chat |
|
- creative |
|
- roleplay |
|
- conversational |
|
- creative-writing |
|
- not-for-all-audiences |
|
language: |
|
- en |
|
- ru |
|
|
|
--- |
|
# AbominationScience-12B-v4 |
|
>When the choice is not random. |
|
|
|
![AbominationScienceLogo256.png](https://cdn-uploads.huggingface.co/production/uploads/673125091920e70ac26c8a2e/mrBCmxkidQ9KNQsRO_fOy.png) |
|
|
|
This is an interesting merge of **11 cool models**, created using [mergekit](https://github.com/arcee-ai/mergekit). |
|
Enjoy exploring :) |
|
|
|
## Merge Details |
|
### Method |
|
|
|
This model was merged using the multistep process and remerge with some model variations for best result. |
|
|
|
### Models |
|
|
|
The following models were included in the merge: |
|
|
|
* [Azazelle/MN-Halide-12b-v1.0](https://huggingface.co/Azazelle/MN-Halide-12b-v1.0) |
|
* [benhaotang/nemo-math-science-philosophy-12B](https://huggingface.co/benhaotang/nemo-math-science-philosophy-12B) |
|
* [FallenMerick/MN-Chunky-Lotus-12B](https://huggingface.co/FallenMerick/MN-Chunky-Lotus-12B) |
|
* [FallenMerick/MN-Violet-Lotus-12B](https://huggingface.co/FallenMerick/MN-Violet-Lotus-12B) |
|
* [GalrionSoftworks/Canidori-12B-v1](https://huggingface.co/GalrionSoftworks/Canidori-12B-v1) |
|
* [GalrionSoftworks/Pleiades-12B-v1](https://huggingface.co/GalrionSoftworks/Pleiades-12B-v1) |
|
* [inflatebot/MN-12B-Mag-Mell-R1](https://huggingface.co/inflatebot/MN-12B-Mag-Mell-R1) |
|
* [Nohobby/MN-12B-Siskin-v0.2](https://huggingface.co/Nohobby/MN-12B-Siskin-v0.2) |
|
* [ThijsL202/MadMix-Unleashed-12B](https://huggingface.co/ThijsL202/MadMix-Unleashed-12B) |
|
* [Trappu/Abomination-merge-attempt-12B](https://huggingface.co/Trappu/Abomination-merge-attempt-12B) |
|
* [VongolaChouko/Starcannon-Unleashed-12B-v1.0](https://huggingface.co/VongolaChouko/Starcannon-Unleashed-12B-v1.0) |
|
|
|
### Configuration |
|
|
|
The following YAML configurations was used to produce this model: |
|
|
|
```yaml |
|
# AbominationScience |
|
# It's a good model, I used it as a base for this merge. |
|
models: |
|
- model: Trappu/Abomination-merge-attempt-12B |
|
- model: benhaotang/nemo-math-science-philosophy-12B |
|
merge_method: slerp |
|
base_model: Trappu/Abomination-merge-attempt-12B |
|
dtype: bfloat16 |
|
parameters: |
|
t: [0.8, 0.2, 0.8, 0.2, 0.8, 0.2, 0.8] |
|
|
|
# SCUMCL |
|
models: |
|
- model: VongolaChouko/Starcannon-Unleashed-12B-v1.0 |
|
- model: FallenMerick/MN-Chunky-Lotus-12B |
|
merge_method: slerp |
|
base_model: VongolaChouko/Starcannon-Unleashed-12B-v1.0 |
|
dtype: bfloat16 |
|
parameters: |
|
t: [0.7, 0.3, 0.7, 0.3, 0.7, 0.3, 0.7] |
|
|
|
# SISMMU |
|
models: |
|
- model: Nohobby/MN-12B-Siskin-v0.2 |
|
- model: ThijsL202/MadMix-Unleashed-12B |
|
merge_method: slerp |
|
base_model: Nohobby/MN-12B-Siskin-v0.2 |
|
dtype: bfloat16 |
|
parameters: |
|
t: [0, 0.5, 1, 0.5, 0] |
|
|
|
# PLECAD |
|
models: |
|
- model: GalrionSoftworks/Pleiades-12B-v1 |
|
- model: GalrionSoftworks/Canidori-12B-v1 |
|
merge_method: slerp |
|
base_model: GalrionSoftworks/Pleiades-12B-v1 |
|
dtype: bfloat16 |
|
parameters: |
|
t: [0.7, 0.3, 0.7, 0.3, 0.7, 0.3, 0.7] |
|
|
|
# Positive-12B-v1 and Negative-12B-v1 are the basis of diversity for the base model. |
|
# I've lost the exact config, but it was most likely a slerp like the one in SCUMCL/SISMMU/PLECAD. |
|
# Positive-12B-v1 = SCUMCL + SISMMU. |
|
# Negative-12B-v1 = PLECAD + AbominationScience. |
|
|
|
# AbominationScience-12B-v2 |
|
models: |
|
- model: F:/Positive-12B-v1 |
|
parameters: |
|
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5] |
|
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5] |
|
- model: F:/Negative-12B-v1 |
|
parameters: |
|
density: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5] |
|
weight: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.1, 0.9, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5] |
|
merge_method: dare_ties |
|
base_model: F:/AbominationScience |
|
dtype: bfloat16 |
|
|
|
# AbominationScience-12B-v3 |
|
# Della merge with a good base to form an interesting core |
|
models: |
|
- model: F:/AbominationScience |
|
parameters: |
|
weight: [0.5, 0.6, 0.4, 0.7, 0.3, 0.8, 0.2, 0.8, 0.2, 0.7, 0.3, 0.6, 0.4, 0.5] |
|
density: [0.5, 0.4, 0.6, 0.3, 0.7, 0.2, 0.8, 0.2, 0.8, 0.3, 0.7, 0.4, 0.6, 0.5] |
|
merge_method: della |
|
parameters: |
|
epsilon: 0.123456789 |
|
lambda: 0.987654321 |
|
base_model: F:/AbominationScience-12B-v2 |
|
dtype: bfloat16 |
|
|
|
# AbominationScience-12B-v4 |
|
# Final shift the model to three very good bases. |
|
models: |
|
- model: inflatebot/MN-12B-Mag-Mell-R1 |
|
- model: FallenMerick/MN-Violet-Lotus-12B |
|
- model: Azazelle/MN-Halide-12b-v1.0 |
|
merge_method: model_stock |
|
base_model: F:/AbominationScience-12B-v3 |
|
dtype: bfloat16 |
|
``` |
|
|
|
>My thanks to the authors of the original models, your work is incredible. Have a good time 🖤 |
|
|