--- license: apache-2.0 tags: - merge - mergekit - mlabonne/AlphaMonarch-7B - mlabonne/NeuralMonarch-7B - bardsai/jaskier-7b-dpo-v5.6 --- # neurotic-crown-clown-7B-ties ![neurotic-crown-clown-image](./neurotic-crown-clown.png) neurotic-crown-clown-7B-ties is a TRIM, ELECT SIGN & MERGE (TIES) merge of the following models using [mergekit](https://github.com/cg123/mergekit): * [mlabonne/NeuralMonarch-7B](https://huggingface.co/mlabonne/NeuralMonarch-7B) * [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B) * [bardsai/jaskier-7b-dpo-v5.6](https://huggingface.co/bardsai/jaskier-7b-dpo-v5.6) See the paper [TIES-Merging: Resolving Interference When Merging Models](https://arxiv.org/abs/2306.01708) for more on the method. ## 🧩 Configuration ```yaml models: - model: mlabonne/NeuralMonarch-7B # no parameters necessary for base model - model: mlabonne/AlphaMonarch-7B parameters: density: 0.5 weight: 0.5 - model: bardsai/jaskier-7b-dpo-v5.6 parameters: density: 0.5 weight: 0.3 merge_method: ties base_model: mlabonne/NeuralMonarch-7B parameters: normalize: true dtype: float16 ```