--- base_model: - elinas/Chronos-Gold-12B-1.0 - nbeerbower/Lyra-Gutenberg-mistral-nemo-12B - cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b - anthracite-org/magnum-v2.5-12b-kto library_name: transformers tags: - mergekit - merge --- # merge This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [anthracite-org/magnum-v2.5-12b-kto](https://huggingface.co/anthracite-org/magnum-v2.5-12b-kto) as a base. ### Models Merged The following models were included in the merge: * [elinas/Chronos-Gold-12B-1.0](https://huggingface.co/elinas/Chronos-Gold-12B-1.0) * [nbeerbower/Lyra-Gutenberg-mistral-nemo-12B](https://huggingface.co/nbeerbower/Lyra-Gutenberg-mistral-nemo-12B) * [cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b](https://huggingface.co/cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: anthracite-org/magnum-v2.5-12b-kto - model: nbeerbower/Lyra-Gutenberg-mistral-nemo-12B parameters: density: 0.4 weight: 0.3 - model: cognitivecomputations/dolphin-2.9.3-mistral-nemo-12b parameters: density: 0.6 weight: 0.4 - model: elinas/Chronos-Gold-12B-1.0 parameters: density: 0.5 weight: 0.3 merge_method: dare_ties base_model: anthracite-org/magnum-v2.5-12b-kto parameters: normalize: false int8_mask: true dtype: float16 ```