--- base_model: - Open-Orca/Mistral-7B-SlimOrca - cognitivecomputations/samantha-1.1-westlake-7b-laser - cognitivecomputations/dolphin-2.2.1-mistral-7b - Nexusflow/Starling-LM-7B-beta - cognitivecomputations/openchat-3.5-0106-laser library_name: transformers tags: - mergekit - merge --- # mistral-dareties-v92 This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [Open-Orca/Mistral-7B-SlimOrca](https://huggingface.co/Open-Orca/Mistral-7B-SlimOrca) as a base. ### Models Merged The following models were included in the merge: * [cognitivecomputations/samantha-1.1-westlake-7b-laser](https://huggingface.co/cognitivecomputations/samantha-1.1-westlake-7b-laser) * [cognitivecomputations/dolphin-2.2.1-mistral-7b](https://huggingface.co/cognitivecomputations/dolphin-2.2.1-mistral-7b) * [Nexusflow/Starling-LM-7B-beta](https://huggingface.co/Nexusflow/Starling-LM-7B-beta) * [cognitivecomputations/openchat-3.5-0106-laser](https://huggingface.co/cognitivecomputations/openchat-3.5-0106-laser) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: Open-Orca/Mistral-7B-SlimOrca dtype: bfloat16 merge_method: dare_ties models: - model: cognitivecomputations/openchat-3.5-0106-laser parameters: density: 0.74 weight: [0, 0.2, 0.4] - model: Nexusflow/Starling-LM-7B-beta parameters: density: 0.74 weight: [0.3, 0.4] - model: cognitivecomputations/samantha-1.1-westlake-7b-laser parameters: density: 0.74 weight: [0.3, 0.4] - model: cognitivecomputations/dolphin-2.2.1-mistral-7b parameters: density: 0.74 weight: [0.3, 0.4] parameters: int8_mask: true normalize: true ```