--- base_model: - Severian/Nexus-IKM-Mistral-Instruct-v0.2-7B - son-of-man/HoloViolet-7B-test3 - alpindale/Mistral-7B-v0.2-hf - localfultonextractor/Erosumika-7B-v3 library_name: transformers tags: - mergekit - merge --- # merged This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [task arithmetic](https://arxiv.org/abs/2212.04089) merge method using [alpindale/Mistral-7B-v0.2-hf](https://huggingface.co/alpindale/Mistral-7B-v0.2-hf) as a base. ### Models Merged The following models were included in the merge: * [Severian/Nexus-IKM-Mistral-Instruct-v0.2-7B](https://huggingface.co/Severian/Nexus-IKM-Mistral-Instruct-v0.2-7B) * [son-of-man/HoloViolet-7B-test3](https://huggingface.co/son-of-man/HoloViolet-7B-test3) * [localfultonextractor/Erosumika-7B-v3](https://huggingface.co/localfultonextractor/Erosumika-7B-v3) ### Configuration The following YAML configuration was used to produce this model: ```yaml base_model: model: path: alpindale/Mistral-7B-v0.2-hf dtype: bfloat16 merge_method: task_arithmetic slices: - sources: - layer_range: [0, 32] model: model: path: alpindale/Mistral-7B-v0.2-hf parameters: weight: 0.3 - layer_range: [0, 32] model: model: path: son-of-man/HoloViolet-7B-test3 parameters: weight: 0.2 - layer_range: [0, 32] model: model: path: localfultonextractor/Erosumika-7B-v3 parameters: weight: 0.3 - layer_range: [0, 32] model: model: path: Severian/Nexus-IKM-Mistral-Instruct-v0.2-7B parameters: weight: 0.2 ```