--- base_model: - nayohan/llama3-8b-it-translation-general-en-ko-1sent - MaziyarPanahi/Llama-3-8B-Instruct-DPO-v0.3 - cognitivecomputations/dolphin-2.9-llama3-8b - NousResearch/Hermes-2-Pro-Llama-3-8B - winglian/llama-3-8b-1m-PoSE - asiansoul/Llama-3-Open-Ko-Linear-8B - NousResearch/Meta-Llama-3-8B - Danielbrdz/Barcenas-Llama3-8b-ORPO - NousResearch/Meta-Llama-3-8B-Instruct library_name: transformers tags: - mergekit - merge --- # Versatile-Llama-3-8B I'm not going to say that this Merge model is the best model ever made. I'm not going to tell you that you'll enjoy chatting with my merge model. All I want to say is thank you for taking time out of your day to visit. Without users like you, my work would be meaningless. Not test this model, just based on brainstorming, release this. ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [NousResearch/Meta-Llama-3-8B](https://huggingface.co/NousResearch/Meta-Llama-3-8B) as a base. ### Models Merged The following models were included in the merge: * [nayohan/llama3-8b-it-translation-general-en-ko-1sent](https://huggingface.co/nayohan/llama3-8b-it-translation-general-en-ko-1sent) * [MaziyarPanahi/Llama-3-8B-Instruct-DPO-v0.3](https://huggingface.co/MaziyarPanahi/Llama-3-8B-Instruct-DPO-v0.3) * [cognitivecomputations/dolphin-2.9-llama3-8b](https://huggingface.co/cognitivecomputations/dolphin-2.9-llama3-8b) * [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B) * [winglian/llama-3-8b-1m-PoSE](https://huggingface.co/winglian/llama-3-8b-1m-PoSE) * [asiansoul/Llama-3-Open-Ko-Linear-8B](https://huggingface.co/asiansoul/Llama-3-Open-Ko-Linear-8B) * [Danielbrdz/Barcenas-Llama3-8b-ORPO](https://huggingface.co/Danielbrdz/Barcenas-Llama3-8b-ORPO) * [NousResearch/Meta-Llama-3-8B-Instruct](https://huggingface.co/NousResearch/Meta-Llama-3-8B-Instruct) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: NousResearch/Meta-Llama-3-8B # Base model providing a general foundation without specific parameters - model: NousResearch/Meta-Llama-3-8B-Instruct parameters: density: 0.60 weight: 0.25 - model: winglian/llama-3-8b-1m-PoSE parameters: density: 0.55 weight: 0.15 - model: MaziyarPanahi/Llama-3-8B-Instruct-DPO-v0.3 parameters: density: 0.55 weight: 0.15 - model: asiansoul/Llama-3-Open-Ko-Linear-8B parameters: density: 0.55 weight: 0.2 - model: nayohan/llama3-8b-it-translation-general-en-ko-1sent parameters: density: 0.55 weight: 0.1 - model: cognitivecomputations/dolphin-2.9-llama3-8b parameters: density: 0.55 weight: 0.1 - model: Danielbrdz/Barcenas-Llama3-8b-ORPO parameters: density: 0.55 weight: 0.05 - model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 0.55 weight: 0.1 merge_method: dare_ties base_model: NousResearch/Meta-Llama-3-8B parameters: int8_mask: true dtype: bfloat16 ```