--- base_model: - meta-llama/Meta-Llama-3-8B-Instruct - NousResearch/Hermes-2-Pro-Llama-3-8B - aaditya/Llama3-OpenBioLLM-8B library_name: transformers tags: - mergekit - merge license: llama3 --- # Llama3-merge-biomed-8b This is a DARE-TIES Merge of Llama3-8b-Instruct + NousResearch/Hermes-2-Pro-Llama-3-8B + aaditya/Llama3-OpenBioLLM-8B ## Leaderboard metrics | Task | Metric | Llama3-merge-biomed-8b (%) | Llama3-8B-Inst (%) | Llama3-OpenBioLLM-8B (%) | |--------------------------------------|--------------------------|------------------|------------|-------------| | **ARC Challenge** | Accuracy | 59.39 | 57.17 | 55.38 | | | Normalized Accuracy | 63.65 | 60.75 | 58.62 | | **Hellaswag** | Accuracy | 62.59 | 59.04 | 61.83 | | | Normalized Accuracy | 81.53 | 78.55 | 80.76 | | **Winogrande** | Accuracy | 75.93 | 74.51 | 70.88 | | **GSM8K** | Accuracy | 59.36 | 68.69 | 10.16 | | **HendrycksTest-Average** | Accuracy | 67.85 | 67.07 | 64.40 | | | Normalized Accuracy | 67.85 | 67.07 | 64.40 | | **HendrycksTest-Anatomy** | Accuracy | 72.59 | 65.19 | 56.30 | | **HendrycksTest-Clinical Knowledge** | Accuracy | 77.83 | 74.72 | 60.38 | | **HendrycksTest-College Biology** | Accuracy | 79.86 | 79.86 | 79.86 | | **HendrycksTest-College Medicine** | Accuracy | 70.81 | 63.58 | 62.28 | | **HendrycksTest-Medical Genetics** | Accuracy | 84.00 | 80.00 | 76.00 | | **HendrycksTest-Professional Medicine** | Accuracy | 71.69 | 71.69 | 69.41 | This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit). ## Merge Details ### Merge Method This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) as a base. ### Models Merged The following models were included in the merge: * [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B) * [aaditya/Llama3-OpenBioLLM-8B](https://huggingface.co/aaditya/Llama3-OpenBioLLM-8B) ### Configuration The following YAML configuration was used to produce this model: ```yaml models: - model: meta-llama/Meta-Llama-3-8B-Instruct # Base model providing a general foundation without specific parameters - model: meta-llama/Meta-Llama-3-8B-Instruct parameters: density: 0.60 weight: 0.5 - model: NousResearch/Hermes-2-Pro-Llama-3-8B parameters: density: 0.55 weight: 0.1 - model: aaditya/Llama3-OpenBioLLM-8B parameters: density: 0.55 weight: 0.4 merge_method: dare_ties base_model: meta-llama/Meta-Llama-3-8B-Instruct parameters: int8_mask: true dtype: bfloat16 ```