lighteternal's picture
Update README.md
3b97d1f verified
---
base_model:
- meta-llama/Meta-Llama-3-8B-Instruct
- NousResearch/Hermes-2-Pro-Llama-3-8B
- aaditya/Llama3-OpenBioLLM-8B
library_name: transformers
tags:
- mergekit
- merge
license: llama3
---
# Llama3-merge-biomed-8b
This is a DARE-TIES Merge of Llama3-8b-Instruct + NousResearch/Hermes-2-Pro-Llama-3-8B + aaditya/Llama3-OpenBioLLM-8B.
It is a simple experiment to assess whether combining models with strengths in general language understanding and biomedical knowledge can enhance performance on specialized tasks without compromising general applicability.
The results indicate promising outcomes in areas like HendrycksTest tasks related to Biology and Medicine, as well as improvements in complex reasoning as seen in the ARC Challenge and Winogrande benchmarks.
## Usage
I recommend using the prompt template of Llama3: https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3/
## Leaderboard metrics according to 🤗 Open LLM Leaderboard
| Task | Metric | Ours (%) | Llama38BInstr. (%) |OpenBioLLM8B (%) |
|--------------------------------------|--------------------------|------------------|------------|-------------|
| **ARC Challenge** | Accuracy | **59.39** | 57.17 | 55.38 |
| | Normalized Accuracy | **63.65** | 60.75 | 58.62 |
| **Hellaswag** | Accuracy | **62.59** | 59.04 | 61.83 |
| | Normalized Accuracy | **81.53** | 78.55 | 80.76 |
| **Winogrande** | Accuracy | **75.93** | 74.51 | 70.88 |
| **GSM8K** | Accuracy | 59.36 | **68.69** | 10.15 |
| **HendrycksTest-Anatomy** | Accuracy | **72.59** | 65.19 | 69.62 |
| **HendrycksTest-Clinical Knowledge** | Accuracy | **77.83** | 74.72 | 60.38 |
| **HendrycksTest-College Biology** | Accuracy | **81.94** | 79.86 | 79.86 |
| **HendrycksTest-College Medicine** | Accuracy | 69.36 | 63.58 | **70.52** |
| **HendrycksTest-Medical Genetics** | Accuracy | **86.00** | 80.00 | 80.00 |
| **HendrycksTest-Professional Medicine** | Accuracy | **77.94** | 71.69 | 77.94 |
This is a merge of pre-trained language models created using [mergekit](https://github.com/cg123/mergekit).
## Merge Details
### Merge Method
This model was merged using the [DARE](https://arxiv.org/abs/2311.03099) [TIES](https://arxiv.org/abs/2306.01708) merge method using [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) as a base.
### Models Merged
The following models were included in the merge:
* [NousResearch/Hermes-2-Pro-Llama-3-8B](https://huggingface.co/NousResearch/Hermes-2-Pro-Llama-3-8B)
* [aaditya/Llama3-OpenBioLLM-8B](https://huggingface.co/aaditya/Llama3-OpenBioLLM-8B)
### Configuration
The following YAML configuration was used to produce this model:
```yaml
models:
- model: meta-llama/Meta-Llama-3-8B-Instruct
# Base model providing a general foundation without specific parameters
- model: meta-llama/Meta-Llama-3-8B-Instruct
parameters:
density: 0.60
weight: 0.5
- model: NousResearch/Hermes-2-Pro-Llama-3-8B
parameters:
density: 0.55
weight: 0.1
- model: aaditya/Llama3-OpenBioLLM-8B
parameters:
density: 0.55
weight: 0.4
merge_method: dare_ties
base_model: meta-llama/Meta-Llama-3-8B-Instruct
parameters:
int8_mask: true
dtype: bfloat16
```