lighteternal
commited on
Commit
•
c09ab32
1
Parent(s):
ea87f9f
Update README.md
Browse files
README.md
CHANGED
@@ -12,11 +12,17 @@ license: llama3
|
|
12 |
---
|
13 |
# Llama3-merge-biomed-8b
|
14 |
|
15 |
-
This is a DARE-TIES Merge of Llama3-8b-Instruct + NousResearch/Hermes-2-Pro-Llama-3-8B + aaditya/Llama3-OpenBioLLM-8B
|
|
|
|
|
16 |
|
17 |
-
##
|
18 |
|
19 |
-
|
|
|
|
|
|
|
|
|
20 |
|--------------------------------------|--------------------------|------------------|------------|-------------|
|
21 |
| **ARC Challenge** | Accuracy | 59.39 | 57.17 | 55.38 |
|
22 |
| | Normalized Accuracy | 63.65 | 60.75 | 58.62 |
|
|
|
12 |
---
|
13 |
# Llama3-merge-biomed-8b
|
14 |
|
15 |
+
This is a DARE-TIES Merge of Llama3-8b-Instruct + NousResearch/Hermes-2-Pro-Llama-3-8B + aaditya/Llama3-OpenBioLLM-8B.
|
16 |
+
It is a simple experiment to assess whether combining models with strengths in general language understanding and biomedical knowledge can enhance performance on specialized tasks without compromising general applicability.
|
17 |
+
The results indicate promising outcomes in areas like HendrycksTest tasks related to Biology and Medicine, as well as improvements in complex reasoning as seen in the ARC Challenge and Winogrande benchmarks.
|
18 |
|
19 |
+
## Usage
|
20 |
|
21 |
+
I recommend using the prompt template of Llama3: https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3/
|
22 |
+
|
23 |
+
## Leaderboard metrics according to 🤗 Open LLM Leaderboard
|
24 |
+
|
25 |
+
| Task | Metric | Ours (%) | Llama38BInstr. (%) |OpenBioLLM8B (%) |
|
26 |
|--------------------------------------|--------------------------|------------------|------------|-------------|
|
27 |
| **ARC Challenge** | Accuracy | 59.39 | 57.17 | 55.38 |
|
28 |
| | Normalized Accuracy | 63.65 | 60.75 | 58.62 |
|