lighteternal commited on
Commit
c09ab32
1 Parent(s): ea87f9f

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +9 -3
README.md CHANGED
@@ -12,11 +12,17 @@ license: llama3
12
  ---
13
  # Llama3-merge-biomed-8b
14
 
15
- This is a DARE-TIES Merge of Llama3-8b-Instruct + NousResearch/Hermes-2-Pro-Llama-3-8B + aaditya/Llama3-OpenBioLLM-8B
 
 
16
 
17
- ## Leaderboard metrics
18
 
19
- | Task | Metric | Llama3-merge-biomed-8b (%) | Llama3-8B-Inst (%) | Llama3-OpenBioLLM-8B (%) |
 
 
 
 
20
  |--------------------------------------|--------------------------|------------------|------------|-------------|
21
  | **ARC Challenge** | Accuracy | 59.39 | 57.17 | 55.38 |
22
  | | Normalized Accuracy | 63.65 | 60.75 | 58.62 |
 
12
  ---
13
  # Llama3-merge-biomed-8b
14
 
15
+ This is a DARE-TIES Merge of Llama3-8b-Instruct + NousResearch/Hermes-2-Pro-Llama-3-8B + aaditya/Llama3-OpenBioLLM-8B.
16
+ It is a simple experiment to assess whether combining models with strengths in general language understanding and biomedical knowledge can enhance performance on specialized tasks without compromising general applicability.
17
+ The results indicate promising outcomes in areas like HendrycksTest tasks related to Biology and Medicine, as well as improvements in complex reasoning as seen in the ARC Challenge and Winogrande benchmarks.
18
 
19
+ ## Usage
20
 
21
+ I recommend using the prompt template of Llama3: https://llama.meta.com/docs/model-cards-and-prompt-formats/meta-llama-3/
22
+
23
+ ## Leaderboard metrics according to 🤗 Open LLM Leaderboard
24
+
25
+ | Task | Metric | Ours (%) | Llama38BInstr. (%) |OpenBioLLM8B (%) |
26
  |--------------------------------------|--------------------------|------------------|------------|-------------|
27
  | **ARC Challenge** | Accuracy | 59.39 | 57.17 | 55.38 |
28
  | | Normalized Accuracy | 63.65 | 60.75 | 58.62 |