abideen commited on
Commit
d631c24
1 Parent(s): 46761b6

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +13 -14
README.md CHANGED
@@ -127,23 +127,22 @@ MonarchCoder-MoE-2x7B is a Mixure of Experts (MoE) made with the following model
127
  * [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B)
128
  * [Syed-Hasan-8503/Tess-Coder-7B-Mistral-v1.0](https://huggingface.co/Syed-Hasan-8503/Tess-Coder-7B-Mistral-v1.0)
129
 
130
- The main aim behind creating this model is to create a model that performs well in reasoning, conversation, and coding. AlphaMonarch pperforms amazing on reasoning and conversation tasks. Merging AlphaMonarch with a coding model yielded MonarchCoder-2x7B which performs better on OpenLLM, Nous, and HumanEval benchmark.
131
 
132
  ## 🏆 Evaluation results
133
-
134
  ```
135
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
136
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_abideen__MonarchCoder-MoE-2x7B)
137
-
138
- | Metric |Value|
139
- |---------------------------------|----:|
140
- |Avg. |74.23|
141
- |AI2 Reasoning Challenge (25-Shot)|70.99|
142
- |HellaSwag (10-Shot) |87.99|
143
- |MMLU (5-Shot) |65.11|
144
- |TruthfulQA (0-shot) |71.25|
145
- |Winogrande (5-shot) |80.66|
146
- |GSM8k (5-shot) |69.37|
147
  ```
148
 
149
  ## 🧩 Configuration
 
127
  * [mlabonne/AlphaMonarch-7B](https://huggingface.co/mlabonne/AlphaMonarch-7B)
128
  * [Syed-Hasan-8503/Tess-Coder-7B-Mistral-v1.0](https://huggingface.co/Syed-Hasan-8503/Tess-Coder-7B-Mistral-v1.0)
129
 
130
+ The main aim behind creating this model is to create a model that performs well in reasoning, conversation, and coding. AlphaMonarch performs amazing on reasoning and conversation tasks. Merging AlphaMonarch with a coding model yielded MonarchCoder-2x7B which performs better on OpenLLM, Nous, and HumanEval benchmark.
131
 
132
  ## 🏆 Evaluation results
 
133
  ```
134
+ | Metric |MonarchCoder-Moe-2x7B||MonarchCoder-7B||AlphaMonarch|
135
+ |---------------------------------|---------------------|-----------------|------------|
136
+ |Avg. | 74.23 | 71.17 | 75.99 |
137
+ |HumanEval | 41.15 | 39.02 | 34.14 |
138
+ |HumanEval+ | 29.87 | 31.70 | 29.26 |
139
+ |MBPP | 40.60 | * | * |
140
+ |AI2 Reasoning Challenge (25-Shot)| 70.99 | 68.52 | 73.04 |
141
+ |HellaSwag (10-Shot) | 87.99 | 87.30 | 89.18 |
142
+ |MMLU (5-Shot) | 65.11 | 64.65 | 64.40 |
143
+ |TruthfulQA (0-shot) | 71.25 | 61.21 | 77.91 |
144
+ |Winogrande (5-shot) | 80.66 | 80.19 .| 84.69 |
145
+ |GSM8k (5-shot) . | 69.37 | 65.13 | 66.72 |
146
  ```
147
 
148
  ## 🧩 Configuration