weezywitasneezy commited on
Commit
54e0fde
1 Parent(s): 29b30af

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +16 -12
README.md CHANGED
@@ -116,10 +116,26 @@ model-index:
116
 
117
  # BenchmarkEngineering-7B-slerp
118
 
 
 
119
  BenchmarkEngineering-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
120
  * [paulml/OmniBeagleSquaredMBX-v3-7B](https://huggingface.co/paulml/OmniBeagleSquaredMBX-v3-7B)
121
  * [automerger/YamshadowExperiment28-7B](https://huggingface.co/automerger/YamshadowExperiment28-7B)
122
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
123
  ## 🧩 Configuration
124
 
125
  ```yaml
@@ -165,16 +181,4 @@ pipeline = transformers.pipeline(
165
  outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
166
  print(outputs[0]["generated_text"])
167
  ```
168
- # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
169
- Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_weezywitasneezy__BenchmarkEngineering-7B-slerp)
170
-
171
- | Metric |Value|
172
- |---------------------------------|----:|
173
- |Avg. |76.40|
174
- |AI2 Reasoning Challenge (25-Shot)|74.15|
175
- |HellaSwag (10-Shot) |89.09|
176
- |MMLU (5-Shot) |64.69|
177
- |TruthfulQA (0-shot) |75.93|
178
- |Winogrande (5-shot) |85.32|
179
- |GSM8k (5-shot) |69.22|
180
 
 
116
 
117
  # BenchmarkEngineering-7B-slerp
118
 
119
+ This model was merged with the intent of producing excellent Open-LLM benchmarking results but smashing two of the highest performant models in their class together
120
+
121
  BenchmarkEngineering-7B-slerp is a merge of the following models using [LazyMergekit](https://colab.research.google.com/drive/1obulZ1ROXHjYLn6PPZJwRR6GzgQogxxb?usp=sharing):
122
  * [paulml/OmniBeagleSquaredMBX-v3-7B](https://huggingface.co/paulml/OmniBeagleSquaredMBX-v3-7B)
123
  * [automerger/YamshadowExperiment28-7B](https://huggingface.co/automerger/YamshadowExperiment28-7B)
124
 
125
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
126
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_weezywitasneezy__BenchmarkEngineering-7B-slerp)
127
+
128
+ | Metric |Value|
129
+ |---------------------------------|----:|
130
+ |Avg. |76.40|
131
+ |AI2 Reasoning Challenge (25-Shot)|74.15|
132
+ |HellaSwag (10-Shot) |89.09|
133
+ |MMLU (5-Shot) |64.69|
134
+ |TruthfulQA (0-shot) |75.93|
135
+ |Winogrande (5-shot) |85.32|
136
+ |GSM8k (5-shot) |69.22|
137
+
138
+
139
  ## 🧩 Configuration
140
 
141
  ```yaml
 
181
  outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
182
  print(outputs[0]["generated_text"])
183
  ```
 
 
 
 
 
 
 
 
 
 
 
 
184