aashish1904 commited on
Commit
2e813da
β€’
1 Parent(s): a15cdf6

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +173 -18
README.md CHANGED
@@ -1,4 +1,6 @@
 
1
  ---
 
2
  language:
3
  - en
4
  license: other
@@ -17,7 +19,7 @@ tags:
17
  - math
18
  - llama
19
  - llama3
20
- base_model: Weyaxi/Einstein-v6.1-Llama3-8B
21
  datasets:
22
  - allenai/ai2_arc
23
  - camel-ai/physics
@@ -75,8 +77,7 @@ model-index:
75
  value: 62.46
76
  name: normalized accuracy
77
  source:
78
- url: >-
79
- https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
80
  name: Open LLM Leaderboard
81
  - task:
82
  type: text-generation
@@ -92,8 +93,7 @@ model-index:
92
  value: 82.41
93
  name: normalized accuracy
94
  source:
95
- url: >-
96
- https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
97
  name: Open LLM Leaderboard
98
  - task:
99
  type: text-generation
@@ -110,8 +110,7 @@ model-index:
110
  value: 66.19
111
  name: accuracy
112
  source:
113
- url: >-
114
- https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
115
  name: Open LLM Leaderboard
116
  - task:
117
  type: text-generation
@@ -127,8 +126,7 @@ model-index:
127
  - type: mc2
128
  value: 55.1
129
  source:
130
- url: >-
131
- https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
132
  name: Open LLM Leaderboard
133
  - task:
134
  type: text-generation
@@ -145,8 +143,7 @@ model-index:
145
  value: 79.32
146
  name: accuracy
147
  source:
148
- url: >-
149
- https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
150
  name: Open LLM Leaderboard
151
  - task:
152
  type: text-generation
@@ -163,17 +160,114 @@ model-index:
163
  value: 66.11
164
  name: accuracy
165
  source:
166
- url: >-
167
- https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
168
  name: Open LLM Leaderboard
169
- pipeline_tag: text-generation
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
170
  ---
171
 
172
- # πŸ”¬ Einstein-v6.1-Llama3-8B - GGUF
 
 
 
 
173
 
174
- - This is quantized version of [Weyaxi/Einstein-v6.1-Llama3-8B](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) created using llama.cpp
175
 
176
- # Model Description
 
 
177
 
178
  This model is a full fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on diverse datasets.
179
 
@@ -372,6 +466,19 @@ The results of this filtering process and its outcomes are in the data folder of
372
 
373
  [Weyaxi/Einstein-v6.1-Llama3-8B/data](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B/tree/main/data)
374
 
 
 
 
 
 
 
 
 
 
 
 
 
 
375
 
376
  # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
377
  Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v6.1-Llama3-8B)
@@ -386,6 +493,38 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
386
  |Winogrande (5-shot) |79.32|
387
  |GSM8k (5-shot) |66.11|
388
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
389
  # πŸ€– Additional information about training
390
 
391
  This model is full fine-tuned for 2 epoch.
@@ -396,4 +535,20 @@ Total number of steps was 2026.
396
 
397
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/Ycs7ZpoqmxFt0u9rybCO1.png)
398
 
399
- </details><br>
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+
2
  ---
3
+
4
  language:
5
  - en
6
  license: other
 
19
  - math
20
  - llama
21
  - llama3
22
+ base_model: meta-llama/Meta-Llama-3-8B
23
  datasets:
24
  - allenai/ai2_arc
25
  - camel-ai/physics
 
77
  value: 62.46
78
  name: normalized accuracy
79
  source:
80
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
 
81
  name: Open LLM Leaderboard
82
  - task:
83
  type: text-generation
 
93
  value: 82.41
94
  name: normalized accuracy
95
  source:
96
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
 
97
  name: Open LLM Leaderboard
98
  - task:
99
  type: text-generation
 
110
  value: 66.19
111
  name: accuracy
112
  source:
113
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
 
114
  name: Open LLM Leaderboard
115
  - task:
116
  type: text-generation
 
126
  - type: mc2
127
  value: 55.1
128
  source:
129
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
 
130
  name: Open LLM Leaderboard
131
  - task:
132
  type: text-generation
 
143
  value: 79.32
144
  name: accuracy
145
  source:
146
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
 
147
  name: Open LLM Leaderboard
148
  - task:
149
  type: text-generation
 
160
  value: 66.11
161
  name: accuracy
162
  source:
163
+ url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
164
+ name: Open LLM Leaderboard
165
+ - task:
166
+ type: text-generation
167
+ name: Text Generation
168
+ dataset:
169
+ name: IFEval (0-Shot)
170
+ type: HuggingFaceH4/ifeval
171
+ args:
172
+ num_few_shot: 0
173
+ metrics:
174
+ - type: inst_level_strict_acc and prompt_level_strict_acc
175
+ value: 45.68
176
+ name: strict accuracy
177
+ source:
178
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
179
+ name: Open LLM Leaderboard
180
+ - task:
181
+ type: text-generation
182
+ name: Text Generation
183
+ dataset:
184
+ name: BBH (3-Shot)
185
+ type: BBH
186
+ args:
187
+ num_few_shot: 3
188
+ metrics:
189
+ - type: acc_norm
190
+ value: 29.38
191
+ name: normalized accuracy
192
+ source:
193
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
194
+ name: Open LLM Leaderboard
195
+ - task:
196
+ type: text-generation
197
+ name: Text Generation
198
+ dataset:
199
+ name: MATH Lvl 5 (4-Shot)
200
+ type: hendrycks/competition_math
201
+ args:
202
+ num_few_shot: 4
203
+ metrics:
204
+ - type: exact_match
205
+ value: 5.74
206
+ name: exact match
207
+ source:
208
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
209
+ name: Open LLM Leaderboard
210
+ - task:
211
+ type: text-generation
212
+ name: Text Generation
213
+ dataset:
214
+ name: GPQA (0-shot)
215
+ type: Idavidrein/gpqa
216
+ args:
217
+ num_few_shot: 0
218
+ metrics:
219
+ - type: acc_norm
220
+ value: 4.25
221
+ name: acc_norm
222
+ source:
223
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
224
  name: Open LLM Leaderboard
225
+ - task:
226
+ type: text-generation
227
+ name: Text Generation
228
+ dataset:
229
+ name: MuSR (0-shot)
230
+ type: TAUR-Lab/MuSR
231
+ args:
232
+ num_few_shot: 0
233
+ metrics:
234
+ - type: acc_norm
235
+ value: 11.23
236
+ name: acc_norm
237
+ source:
238
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
239
+ name: Open LLM Leaderboard
240
+ - task:
241
+ type: text-generation
242
+ name: Text Generation
243
+ dataset:
244
+ name: MMLU-PRO (5-shot)
245
+ type: TIGER-Lab/MMLU-Pro
246
+ config: main
247
+ split: test
248
+ args:
249
+ num_few_shot: 5
250
+ metrics:
251
+ - type: acc
252
+ value: 23.68
253
+ name: accuracy
254
+ source:
255
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=Weyaxi/Einstein-v6.1-Llama3-8B
256
+ name: Open LLM Leaderboard
257
+
258
  ---
259
 
260
+ [![QuantFactory Banner](https://lh7-rt.googleusercontent.com/docsz/AD_4nXeiuCm7c8lEwEJuRey9kiVZsRn2W-b4pWlu3-X534V3YmVuVc2ZL-NXg2RkzSOOS2JXGHutDuyyNAUtdJI65jGTo8jT9Y99tMi4H4MqL44Uc5QKG77B0d6-JfIkZHFaUA71-RtjyYZWVIhqsNZcx8-OMaA?key=xt3VSDoCbmTY7o-cwwOFwQ)](https://hf.co/QuantFactory)
261
+
262
+
263
+ # QuantFactory/Einstein-v6.1-Llama3-8B-GGUF
264
+ This is quantized version of [Weyaxi/Einstein-v6.1-Llama3-8B](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) created using llama.cpp
265
 
266
+ # Original Model Card
267
 
268
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/5s12oq859qLfDkkTNam_C.png)
269
+
270
+ # πŸ”¬ Einstein-v6.1-Llama3-8B
271
 
272
  This model is a full fine-tuned version of [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) on diverse datasets.
273
 
 
466
 
467
  [Weyaxi/Einstein-v6.1-Llama3-8B/data](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B/tree/main/data)
468
 
469
+ # πŸ”„ Quantizationed versions
470
+
471
+ ## GGUF [@bartowski](https://huggingface.co/bartowski)
472
+
473
+ - https://huggingface.co/bartowski/Einstein-v6.1-Llama3-8B-GGUF
474
+
475
+ ## ExLlamaV2 [@bartowski](https://huggingface.co/bartowski)
476
+
477
+ - https://huggingface.co/bartowski/Einstein-v6.1-Llama3-8B-exl2
478
+
479
+ ## AWQ [@solidrust](https://huggingface.co/solidrust)
480
+
481
+ - https://huggingface.co/solidrust/Einstein-v6.1-Llama3-8B-AWQ
482
 
483
  # 🎯 [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
484
  Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v6.1-Llama3-8B)
 
493
  |Winogrande (5-shot) |79.32|
494
  |GSM8k (5-shot) |66.11|
495
 
496
+ # 🎯 [Open LLM Leaderboard v2 Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
497
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_Weyaxi__Einstein-v6.1-Llama3-8B)
498
+
499
+ | Metric |Value|
500
+ |-------------------|----:|
501
+ |Avg. |19.99|
502
+ |IFEval (0-Shot) |45.68|
503
+ |BBH (3-Shot) |29.38|
504
+ |MATH Lvl 5 (4-Shot)| 5.74|
505
+ |GPQA (0-shot) | 4.25|
506
+ |MuSR (0-shot) |11.23|
507
+ |MMLU-PRO (5-shot) |23.68|
508
+
509
+
510
+ # πŸ“š Some resources, discussions and reviews aboout this model
511
+
512
+ #### 🐦 Announcement tweet:
513
+
514
+ - https://twitter.com/Weyaxi/status/1783050724659675627
515
+
516
+ #### πŸ” Reddit post in r/LocalLLaMA:
517
+
518
+ - https://www.reddit.com/r/LocalLLaMA/comments/1cdlym1/introducing_einstein_v61_based_on_the_new_llama3/
519
+
520
+ #### ▢️ Youtube Video(s)
521
+
522
+ - [Install Einstein v6.1 Llama3-8B Locally on Windows](https://www.youtube.com/watch?v=VePvv6OM0JY)
523
+
524
+ #### πŸ“± Octopus-V4-3B
525
+
526
+ - [Octopus-V4-3B](https://huggingface.co/NexaAIDev/Octopus-v4) leverages the incredible physics capabilities of [Einstein-v6.1-Llama3-8B](https://huggingface.co/Weyaxi/Einstein-v6.1-Llama3-8B) in their model.
527
+
528
  # πŸ€– Additional information about training
529
 
530
  This model is full fine-tuned for 2 epoch.
 
535
 
536
  ![image/png](https://cdn-uploads.huggingface.co/production/uploads/6468ce47e134d050a58aa89c/Ycs7ZpoqmxFt0u9rybCO1.png)
537
 
538
+ </details><br>
539
+
540
+ # 🀝 Acknowledgments
541
+
542
+ Thanks to [sablo.ai](https://sablo.ai) for sponsoring this model.
543
+
544
+ Thanks to all the dataset authors mentioned in the datasets section.
545
+
546
+ Thanks to [axolotl](https://github.com/OpenAccess-AI-Collective/axolotl) for making the repository I used to make this model.
547
+
548
+ Thanks to all open source AI community.
549
+
550
+ [<img src="https://raw.githubusercontent.com/OpenAccess-AI-Collective/axolotl/main/image/axolotl-badge-web.png" alt="Built with Axolotl" width="200" height="32"/>](https://github.com/OpenAccess-AI-Collective/axolotl)
551
+
552
+ If you would like to support me:
553
+
554
+ [β˜• Buy Me a Coffee](https://www.buymeacoffee.com/weyaxi)