Files changed (1) hide show
  1. README.md +106 -0
README.md CHANGED
@@ -111,6 +111,98 @@ model-index:
111
  source:
112
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
113
  name: Open LLM Leaderboard
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
114
  ---
115
 
116
 
@@ -264,3 +356,17 @@ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-le
264
  |Winogrande (5-shot) |72.77|
265
  |GSM8k (5-shot) | 8.34|
266
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
111
  source:
112
  url: https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
113
  name: Open LLM Leaderboard
114
+ - task:
115
+ type: text-generation
116
+ name: Text Generation
117
+ dataset:
118
+ name: IFEval (0-Shot)
119
+ type: HuggingFaceH4/ifeval
120
+ args:
121
+ num_few_shot: 0
122
+ metrics:
123
+ - type: inst_level_strict_acc and prompt_level_strict_acc
124
+ value: 24.54
125
+ name: strict accuracy
126
+ source:
127
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
128
+ name: Open LLM Leaderboard
129
+ - task:
130
+ type: text-generation
131
+ name: Text Generation
132
+ dataset:
133
+ name: BBH (3-Shot)
134
+ type: BBH
135
+ args:
136
+ num_few_shot: 3
137
+ metrics:
138
+ - type: acc_norm
139
+ value: 13.52
140
+ name: normalized accuracy
141
+ source:
142
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
143
+ name: Open LLM Leaderboard
144
+ - task:
145
+ type: text-generation
146
+ name: Text Generation
147
+ dataset:
148
+ name: MATH Lvl 5 (4-Shot)
149
+ type: hendrycks/competition_math
150
+ args:
151
+ num_few_shot: 4
152
+ metrics:
153
+ - type: exact_match
154
+ value: 0.83
155
+ name: exact match
156
+ source:
157
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
158
+ name: Open LLM Leaderboard
159
+ - task:
160
+ type: text-generation
161
+ name: Text Generation
162
+ dataset:
163
+ name: GPQA (0-shot)
164
+ type: Idavidrein/gpqa
165
+ args:
166
+ num_few_shot: 0
167
+ metrics:
168
+ - type: acc_norm
169
+ value: 2.01
170
+ name: acc_norm
171
+ source:
172
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
173
+ name: Open LLM Leaderboard
174
+ - task:
175
+ type: text-generation
176
+ name: Text Generation
177
+ dataset:
178
+ name: MuSR (0-shot)
179
+ type: TAUR-Lab/MuSR
180
+ args:
181
+ num_few_shot: 0
182
+ metrics:
183
+ - type: acc_norm
184
+ value: 6.61
185
+ name: acc_norm
186
+ source:
187
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
188
+ name: Open LLM Leaderboard
189
+ - task:
190
+ type: text-generation
191
+ name: Text Generation
192
+ dataset:
193
+ name: MMLU-PRO (5-shot)
194
+ type: TIGER-Lab/MMLU-Pro
195
+ config: main
196
+ split: test
197
+ args:
198
+ num_few_shot: 5
199
+ metrics:
200
+ - type: acc
201
+ value: 12.7
202
+ name: accuracy
203
+ source:
204
+ url: https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard?query=pszemraj/Mistral-v0.3-6B
205
+ name: Open LLM Leaderboard
206
  ---
207
 
208
 
 
356
  |Winogrande (5-shot) |72.77|
357
  |GSM8k (5-shot) | 8.34|
358
 
359
+
360
+ # [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/open-llm-leaderboard/open_llm_leaderboard)
361
+ Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_pszemraj__Mistral-v0.3-6B)
362
+
363
+ | Metric |Value|
364
+ |-------------------|----:|
365
+ |Avg. |10.03|
366
+ |IFEval (0-Shot) |24.54|
367
+ |BBH (3-Shot) |13.52|
368
+ |MATH Lvl 5 (4-Shot)| 0.83|
369
+ |GPQA (0-shot) | 2.01|
370
+ |MuSR (0-shot) | 6.61|
371
+ |MMLU-PRO (5-shot) |12.70|
372
+