Spaces:
Running
Running
Add mini models
#1
by
jansowa
- opened
data.json
CHANGED
|
@@ -206,5 +206,109 @@
|
|
| 206 |
"Sentiment": 3.301282051282051,
|
| 207 |
"Language understanding": 2.755,
|
| 208 |
"Phraseology": 2.165
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 209 |
}
|
| 210 |
]
|
|
|
|
| 206 |
"Sentiment": 3.301282051282051,
|
| 207 |
"Language understanding": 2.755,
|
| 208 |
"Phraseology": 2.165
|
| 209 |
+
},
|
| 210 |
+
{
|
| 211 |
+
"Model": "google/gemma-2-2b-it",
|
| 212 |
+
"Params": "2.61B",
|
| 213 |
+
"Average": 2.7974786324786325,
|
| 214 |
+
"Sentiment": 3.3974358974359,
|
| 215 |
+
"Language understanding": 2.9,
|
| 216 |
+
"Phraseology": 2.095
|
| 217 |
+
},
|
| 218 |
+
{
|
| 219 |
+
"Model": "Qwen/Qwen2.5-3B-Instruct",
|
| 220 |
+
"Params": "3.09B",
|
| 221 |
+
"Average": 2.734572649572649,
|
| 222 |
+
"Sentiment": 2.948717948717949,
|
| 223 |
+
"Language understanding": 2.455,
|
| 224 |
+
"Phraseology": 2.8
|
| 225 |
+
},
|
| 226 |
+
{
|
| 227 |
+
"Model": "NousResearch/Hermes-3-Llama-3.2-3B",
|
| 228 |
+
"Params": "3.21B",
|
| 229 |
+
"Average": 2.695128205128205,
|
| 230 |
+
"Sentiment": 2.6153846153846154,
|
| 231 |
+
"Language understanding": 2.705,
|
| 232 |
+
"Phraseology": 2.765
|
| 233 |
+
},
|
| 234 |
+
{
|
| 235 |
+
"Model": "ibm-granite/granite-3.1-2b-instruct",
|
| 236 |
+
"Params": "2.53B",
|
| 237 |
+
"Average": 2.397307692307692,
|
| 238 |
+
"Sentiment": 3.076923076923077,
|
| 239 |
+
"Language understanding": 2.235,
|
| 240 |
+
"Phraseology": 1.88
|
| 241 |
+
},
|
| 242 |
+
{
|
| 243 |
+
"Model": "meta-llama/Llama-3.2-1B-Instruct",
|
| 244 |
+
"Params": "1.24B",
|
| 245 |
+
"Average": 2.383974358974359,
|
| 246 |
+
"Sentiment": 3.076923076923077,
|
| 247 |
+
"Language understanding": 1.735,
|
| 248 |
+
"Phraseology": 2.34
|
| 249 |
+
},
|
| 250 |
+
{
|
| 251 |
+
"Model": "microsoft/Phi-3.5-mini-instruct",
|
| 252 |
+
"Params": "3.82B",
|
| 253 |
+
"Average": 2.331965811965812,
|
| 254 |
+
"Sentiment": 2.435897435897436,
|
| 255 |
+
"Language understanding": 2.135,
|
| 256 |
+
"Phraseology": 2.425
|
| 257 |
+
},
|
| 258 |
+
{
|
| 259 |
+
"Model": "meta-llama/Llama-3.2-3B-Instruct",
|
| 260 |
+
"Params": "3.21B",
|
| 261 |
+
"Average": 2.257136752136752,
|
| 262 |
+
"Sentiment": 2.7564102564102564,
|
| 263 |
+
"Language understanding": 2.295,
|
| 264 |
+
"Phraseology": 1.72
|
| 265 |
+
},
|
| 266 |
+
{
|
| 267 |
+
"Model": "h2oai/h2o-danube2-1.8b-chat",
|
| 268 |
+
"Params": "1.83B",
|
| 269 |
+
"Average": 2.1455982905982904,
|
| 270 |
+
"Sentiment": 2.371794871794872,
|
| 271 |
+
"Language understanding": 1.595,
|
| 272 |
+
"Phraseology": 2.47
|
| 273 |
+
},
|
| 274 |
+
{
|
| 275 |
+
"Model": "Qwen/Qwen2.5-1.5B-Instruct",
|
| 276 |
+
"Params": "1.54B",
|
| 277 |
+
"Average": 2.1232905982905983,
|
| 278 |
+
"Sentiment": 2.7948717948717947,
|
| 279 |
+
"Language understanding": 1.35,
|
| 280 |
+
"Phraseology": 2.225
|
| 281 |
+
},
|
| 282 |
+
{
|
| 283 |
+
"Model": "utter-project/EuroLLM-1.7B-Instruct",
|
| 284 |
+
"Params": "1.66B",
|
| 285 |
+
"Average": 2.097863247863248,
|
| 286 |
+
"Sentiment": 2.243589743589744,
|
| 287 |
+
"Language understanding": 1.79,
|
| 288 |
+
"Phraseology": 2.26
|
| 289 |
+
},
|
| 290 |
+
{
|
| 291 |
+
"Model": "LGAI-EXAONE/EXAONE-3.5-2.4B-Instruct",
|
| 292 |
+
"Params": "2.41B",
|
| 293 |
+
"Average": 2.062846282695529,
|
| 294 |
+
"Sentiment": 1.9423076923076923,
|
| 295 |
+
"Language understanding": 2.1155778894472363,
|
| 296 |
+
"Phraseology": 2.130653266331658
|
| 297 |
+
},
|
| 298 |
+
{
|
| 299 |
+
"Model": "HuggingFaceTB/SmolLM2-1.7B-Instruct",
|
| 300 |
+
"Params": "1.71B",
|
| 301 |
+
"Average": 1.9102136752136751,
|
| 302 |
+
"Sentiment": 2.275641025641025,
|
| 303 |
+
"Language understanding": 1.1,
|
| 304 |
+
"Phraseology": 2.355
|
| 305 |
+
},
|
| 306 |
+
{
|
| 307 |
+
"Model": "Qwen/Qwen2.5-0.5B-Instruct",
|
| 308 |
+
"Params": "0.49B",
|
| 309 |
+
"Average": 1.7950427350427354,
|
| 310 |
+
"Sentiment": 1.955128205128205,
|
| 311 |
+
"Language understanding": 0.835,
|
| 312 |
+
"Phraseology": 2.595
|
| 313 |
}
|
| 314 |
]
|