Muennighoff commited on
Commit
d7e613a
1 Parent(s): bb38d52

Update results/LaBSE/MassiveScenarioClassification.json

Browse files
results/LaBSE/MassiveScenarioClassification.json CHANGED
@@ -1,15 +1,365 @@
1
  {
2
- "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
- "mteb_dataset_name": "MassiveScenarioClassification",
4
- "mteb_version": "1.1.1",
5
- "test": {
6
- "evaluation_time": 3.32,
7
- "pl": {
8
- "accuracy": 0.645796906523201,
9
- "accuracy_stderr": 0.01695247609678737,
10
- "f1": 0.6313195906066588,
11
- "f1_stderr": 0.01393254681254561,
12
- "main_score": 0.645796906523201
13
- }
14
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
15
  }
 
1
  {
2
+ "mteb_version": "0.0.2",
3
+ "test": {
4
+ "af": {
5
+ "accuracy": 0.6338601210490922,
6
+ "accuracy_stderr": 0.020796002685561546,
7
+ "f1": 0.6099940718223488,
8
+ "f1_stderr": 0.01841464431126804,
9
+ "main_score": 0.6338601210490922
10
+ },
11
+ "am": {
12
+ "accuracy": 0.6202084734364492,
13
+ "accuracy_stderr": 0.01815066876992255,
14
+ "f1": 0.5967740081153838,
15
+ "f1_stderr": 0.018058548623767733,
16
+ "main_score": 0.6202084734364492
17
+ },
18
+ "ar": {
19
+ "accuracy": 0.577236045729657,
20
+ "accuracy_stderr": 0.01902640838070928,
21
+ "f1": 0.5682019373821332,
22
+ "f1_stderr": 0.018960215329018097,
23
+ "main_score": 0.577236045729657
24
+ },
25
+ "az": {
26
+ "accuracy": 0.6348352387357095,
27
+ "accuracy_stderr": 0.01985739936668887,
28
+ "f1": 0.6182372131316041,
29
+ "f1_stderr": 0.020831225444742495,
30
+ "main_score": 0.6348352387357095
31
+ },
32
+ "bn": {
33
+ "accuracy": 0.6183591123066576,
34
+ "accuracy_stderr": 0.020491242444606123,
35
+ "f1": 0.602302706858818,
36
+ "f1_stderr": 0.021391943738139053,
37
+ "main_score": 0.6183591123066576
38
+ },
39
+ "cy": {
40
+ "accuracy": 0.5613315400134499,
41
+ "accuracy_stderr": 0.017472707994897536,
42
+ "f1": 0.5238803678743691,
43
+ "f1_stderr": 0.016429323703196226,
44
+ "main_score": 0.5613315400134499
45
+ },
46
+ "da": {
47
+ "accuracy": 0.6523873570948219,
48
+ "accuracy_stderr": 0.017971350361487184,
49
+ "f1": 0.6330101816140603,
50
+ "f1_stderr": 0.017387997719007987,
51
+ "main_score": 0.6523873570948219
52
+ },
53
+ "de": {
54
+ "accuracy": 0.6238735709482179,
55
+ "accuracy_stderr": 0.014762145094762657,
56
+ "f1": 0.6037455953818265,
57
+ "f1_stderr": 0.016045304426233677,
58
+ "main_score": 0.6238735709482179
59
+ },
60
+ "el": {
61
+ "accuracy": 0.6457632817753867,
62
+ "accuracy_stderr": 0.013769118685050997,
63
+ "f1": 0.621090672305202,
64
+ "f1_stderr": 0.014362362038430058,
65
+ "main_score": 0.6457632817753867
66
+ },
67
+ "en": {
68
+ "accuracy": 0.6640551445864157,
69
+ "accuracy_stderr": 0.018188129307017843,
70
+ "f1": 0.6413643155720361,
71
+ "f1_stderr": 0.019284478720966253,
72
+ "main_score": 0.6640551445864157
73
+ },
74
+ "es": {
75
+ "accuracy": 0.6361129791526563,
76
+ "accuracy_stderr": 0.013220812993315708,
77
+ "f1": 0.6217726502690191,
78
+ "f1_stderr": 0.015117153814962439,
79
+ "main_score": 0.6361129791526563
80
+ },
81
+ "evaluation_time": 173.37,
82
+ "fa": {
83
+ "accuracy": 0.67457969065232,
84
+ "accuracy_stderr": 0.01896853059253584,
85
+ "f1": 0.6557868313320845,
86
+ "f1_stderr": 0.02022274261697446,
87
+ "main_score": 0.67457969065232
88
+ },
89
+ "fi": {
90
+ "accuracy": 0.645796906523201,
91
+ "accuracy_stderr": 0.020912430665552845,
92
+ "f1": 0.6294630817365124,
93
+ "f1_stderr": 0.020135664379759458,
94
+ "main_score": 0.645796906523201
95
+ },
96
+ "fr": {
97
+ "accuracy": 0.6510423671822462,
98
+ "accuracy_stderr": 0.014213209140494122,
99
+ "f1": 0.6376153901218867,
100
+ "f1_stderr": 0.015419098040681577,
101
+ "main_score": 0.6510423671822462
102
+ },
103
+ "he": {
104
+ "accuracy": 0.635305985205111,
105
+ "accuracy_stderr": 0.013806798054280704,
106
+ "f1": 0.6186780542893299,
107
+ "f1_stderr": 0.015999534462356246,
108
+ "main_score": 0.635305985205111
109
+ },
110
+ "hi": {
111
+ "accuracy": 0.6440147948890383,
112
+ "accuracy_stderr": 0.023100495412540042,
113
+ "f1": 0.6259674137995981,
114
+ "f1_stderr": 0.022396786160345798,
115
+ "main_score": 0.6440147948890383
116
+ },
117
+ "hu": {
118
+ "accuracy": 0.6582380632145259,
119
+ "accuracy_stderr": 0.01718366081479456,
120
+ "f1": 0.6422229714958034,
121
+ "f1_stderr": 0.017977882656527718,
122
+ "main_score": 0.6582380632145259
123
+ },
124
+ "hy": {
125
+ "accuracy": 0.612542030934768,
126
+ "accuracy_stderr": 0.016765266889829294,
127
+ "f1": 0.5984886286461146,
128
+ "f1_stderr": 0.018147557735502366,
129
+ "main_score": 0.612542030934768
130
+ },
131
+ "id": {
132
+ "accuracy": 0.6584398117014122,
133
+ "accuracy_stderr": 0.011284898191673205,
134
+ "f1": 0.641771604889956,
135
+ "f1_stderr": 0.01276975957788136,
136
+ "main_score": 0.6584398117014122
137
+ },
138
+ "is": {
139
+ "accuracy": 0.6194014794889038,
140
+ "accuracy_stderr": 0.01817755857211039,
141
+ "f1": 0.5906267988780722,
142
+ "f1_stderr": 0.0196952877595304,
143
+ "main_score": 0.6194014794889038
144
+ },
145
+ "it": {
146
+ "accuracy": 0.6408540685944856,
147
+ "accuracy_stderr": 0.013814370701911194,
148
+ "f1": 0.6261786684273842,
149
+ "f1_stderr": 0.012287900660372282,
150
+ "main_score": 0.6408540685944856
151
+ },
152
+ "ja": {
153
+ "accuracy": 0.677236045729657,
154
+ "accuracy_stderr": 0.015110536676124438,
155
+ "f1": 0.6673296622305556,
156
+ "f1_stderr": 0.014545265726506836,
157
+ "main_score": 0.677236045729657
158
+ },
159
+ "jv": {
160
+ "accuracy": 0.5829186281102892,
161
+ "accuracy_stderr": 0.008523989631910896,
162
+ "f1": 0.5680383317289192,
163
+ "f1_stderr": 0.011313508026952397,
164
+ "main_score": 0.5829186281102892
165
+ },
166
+ "ka": {
167
+ "accuracy": 0.5338264963012778,
168
+ "accuracy_stderr": 0.025989211229142786,
169
+ "f1": 0.5217669546768694,
170
+ "f1_stderr": 0.02247329105813734,
171
+ "main_score": 0.5338264963012778
172
+ },
173
+ "km": {
174
+ "accuracy": 0.5618359112306657,
175
+ "accuracy_stderr": 0.013807821627734375,
176
+ "f1": 0.5240173523119068,
177
+ "f1_stderr": 0.015390629080603253,
178
+ "main_score": 0.5618359112306657
179
+ },
180
+ "kn": {
181
+ "accuracy": 0.6173839946200405,
182
+ "accuracy_stderr": 0.022463470737220846,
183
+ "f1": 0.5947574042335859,
184
+ "f1_stderr": 0.024678486777218324,
185
+ "main_score": 0.6173839946200405
186
+ },
187
+ "ko": {
188
+ "accuracy": 0.6726294552790855,
189
+ "accuracy_stderr": 0.009797676852321914,
190
+ "f1": 0.6610832939650482,
191
+ "f1_stderr": 0.01204782326863115,
192
+ "main_score": 0.6726294552790855
193
+ },
194
+ "lv": {
195
+ "accuracy": 0.6186953597848015,
196
+ "accuracy_stderr": 0.01811746843771493,
197
+ "f1": 0.6008441860590075,
198
+ "f1_stderr": 0.017567988137501225,
199
+ "main_score": 0.6186953597848015
200
+ },
201
+ "ml": {
202
+ "accuracy": 0.6226294552790854,
203
+ "accuracy_stderr": 0.025463526667790882,
204
+ "f1": 0.6042233053035149,
205
+ "f1_stderr": 0.023024555660579968,
206
+ "main_score": 0.6226294552790854
207
+ },
208
+ "mn": {
209
+ "accuracy": 0.626025554808339,
210
+ "accuracy_stderr": 0.02499510957432111,
211
+ "f1": 0.6089167910409335,
212
+ "f1_stderr": 0.022766453349033586,
213
+ "main_score": 0.626025554808339
214
+ },
215
+ "ms": {
216
+ "accuracy": 0.6563214525891056,
217
+ "accuracy_stderr": 0.019678637683402438,
218
+ "f1": 0.6310162425478862,
219
+ "f1_stderr": 0.018392294088964707,
220
+ "main_score": 0.6563214525891056
221
+ },
222
+ "my": {
223
+ "accuracy": 0.6293880295897781,
224
+ "accuracy_stderr": 0.018817725940331854,
225
+ "f1": 0.6047944369727123,
226
+ "f1_stderr": 0.015766961962910157,
227
+ "main_score": 0.6293880295897781
228
+ },
229
+ "nb": {
230
+ "accuracy": 0.642871553463349,
231
+ "accuracy_stderr": 0.02334460927972955,
232
+ "f1": 0.6240202970047466,
233
+ "f1_stderr": 0.023415024180900877,
234
+ "main_score": 0.642871553463349
235
+ },
236
+ "nl": {
237
+ "accuracy": 0.6515803631472764,
238
+ "accuracy_stderr": 0.01549280096394042,
239
+ "f1": 0.6314226773432929,
240
+ "f1_stderr": 0.01521046780336452,
241
+ "main_score": 0.6515803631472764
242
+ },
243
+ "pl": {
244
+ "accuracy": 0.645796906523201,
245
+ "accuracy_stderr": 0.01695247609678737,
246
+ "f1": 0.6313195906066588,
247
+ "f1_stderr": 0.01393254681254561,
248
+ "main_score": 0.645796906523201
249
+ },
250
+ "pt": {
251
+ "accuracy": 0.6327505043712172,
252
+ "accuracy_stderr": 0.016159341549333697,
253
+ "f1": 0.6223577258815645,
254
+ "f1_stderr": 0.016447885132132502,
255
+ "main_score": 0.6327505043712172
256
+ },
257
+ "ro": {
258
+ "accuracy": 0.6241089441829187,
259
+ "accuracy_stderr": 0.017778208995554653,
260
+ "f1": 0.6130555712806881,
261
+ "f1_stderr": 0.015285219550136373,
262
+ "main_score": 0.6241089441829187
263
+ },
264
+ "ru": {
265
+ "accuracy": 0.6524546065904506,
266
+ "accuracy_stderr": 0.01314568535840283,
267
+ "f1": 0.6407016018626791,
268
+ "f1_stderr": 0.012908135250465469,
269
+ "main_score": 0.6524546065904506
270
+ },
271
+ "sl": {
272
+ "accuracy": 0.6424680564895764,
273
+ "accuracy_stderr": 0.015024694728800705,
274
+ "f1": 0.6261012971718346,
275
+ "f1_stderr": 0.012170667095875231,
276
+ "main_score": 0.6424680564895764
277
+ },
278
+ "sq": {
279
+ "accuracy": 0.6453934095494285,
280
+ "accuracy_stderr": 0.01969285250513722,
281
+ "f1": 0.6182473942587723,
282
+ "f1_stderr": 0.019780635650422644,
283
+ "main_score": 0.6453934095494285
284
+ },
285
+ "sv": {
286
+ "accuracy": 0.6600874243443174,
287
+ "accuracy_stderr": 0.017692785073748635,
288
+ "f1": 0.6386055303219663,
289
+ "f1_stderr": 0.017498030932514213,
290
+ "main_score": 0.6600874243443174
291
+ },
292
+ "sw": {
293
+ "accuracy": 0.5835574983187627,
294
+ "accuracy_stderr": 0.018941360951120256,
295
+ "f1": 0.5664574114155856,
296
+ "f1_stderr": 0.020324479741315266,
297
+ "main_score": 0.5835574983187627
298
+ },
299
+ "ta": {
300
+ "accuracy": 0.5907868190988568,
301
+ "accuracy_stderr": 0.020551846364448793,
302
+ "f1": 0.5762261681371181,
303
+ "f1_stderr": 0.019093653511054066,
304
+ "main_score": 0.5907868190988568
305
+ },
306
+ "te": {
307
+ "accuracy": 0.6412575655682582,
308
+ "accuracy_stderr": 0.027359244638229925,
309
+ "f1": 0.6165311839093729,
310
+ "f1_stderr": 0.02617785896886573,
311
+ "main_score": 0.6412575655682582
312
+ },
313
+ "th": {
314
+ "accuracy": 0.6434095494283792,
315
+ "accuracy_stderr": 0.011984994100699071,
316
+ "f1": 0.6232423578482493,
317
+ "f1_stderr": 0.0130395965672043,
318
+ "main_score": 0.6434095494283792
319
+ },
320
+ "tl": {
321
+ "accuracy": 0.6023201075991931,
322
+ "accuracy_stderr": 0.02225544306549688,
323
+ "f1": 0.580787809209993,
324
+ "f1_stderr": 0.019404074813735208,
325
+ "main_score": 0.6023201075991931
326
+ },
327
+ "tr": {
328
+ "accuracy": 0.6542703429724277,
329
+ "accuracy_stderr": 0.018136055643283747,
330
+ "f1": 0.6394731268332807,
331
+ "f1_stderr": 0.01798508974291203,
332
+ "main_score": 0.6542703429724277
333
+ },
334
+ "ur": {
335
+ "accuracy": 0.6152320107599193,
336
+ "accuracy_stderr": 0.017061403798171403,
337
+ "f1": 0.5933963394603274,
338
+ "f1_stderr": 0.018439116553404755,
339
+ "main_score": 0.6152320107599193
340
+ },
341
+ "vi": {
342
+ "accuracy": 0.6105245460659046,
343
+ "accuracy_stderr": 0.015053262957893667,
344
+ "f1": 0.5968939699243327,
345
+ "f1_stderr": 0.0142592079200388,
346
+ "main_score": 0.6105245460659046
347
+ },
348
+ "zh-CN": {
349
+ "accuracy": 0.7085406859448554,
350
+ "accuracy_stderr": 0.01673472316681262,
351
+ "f1": 0.6924609266534586,
352
+ "f1_stderr": 0.015743076107271187,
353
+ "main_score": 0.7085406859448554
354
+ },
355
+ "zh-TW": {
356
+ "accuracy": 0.6708473436449226,
357
+ "accuracy_stderr": 0.01732536136279832,
358
+ "f1": 0.6581357591960727,
359
+ "f1_stderr": 0.019380735172867463,
360
+ "main_score": 0.6708473436449226
361
+ }
362
+ },
363
+ "mteb_dataset_name": "MassiveScenarioClassification",
364
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634"
365
  }