results / results /LASER2 /MassiveIntentClassification.json
Muennighoff's picture
Update results
4951470
raw history blame
No virus
13.2 kB
{
"mteb_version": "0.0.2",
"test": {
"af": {
"accuracy": 0.38012777404169473,
"accuracy_stderr": 0.018054485755774126,
"f1": 0.36372620791961896,
"f1_stderr": 0.011605337850752963,
"main_score": 0.38012777404169473
},
"am": {
"accuracy": 0.1269670477471419,
"accuracy_stderr": 0.024740686380275122,
"f1": 0.10125430985923822,
"f1_stderr": 0.014610275424102794,
"main_score": 0.1269670477471419
},
"ar": {
"accuracy": 0.3716207128446537,
"accuracy_stderr": 0.015242658586425109,
"f1": 0.3513819183071391,
"f1_stderr": 0.00952968827923218,
"main_score": 0.3716207128446537
},
"az": {
"accuracy": 0.19979825151311364,
"accuracy_stderr": 0.023512716506208713,
"f1": 0.20755182437536165,
"f1_stderr": 0.020871688941913907,
"main_score": 0.19979825151311364
},
"bn": {
"accuracy": 0.4251176866173504,
"accuracy_stderr": 0.015417634386490787,
"f1": 0.41808038563926464,
"f1_stderr": 0.015199386985110714,
"main_score": 0.4251176866173504
},
"cy": {
"accuracy": 0.17326832548755883,
"accuracy_stderr": 0.012898104386975457,
"f1": 0.18929774881076256,
"f1_stderr": 0.011323358864015197,
"main_score": 0.17326832548755883
},
"da": {
"accuracy": 0.4560860793544048,
"accuracy_stderr": 0.01951406090649724,
"f1": 0.4499828344382091,
"f1_stderr": 0.012363163669498014,
"main_score": 0.4560860793544048
},
"de": {
"accuracy": 0.44788164088769333,
"accuracy_stderr": 0.014665941706559757,
"f1": 0.4331277340788812,
"f1_stderr": 0.01522229039527465,
"main_score": 0.44788164088769333
},
"el": {
"accuracy": 0.4670813718897109,
"accuracy_stderr": 0.016337233490102906,
"f1": 0.4610946954649364,
"f1_stderr": 0.0176594232674652,
"main_score": 0.4670813718897109
},
"en": {
"accuracy": 0.4790854068594485,
"accuracy_stderr": 0.01717352517420073,
"f1": 0.464939146287881,
"f1_stderr": 0.012710556685691056,
"main_score": 0.4790854068594485
},
"es": {
"accuracy": 0.45437121721587087,
"accuracy_stderr": 0.014659040367775445,
"f1": 0.46286614708385815,
"f1_stderr": 0.01400822366938995,
"main_score": 0.45437121721587087
},
"evaluation_time": 2187.96,
"fa": {
"accuracy": 0.45013449899125757,
"accuracy_stderr": 0.018101267052589458,
"f1": 0.4440276709633043,
"f1_stderr": 0.02139785887315828,
"main_score": 0.45013449899125757
},
"fi": {
"accuracy": 0.45938130464021515,
"accuracy_stderr": 0.019983852698166247,
"f1": 0.4467651558171412,
"f1_stderr": 0.013228992239689133,
"main_score": 0.45938130464021515
},
"fr": {
"accuracy": 0.4613315400134499,
"accuracy_stderr": 0.019672058058454584,
"f1": 0.46149191584616656,
"f1_stderr": 0.01462338783534165,
"main_score": 0.4613315400134499
},
"he": {
"accuracy": 0.42545393409549426,
"accuracy_stderr": 0.015641146997013137,
"f1": 0.406277912154974,
"f1_stderr": 0.015808885221407522,
"main_score": 0.42545393409549426
},
"hi": {
"accuracy": 0.40201748486886346,
"accuracy_stderr": 0.014870504329345259,
"f1": 0.39955511381980663,
"f1_stderr": 0.012540162228863491,
"main_score": 0.40201748486886346
},
"hu": {
"accuracy": 0.4277404169468729,
"accuracy_stderr": 0.012916849604535326,
"f1": 0.41428552778550787,
"f1_stderr": 0.01796504857153477,
"main_score": 0.4277404169468729
},
"hy": {
"accuracy": 0.28073301950235374,
"accuracy_stderr": 0.022799686666722056,
"f1": 0.27060617914010276,
"f1_stderr": 0.011002544912943118,
"main_score": 0.28073301950235374
},
"id": {
"accuracy": 0.4581035642232683,
"accuracy_stderr": 0.01719550009501075,
"f1": 0.44698433663656323,
"f1_stderr": 0.016362307533275874,
"main_score": 0.4581035642232683
},
"is": {
"accuracy": 0.39862138533960995,
"accuracy_stderr": 0.016819937773966665,
"f1": 0.3920844117598715,
"f1_stderr": 0.015188173368411415,
"main_score": 0.39862138533960995
},
"it": {
"accuracy": 0.4825487558843308,
"accuracy_stderr": 0.017093447378236502,
"f1": 0.4800020656000982,
"f1_stderr": 0.014433139843477095,
"main_score": 0.4825487558843308
},
"ja": {
"accuracy": 0.45299260255548085,
"accuracy_stderr": 0.020810866845063326,
"f1": 0.45499154063652886,
"f1_stderr": 0.015152638500548954,
"main_score": 0.45299260255548085
},
"jv": {
"accuracy": 0.24303967720242098,
"accuracy_stderr": 0.018405242193104722,
"f1": 0.24206441275862223,
"f1_stderr": 0.01230622384839183,
"main_score": 0.24303967720242098
},
"ka": {
"accuracy": 0.22700067249495626,
"accuracy_stderr": 0.011315564480105716,
"f1": 0.2262620454603394,
"f1_stderr": 0.013641428742103908,
"main_score": 0.22700067249495626
},
"km": {
"accuracy": 0.22481506388702083,
"accuracy_stderr": 0.017626681644496683,
"f1": 0.20273821769692973,
"f1_stderr": 0.010423853100542939,
"main_score": 0.22481506388702083
},
"kn": {
"accuracy": 0.043207800941492944,
"accuracy_stderr": 0.011439581992919204,
"f1": 0.019591698863663035,
"f1_stderr": 0.004395348203545399,
"main_score": 0.043207800941492944
},
"ko": {
"accuracy": 0.4426025554808339,
"accuracy_stderr": 0.024004890918808426,
"f1": 0.44147618521379844,
"f1_stderr": 0.01371530289091307,
"main_score": 0.4426025554808339
},
"lv": {
"accuracy": 0.3975453934095495,
"accuracy_stderr": 0.02350514180018239,
"f1": 0.4061787392665471,
"f1_stderr": 0.02079966697693215,
"main_score": 0.3975453934095495
},
"ml": {
"accuracy": 0.4133490248823134,
"accuracy_stderr": 0.022600758621339218,
"f1": 0.40286261835633236,
"f1_stderr": 0.019466174082752453,
"main_score": 0.4133490248823134
},
"mn": {
"accuracy": 0.16200403496973773,
"accuracy_stderr": 0.0100095627651343,
"f1": 0.16450176344768086,
"f1_stderr": 0.01474426560593057,
"main_score": 0.16200403496973773
},
"ms": {
"accuracy": 0.4322797579018157,
"accuracy_stderr": 0.01759072503596667,
"f1": 0.4248684950682716,
"f1_stderr": 0.016635102291421434,
"main_score": 0.4322797579018157
},
"my": {
"accuracy": 0.25373234700739744,
"accuracy_stderr": 0.014838536825778608,
"f1": 0.2222952356057096,
"f1_stderr": 0.013501549124465934,
"main_score": 0.25373234700739744
},
"nb": {
"accuracy": 0.3773705447209146,
"accuracy_stderr": 0.021859215543374035,
"f1": 0.3822420276063557,
"f1_stderr": 0.01781935165480971,
"main_score": 0.3773705447209146
},
"nl": {
"accuracy": 0.45000000000000007,
"accuracy_stderr": 0.016798952275448243,
"f1": 0.4390949593879808,
"f1_stderr": 0.013457897429685057,
"main_score": 0.45000000000000007
},
"pl": {
"accuracy": 0.4499327505043712,
"accuracy_stderr": 0.018087145352772453,
"f1": 0.42978932594492064,
"f1_stderr": 0.020420111856607633,
"main_score": 0.4499327505043712
},
"pt": {
"accuracy": 0.4855413584398117,
"accuracy_stderr": 0.01347928867036143,
"f1": 0.48571088423157766,
"f1_stderr": 0.013222183663454462,
"main_score": 0.4855413584398117
},
"ro": {
"accuracy": 0.44300605245460656,
"accuracy_stderr": 0.021637858328300265,
"f1": 0.44740069972880264,
"f1_stderr": 0.01577190962835917,
"main_score": 0.44300605245460656
},
"ru": {
"accuracy": 0.44290517821116343,
"accuracy_stderr": 0.021596827649085326,
"f1": 0.43747519563352766,
"f1_stderr": 0.019105088547632505,
"main_score": 0.44290517821116343
},
"sl": {
"accuracy": 0.44717552118359116,
"accuracy_stderr": 0.01726980272383147,
"f1": 0.4583434006703727,
"f1_stderr": 0.016199198460333516,
"main_score": 0.44717552118359116
},
"sq": {
"accuracy": 0.4612306657700067,
"accuracy_stderr": 0.01641084055632617,
"f1": 0.4549674955169706,
"f1_stderr": 0.013851645495939875,
"main_score": 0.4612306657700067
},
"sv": {
"accuracy": 0.4594821788836584,
"accuracy_stderr": 0.020832831141935834,
"f1": 0.44306077100248525,
"f1_stderr": 0.019118675204039762,
"main_score": 0.4594821788836584
},
"sw": {
"accuracy": 0.31893073301950237,
"accuracy_stderr": 0.011235144820777232,
"f1": 0.31391383206208073,
"f1_stderr": 0.014373830497544215,
"main_score": 0.31893073301950237
},
"ta": {
"accuracy": 0.2963349024882313,
"accuracy_stderr": 0.012793365954792035,
"f1": 0.2888382113446147,
"f1_stderr": 0.01179745104077985,
"main_score": 0.2963349024882313
},
"te": {
"accuracy": 0.36028917283120376,
"accuracy_stderr": 0.018898933014207392,
"f1": 0.3438728561319695,
"f1_stderr": 0.015037693640197282,
"main_score": 0.36028917283120376
},
"th": {
"accuracy": 0.4338601210490921,
"accuracy_stderr": 0.01838326805184409,
"f1": 0.44123009965336957,
"f1_stderr": 0.014967999591595584,
"main_score": 0.4338601210490921
},
"tl": {
"accuracy": 0.2973436449226631,
"accuracy_stderr": 0.01833399973821919,
"f1": 0.28090429460991373,
"f1_stderr": 0.012358692677216883,
"main_score": 0.2973436449226631
},
"tr": {
"accuracy": 0.43930733019502355,
"accuracy_stderr": 0.02477039571658176,
"f1": 0.4265561991600836,
"f1_stderr": 0.022060686946343926,
"main_score": 0.43930733019502355
},
"ur": {
"accuracy": 0.26109616677874925,
"accuracy_stderr": 0.011974612577756474,
"f1": 0.27607059745773804,
"f1_stderr": 0.011634207629783868,
"main_score": 0.26109616677874925
},
"vi": {
"accuracy": 0.44327505043712173,
"accuracy_stderr": 0.022061037910983305,
"f1": 0.43953676685827936,
"f1_stderr": 0.01965182342188036,
"main_score": 0.44327505043712173
},
"zh-CN": {
"accuracy": 0.40618695359784807,
"accuracy_stderr": 0.01878416975952488,
"f1": 0.40304322644486057,
"f1_stderr": 0.016358332856736345,
"main_score": 0.40618695359784807
},
"zh-TW": {
"accuracy": 0.3292871553463349,
"accuracy_stderr": 0.012964553395663082,
"f1": 0.3310118176304551,
"f1_stderr": 0.00939726932836698,
"main_score": 0.3292871553463349
}
},
"mteb_dataset_name": "MassiveIntentClassification",
"dataset_revision": "072a486a144adf7f4479a4a0dddb2152e161e1ea"
}