results / results /SGPT-125M-weightedmean-nli-bitfit /MTOPDomainClassification.json
Muennighoff's picture
Update results
4951470
raw history blame
No virus
3.31 kB
{
"test": {
"de": {
"accuracy": 0.6037475345167653,
"accuracy_stderr": 0.013460239029994478,
"f1": 0.5845264937551703,
"f1_stderr": 0.013734670702604352,
"main_score": 0.6037475345167653
},
"en": {
"accuracy": 0.8151846785225718,
"accuracy_stderr": 0.011461847240672736,
"f1": 0.81648869152345,
"f1_stderr": 0.010871461184853548,
"main_score": 0.8151846785225718
},
"es": {
"accuracy": 0.6736824549699799,
"accuracy_stderr": 0.01407356066869247,
"f1": 0.6535927434998515,
"f1_stderr": 0.01413855160644313,
"main_score": 0.6736824549699799
},
"evaluation_time": 168.65,
"fr": {
"accuracy": 0.6312871907297212,
"accuracy_stderr": 0.02473728093465062,
"f1": 0.6137620329272278,
"f1_stderr": 0.022860658324147663,
"main_score": 0.6312871907297212
},
"hi": {
"accuracy": 0.47045536034420943,
"accuracy_stderr": 0.019444896819423425,
"f1": 0.46203899126445613,
"f1_stderr": 0.01750055815526211,
"main_score": 0.47045536034420943
},
"th": {
"accuracy": 0.5228209764918625,
"accuracy_stderr": 0.030553280316660377,
"f1": 0.5075489206473579,
"f1_stderr": 0.02674134220327261,
"main_score": 0.5228209764918625
}
},
"validation": {
"de": {
"accuracy": 0.5924517906336088,
"accuracy_stderr": 0.010919798846375912,
"f1": 0.5786788916095694,
"f1_stderr": 0.012936802476599582,
"main_score": 0.5924517906336088
},
"en": {
"accuracy": 0.8214317673378077,
"accuracy_stderr": 0.01212263771708471,
"f1": 0.8270274918463174,
"f1_stderr": 0.011321392162566895,
"main_score": 0.8214317673378077
},
"es": {
"accuracy": 0.6726915520628685,
"accuracy_stderr": 0.017154326092686127,
"f1": 0.6610458353302849,
"f1_stderr": 0.01789675352267575,
"main_score": 0.6726915520628685
},
"evaluation_time": 105.77,
"fr": {
"accuracy": 0.6336715282181358,
"accuracy_stderr": 0.026483031232693195,
"f1": 0.6251191192231544,
"f1_stderr": 0.02294236420005774,
"main_score": 0.6336715282181358
},
"hi": {
"accuracy": 0.4635685884691848,
"accuracy_stderr": 0.01708910577573112,
"f1": 0.46090151154109493,
"f1_stderr": 0.01766759890065127,
"main_score": 0.4635685884691848
},
"th": {
"accuracy": 0.5128067025733094,
"accuracy_stderr": 0.020409398758064606,
"f1": 0.5018323841221017,
"f1_stderr": 0.01679125970157935,
"main_score": 0.5128067025733094
}
},
"mteb_version": "0.0.2",
"mteb_dataset_name": "MTOPDomainClassification",
"dataset_revision": "a7e2a951126a26fc8c6a69f835f33a346ba259e3"
}