Muennighoff lorenpe2 commited on
Commit
f022761
1 Parent(s): 913e1f0

add results - distiluse-base-multilingual-cased-v2 (#7)

Browse files

- add results - distiluse-base-multilingual-cased-v2 (a67fa048bfb63d3949db16549b96ed8a46d7b11e)


Co-authored-by: Petr Lorenc <lorenpe2@users.noreply.huggingface.co>

Files changed (30) hide show
  1. results/distiluse-base-multilingual-cased-v2/AmazonCounterfactualClassification.json +83 -0
  2. results/distiluse-base-multilingual-cased-v2/AmazonPolarityClassification.json +15 -0
  3. results/distiluse-base-multilingual-cased-v2/AmazonReviewsClassification.json +95 -0
  4. results/distiluse-base-multilingual-cased-v2/ArxivClusteringP2P.json +10 -0
  5. results/distiluse-base-multilingual-cased-v2/AskUbuntuDupQuestions.json +10 -0
  6. results/distiluse-base-multilingual-cased-v2/BIOSSES.json +20 -0
  7. results/distiluse-base-multilingual-cased-v2/Banking77Classification.json +13 -0
  8. results/distiluse-base-multilingual-cased-v2/EmotionClassification.json +21 -0
  9. results/distiluse-base-multilingual-cased-v2/ImdbClassification.json +15 -0
  10. results/distiluse-base-multilingual-cased-v2/MTOPDomainClassification.json +95 -0
  11. results/distiluse-base-multilingual-cased-v2/MTOPIntentClassification.json +95 -0
  12. results/distiluse-base-multilingual-cased-v2/MassiveIntentClassification.json +725 -0
  13. results/distiluse-base-multilingual-cased-v2/MassiveScenarioClassification.json +725 -0
  14. results/distiluse-base-multilingual-cased-v2/MindSmallReranking.json +10 -0
  15. results/distiluse-base-multilingual-cased-v2/SICK-R.json +20 -0
  16. results/distiluse-base-multilingual-cased-v2/STS12.json +20 -0
  17. results/distiluse-base-multilingual-cased-v2/STS13.json +20 -0
  18. results/distiluse-base-multilingual-cased-v2/STS14.json +20 -0
  19. results/distiluse-base-multilingual-cased-v2/STS15.json +20 -0
  20. results/distiluse-base-multilingual-cased-v2/STS16.json +20 -0
  21. results/distiluse-base-multilingual-cased-v2/STS17.json +162 -0
  22. results/distiluse-base-multilingual-cased-v2/STS22.json +260 -0
  23. results/distiluse-base-multilingual-cased-v2/STSBenchmark.json +35 -0
  24. results/distiluse-base-multilingual-cased-v2/SciDocsRR.json +15 -0
  25. results/distiluse-base-multilingual-cased-v2/SprintDuplicateQuestions.json +93 -0
  26. results/distiluse-base-multilingual-cased-v2/StackOverflowDupQuestions.json +10 -0
  27. results/distiluse-base-multilingual-cased-v2/ToxicConversationsClassification.json +15 -0
  28. results/distiluse-base-multilingual-cased-v2/TweetSentimentExtractionClassification.json +13 -0
  29. results/distiluse-base-multilingual-cased-v2/TwitterSemEval2015.json +49 -0
  30. results/distiluse-base-multilingual-cased-v2/TwitterURLCorpus.json +49 -0
results/distiluse-base-multilingual-cased-v2/AmazonCounterfactualClassification.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e8379541af4e31359cca9fbcf4b00f2671dba205",
3
+ "mteb_dataset_name": "AmazonCounterfactualClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "de": {
7
+ "accuracy": 0.6813704496788009,
8
+ "accuracy_stderr": 0.024612230046488805,
9
+ "ap": 0.806706553308835,
10
+ "ap_stderr": 0.016448423566835353,
11
+ "f1": 0.666468090116337,
12
+ "f1_stderr": 0.023789062068309465,
13
+ "main_score": 0.6813704496788009
14
+ },
15
+ "en": {
16
+ "accuracy": 0.7180597014925373,
17
+ "accuracy_stderr": 0.029021696280176958,
18
+ "ap": 0.3370263085714158,
19
+ "ap_stderr": 0.02417710728701376,
20
+ "f1": 0.6544989712268762,
21
+ "f1_stderr": 0.02527413136493158,
22
+ "main_score": 0.7180597014925373
23
+ },
24
+ "en-ext": {
25
+ "accuracy": 0.7296101949025486,
26
+ "accuracy_stderr": 0.03403669840829533,
27
+ "ap": 0.22209148737301962,
28
+ "ap_stderr": 0.02260533582227882,
29
+ "f1": 0.6042877542046691,
30
+ "f1_stderr": 0.02743732409029803,
31
+ "main_score": 0.7296101949025486
32
+ },
33
+ "evaluation_time": 18.51,
34
+ "ja": {
35
+ "accuracy": 0.6538543897216275,
36
+ "accuracy_stderr": 0.023329909403407633,
37
+ "ap": 0.16135900323284472,
38
+ "ap_stderr": 0.010525937964819992,
39
+ "f1": 0.5320720298606364,
40
+ "f1_stderr": 0.015164867003193814,
41
+ "main_score": 0.6538543897216275
42
+ }
43
+ },
44
+ "validation": {
45
+ "de": {
46
+ "accuracy": 0.7025751072961374,
47
+ "accuracy_stderr": 0.03413568897021146,
48
+ "ap": 0.8195432721434586,
49
+ "ap_stderr": 0.01855963604439618,
50
+ "f1": 0.6865634929258071,
51
+ "f1_stderr": 0.03146723543038631,
52
+ "main_score": 0.7025751072961374
53
+ },
54
+ "en": {
55
+ "accuracy": 0.7,
56
+ "accuracy_stderr": 0.03497126720555219,
57
+ "ap": 0.2959161477290448,
58
+ "ap_stderr": 0.029220638413959707,
59
+ "f1": 0.6269667251442652,
60
+ "f1_stderr": 0.029306318263514188,
61
+ "main_score": 0.7
62
+ },
63
+ "en-ext": {
64
+ "accuracy": 0.6995495495495495,
65
+ "accuracy_stderr": 0.040297626151526884,
66
+ "ap": 0.19518020632989744,
67
+ "ap_stderr": 0.027017350830205716,
68
+ "f1": 0.574144858941869,
69
+ "f1_stderr": 0.03304127410974345,
70
+ "main_score": 0.6995495495495495
71
+ },
72
+ "evaluation_time": 21.33,
73
+ "ja": {
74
+ "accuracy": 0.6564377682403434,
75
+ "accuracy_stderr": 0.030821457446519807,
76
+ "ap": 0.14325859760128967,
77
+ "ap_stderr": 0.013381854116899206,
78
+ "f1": 0.5201054934403554,
79
+ "f1_stderr": 0.019875056929254228,
80
+ "main_score": 0.6564377682403434
81
+ }
82
+ }
83
+ }
results/distiluse-base-multilingual-cased-v2/AmazonPolarityClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e2d317d38cd51312af73b3d32a06d1a08b442046",
3
+ "mteb_dataset_name": "AmazonPolarityClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "accuracy": 0.679988,
7
+ "accuracy_stderr": 0.03355476444932969,
8
+ "ap": 0.6259891275364823,
9
+ "ap_stderr": 0.033532892007831494,
10
+ "evaluation_time": 457.04,
11
+ "f1": 0.6773408963897285,
12
+ "f1_stderr": 0.03419161730008382,
13
+ "main_score": 0.679988
14
+ }
15
+ }
results/distiluse-base-multilingual-cased-v2/AmazonReviewsClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "1399c76144fd37290681b995c656ef9b2e06e26d",
3
+ "mteb_dataset_name": "AmazonReviewsClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "de": {
7
+ "accuracy": 0.35032,
8
+ "accuracy_stderr": 0.024274793510965235,
9
+ "f1": 0.33939764470643535,
10
+ "f1_stderr": 0.02282070229582006,
11
+ "main_score": 0.35032
12
+ },
13
+ "en": {
14
+ "accuracy": 0.35454,
15
+ "accuracy_stderr": 0.022858880112551442,
16
+ "f1": 0.35019589142407004,
17
+ "f1_stderr": 0.02054968398529996,
18
+ "main_score": 0.35454
19
+ },
20
+ "es": {
21
+ "accuracy": 0.36242,
22
+ "accuracy_stderr": 0.010545691063178358,
23
+ "f1": 0.3498879083946539,
24
+ "f1_stderr": 0.01321622772317041,
25
+ "main_score": 0.36242
26
+ },
27
+ "evaluation_time": 62.08,
28
+ "fr": {
29
+ "accuracy": 0.357,
30
+ "accuracy_stderr": 0.022324336496299283,
31
+ "f1": 0.3474911268048424,
32
+ "f1_stderr": 0.022703709419436332,
33
+ "main_score": 0.357
34
+ },
35
+ "ja": {
36
+ "accuracy": 0.31076,
37
+ "accuracy_stderr": 0.01528968279592484,
38
+ "f1": 0.30525865114811995,
39
+ "f1_stderr": 0.01261552083352967,
40
+ "main_score": 0.31076
41
+ },
42
+ "zh": {
43
+ "accuracy": 0.33894,
44
+ "accuracy_stderr": 0.013566149048274542,
45
+ "f1": 0.32638513658296125,
46
+ "f1_stderr": 0.01590406062662379,
47
+ "main_score": 0.33894
48
+ }
49
+ },
50
+ "validation": {
51
+ "de": {
52
+ "accuracy": 0.35140000000000005,
53
+ "accuracy_stderr": 0.03167813125801457,
54
+ "f1": 0.33995664837547496,
55
+ "f1_stderr": 0.029651988528941266,
56
+ "main_score": 0.35140000000000005
57
+ },
58
+ "en": {
59
+ "accuracy": 0.35572,
60
+ "accuracy_stderr": 0.02285260597831241,
61
+ "f1": 0.35182678284471613,
62
+ "f1_stderr": 0.021081362317817858,
63
+ "main_score": 0.35572
64
+ },
65
+ "es": {
66
+ "accuracy": 0.36605999999999994,
67
+ "accuracy_stderr": 0.01066547701699273,
68
+ "f1": 0.35388988527330134,
69
+ "f1_stderr": 0.011414525934042363,
70
+ "main_score": 0.36605999999999994
71
+ },
72
+ "evaluation_time": 64.9,
73
+ "fr": {
74
+ "accuracy": 0.35146,
75
+ "accuracy_stderr": 0.021103374137800814,
76
+ "f1": 0.3420240760686955,
77
+ "f1_stderr": 0.020450899496203788,
78
+ "main_score": 0.35146
79
+ },
80
+ "ja": {
81
+ "accuracy": 0.31432,
82
+ "accuracy_stderr": 0.01748123565426655,
83
+ "f1": 0.3085361473217993,
84
+ "f1_stderr": 0.01532611041574829,
85
+ "main_score": 0.31432
86
+ },
87
+ "zh": {
88
+ "accuracy": 0.33161999999999997,
89
+ "accuracy_stderr": 0.011498330313571618,
90
+ "f1": 0.31936819236765673,
91
+ "f1_stderr": 0.011252640258108498,
92
+ "main_score": 0.33161999999999997
93
+ }
94
+ }
95
+ }
results/distiluse-base-multilingual-cased-v2/ArxivClusteringP2P.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "a122ad7f3f0291bf49cc6f4d32aa80929df69d5d",
3
+ "mteb_dataset_name": "ArxivClusteringP2P",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "evaluation_time": 1142.85,
7
+ "v_measure": 0.33593722530350373,
8
+ "v_measure_std": 0.13938796887848565
9
+ }
10
+ }
results/distiluse-base-multilingual-cased-v2/AskUbuntuDupQuestions.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "2000358ca161889fa9c082cb41daa8dcfb161a54",
3
+ "mteb_dataset_name": "AskUbuntuDupQuestions",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "evaluation_time": 3.23,
7
+ "map": 0.5375229202972581,
8
+ "mrr": 0.6826968737633557
9
+ }
10
+ }
results/distiluse-base-multilingual-cased-v2/BIOSSES.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d3fb88f8f02e40887cd149695127462bbcf29b4a",
3
+ "mteb_dataset_name": "BIOSSES",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.7926094784825986,
8
+ "spearman": 0.7834033925464169
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.7743607353262967,
12
+ "spearman": 0.7677765304536669
13
+ },
14
+ "evaluation_time": 1.56,
15
+ "manhattan": {
16
+ "pearson": 0.7743287991423313,
17
+ "spearman": 0.76849341425823
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/Banking77Classification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "0fd18e25b25c072e09e0d92ab615fda904d66300",
3
+ "mteb_dataset_name": "Banking77Classification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "accuracy": 0.7148051948051949,
7
+ "accuracy_stderr": 0.007638086209107858,
8
+ "evaluation_time": 9.79,
9
+ "f1": 0.704571388461755,
10
+ "f1_stderr": 0.009624885542777998,
11
+ "main_score": 0.7148051948051949
12
+ }
13
+ }
results/distiluse-base-multilingual-cased-v2/EmotionClassification.json ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4f58c6b202a23cf9a4da393831edf4f9183cad37",
3
+ "mteb_dataset_name": "EmotionClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "accuracy": 0.40045000000000003,
7
+ "accuracy_stderr": 0.020835606542647128,
8
+ "evaluation_time": 1.47,
9
+ "f1": 0.3659544493168501,
10
+ "f1_stderr": 0.018558428267007415,
11
+ "main_score": 0.40045000000000003
12
+ },
13
+ "validation": {
14
+ "accuracy": 0.40679999999999994,
15
+ "accuracy_stderr": 0.02049536532975199,
16
+ "evaluation_time": 1.45,
17
+ "f1": 0.37603848849514077,
18
+ "f1_stderr": 0.018090906704553725,
19
+ "main_score": 0.40679999999999994
20
+ }
21
+ }
results/distiluse-base-multilingual-cased-v2/ImdbClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3d86128a09e091d6018b6d26cad27f2739fc2db7",
3
+ "mteb_dataset_name": "ImdbClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "accuracy": 0.6151679999999999,
7
+ "accuracy_stderr": 0.03937559772244734,
8
+ "ap": 0.5730211495623951,
9
+ "ap_stderr": 0.030273556660917186,
10
+ "evaluation_time": 36.83,
11
+ "f1": 0.6124392423075582,
12
+ "f1_stderr": 0.04030755169060487,
13
+ "main_score": 0.6151679999999999
14
+ }
15
+ }
results/distiluse-base-multilingual-cased-v2/MTOPDomainClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d80d48c1eb48d3562165c59d59d0034df9fff0bf",
3
+ "mteb_dataset_name": "MTOPDomainClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "de": {
7
+ "accuracy": 0.8619047619047618,
8
+ "accuracy_stderr": 0.009279975268553665,
9
+ "f1": 0.8442185095665184,
10
+ "f1_stderr": 0.01099157459261156,
11
+ "main_score": 0.8619047619047618
12
+ },
13
+ "en": {
14
+ "accuracy": 0.9159142726858185,
15
+ "accuracy_stderr": 0.005747905916342653,
16
+ "f1": 0.9116731589297895,
17
+ "f1_stderr": 0.006083518480038355,
18
+ "main_score": 0.9159142726858185
19
+ },
20
+ "es": {
21
+ "accuracy": 0.8774516344229486,
22
+ "accuracy_stderr": 0.009217490654903758,
23
+ "f1": 0.8689629934160831,
24
+ "f1_stderr": 0.010461264595061925,
25
+ "main_score": 0.8774516344229486
26
+ },
27
+ "evaluation_time": 9.29,
28
+ "fr": {
29
+ "accuracy": 0.8461321641089883,
30
+ "accuracy_stderr": 0.008683815611798849,
31
+ "f1": 0.8386194715158407,
32
+ "f1_stderr": 0.00994551331915671,
33
+ "main_score": 0.8461321641089883
34
+ },
35
+ "hi": {
36
+ "accuracy": 0.7641448547866619,
37
+ "accuracy_stderr": 0.009813835549335584,
38
+ "f1": 0.7466143814759416,
39
+ "f1_stderr": 0.011321537828004031,
40
+ "main_score": 0.7641448547866619
41
+ },
42
+ "th": {
43
+ "accuracy": 0.7361663652802893,
44
+ "accuracy_stderr": 0.009496339933087771,
45
+ "f1": 0.7159773512640322,
46
+ "f1_stderr": 0.013316284527249577,
47
+ "main_score": 0.7361663652802893
48
+ }
49
+ },
50
+ "validation": {
51
+ "de": {
52
+ "accuracy": 0.8556473829201103,
53
+ "accuracy_stderr": 0.008056144011651441,
54
+ "f1": 0.8387377179851955,
55
+ "f1_stderr": 0.010951159819488653,
56
+ "main_score": 0.8556473829201103
57
+ },
58
+ "en": {
59
+ "accuracy": 0.9127516778523491,
60
+ "accuracy_stderr": 0.007890466258858354,
61
+ "f1": 0.9110037602177025,
62
+ "f1_stderr": 0.008500038055347383,
63
+ "main_score": 0.9127516778523491
64
+ },
65
+ "es": {
66
+ "accuracy": 0.8756385068762278,
67
+ "accuracy_stderr": 0.01476443206023662,
68
+ "f1": 0.8712497428347236,
69
+ "f1_stderr": 0.016363301926595412,
70
+ "main_score": 0.8756385068762278
71
+ },
72
+ "evaluation_time": 6.95,
73
+ "fr": {
74
+ "accuracy": 0.8384908053265694,
75
+ "accuracy_stderr": 0.01185660167420024,
76
+ "f1": 0.8351704915222973,
77
+ "f1_stderr": 0.01309398488848049,
78
+ "main_score": 0.8384908053265694
79
+ },
80
+ "hi": {
81
+ "accuracy": 0.7288270377733598,
82
+ "accuracy_stderr": 0.011434004373595383,
83
+ "f1": 0.7182544532441011,
84
+ "f1_stderr": 0.012353541684199358,
85
+ "main_score": 0.7288270377733598
86
+ },
87
+ "th": {
88
+ "accuracy": 0.7256134051466188,
89
+ "accuracy_stderr": 0.007217393977886261,
90
+ "f1": 0.7120168370041309,
91
+ "f1_stderr": 0.011166807351882722,
92
+ "main_score": 0.7256134051466188
93
+ }
94
+ }
95
+ }
results/distiluse-base-multilingual-cased-v2/MTOPIntentClassification.json ADDED
@@ -0,0 +1,95 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba",
3
+ "mteb_dataset_name": "MTOPIntentClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "de": {
7
+ "accuracy": 0.5920822766976612,
8
+ "accuracy_stderr": 0.02474701600685877,
9
+ "f1": 0.36594153749624536,
10
+ "f1_stderr": 0.010942758271326127,
11
+ "main_score": 0.5920822766976612
12
+ },
13
+ "en": {
14
+ "accuracy": 0.6640218878248975,
15
+ "accuracy_stderr": 0.01798938840121496,
16
+ "f1": 0.440157655128108,
17
+ "f1_stderr": 0.009354607672096033,
18
+ "main_score": 0.6640218878248975
19
+ },
20
+ "es": {
21
+ "accuracy": 0.5721147431621081,
22
+ "accuracy_stderr": 0.014623897660868575,
23
+ "f1": 0.3846167201793877,
24
+ "f1_stderr": 0.012539758475101621,
25
+ "main_score": 0.5721147431621081
26
+ },
27
+ "evaluation_time": 90.97,
28
+ "fr": {
29
+ "accuracy": 0.5340745380519887,
30
+ "accuracy_stderr": 0.027571336367074026,
31
+ "f1": 0.3687813951228687,
32
+ "f1_stderr": 0.0069233558553558824,
33
+ "main_score": 0.5340745380519887
34
+ },
35
+ "hi": {
36
+ "accuracy": 0.4554320544998208,
37
+ "accuracy_stderr": 0.026878830594223708,
38
+ "f1": 0.2809108688148479,
39
+ "f1_stderr": 0.010259727365242678,
40
+ "main_score": 0.4554320544998208
41
+ },
42
+ "th": {
43
+ "accuracy": 0.4773236889692586,
44
+ "accuracy_stderr": 0.024172988197644237,
45
+ "f1": 0.2987429451601028,
46
+ "f1_stderr": 0.009367086874615469,
47
+ "main_score": 0.4773236889692586
48
+ }
49
+ },
50
+ "validation": {
51
+ "de": {
52
+ "accuracy": 0.5873278236914601,
53
+ "accuracy_stderr": 0.023941280077355232,
54
+ "f1": 0.36695799030694254,
55
+ "f1_stderr": 0.016786355460051364,
56
+ "main_score": 0.5873278236914601
57
+ },
58
+ "en": {
59
+ "accuracy": 0.6718568232662192,
60
+ "accuracy_stderr": 0.017732125620084332,
61
+ "f1": 0.43355788345559115,
62
+ "f1_stderr": 0.015160468260757431,
63
+ "main_score": 0.6718568232662192
64
+ },
65
+ "es": {
66
+ "accuracy": 0.5925343811394892,
67
+ "accuracy_stderr": 0.014038867441106525,
68
+ "f1": 0.399761961435884,
69
+ "f1_stderr": 0.005740116482098513,
70
+ "main_score": 0.5925343811394892
71
+ },
72
+ "evaluation_time": 88.26,
73
+ "fr": {
74
+ "accuracy": 0.5419784400760939,
75
+ "accuracy_stderr": 0.020250135790377707,
76
+ "f1": 0.3466031643658519,
77
+ "f1_stderr": 0.009462256492570287,
78
+ "main_score": 0.5419784400760939
79
+ },
80
+ "hi": {
81
+ "accuracy": 0.45248508946322064,
82
+ "accuracy_stderr": 0.020827599876121216,
83
+ "f1": 0.29262307534575477,
84
+ "f1_stderr": 0.007752611943651198,
85
+ "main_score": 0.45248508946322064
86
+ },
87
+ "th": {
88
+ "accuracy": 0.4751645721125075,
89
+ "accuracy_stderr": 0.022562213559374794,
90
+ "f1": 0.29277985694657527,
91
+ "f1_stderr": 0.012170368969005575,
92
+ "main_score": 0.4751645721125075
93
+ }
94
+ }
95
+ }
results/distiluse-base-multilingual-cased-v2/MassiveIntentClassification.json ADDED
@@ -0,0 +1,725 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "af": {
7
+ "accuracy": 0.4002017484868864,
8
+ "accuracy_stderr": 0.014701207345584978,
9
+ "f1": 0.35758596987693575,
10
+ "f1_stderr": 0.010911285642381738,
11
+ "main_score": 0.4002017484868864
12
+ },
13
+ "am": {
14
+ "accuracy": 0.023470073974445192,
15
+ "accuracy_stderr": 0.0171629882985429,
16
+ "f1": 0.0074653906995346036,
17
+ "f1_stderr": 0.0037882316632745893,
18
+ "main_score": 0.023470073974445192
19
+ },
20
+ "ar": {
21
+ "accuracy": 0.43143913920645593,
22
+ "accuracy_stderr": 0.010804432555558659,
23
+ "f1": 0.38855586375920464,
24
+ "f1_stderr": 0.009412718179704151,
25
+ "main_score": 0.43143913920645593
26
+ },
27
+ "az": {
28
+ "accuracy": 0.25601882985877605,
29
+ "accuracy_stderr": 0.009621849522439484,
30
+ "f1": 0.25205774742990256,
31
+ "f1_stderr": 0.010246663660086893,
32
+ "main_score": 0.25601882985877605
33
+ },
34
+ "bn": {
35
+ "accuracy": 0.0484196368527236,
36
+ "accuracy_stderr": 0.012225728959447102,
37
+ "f1": 0.017486302624639155,
38
+ "f1_stderr": 0.0023696188307934284,
39
+ "main_score": 0.0484196368527236
40
+ },
41
+ "cy": {
42
+ "accuracy": 0.1543375924680565,
43
+ "accuracy_stderr": 0.006123028723653011,
44
+ "f1": 0.14212012285498213,
45
+ "f1_stderr": 0.007508066255760151,
46
+ "main_score": 0.1543375924680565
47
+ },
48
+ "da": {
49
+ "accuracy": 0.5233355749831876,
50
+ "accuracy_stderr": 0.012862246197400182,
51
+ "f1": 0.4818484932318873,
52
+ "f1_stderr": 0.011503482756025075,
53
+ "main_score": 0.5233355749831876
54
+ },
55
+ "de": {
56
+ "accuracy": 0.5157363819771351,
57
+ "accuracy_stderr": 0.021238919642934107,
58
+ "f1": 0.4555934579164648,
59
+ "f1_stderr": 0.017404603288308545,
60
+ "main_score": 0.5157363819771351
61
+ },
62
+ "el": {
63
+ "accuracy": 0.4965366509751178,
64
+ "accuracy_stderr": 0.019028191014635615,
65
+ "f1": 0.4564683808611846,
66
+ "f1_stderr": 0.014785178277635792,
67
+ "main_score": 0.4965366509751178
68
+ },
69
+ "en": {
70
+ "accuracy": 0.6671149966375253,
71
+ "accuracy_stderr": 0.0195925839780822,
72
+ "f1": 0.6378255507050109,
73
+ "f1_stderr": 0.014859849026916973,
74
+ "main_score": 0.6671149966375253
75
+ },
76
+ "es": {
77
+ "accuracy": 0.5657363819771352,
78
+ "accuracy_stderr": 0.01731021481584545,
79
+ "f1": 0.5498029542986489,
80
+ "f1_stderr": 0.017657418197825583,
81
+ "main_score": 0.5657363819771352
82
+ },
83
+ "evaluation_time": 669.8,
84
+ "fa": {
85
+ "accuracy": 0.5535642232683256,
86
+ "accuracy_stderr": 0.013824679202554693,
87
+ "f1": 0.5020214626269123,
88
+ "f1_stderr": 0.014869990851833466,
89
+ "main_score": 0.5535642232683256
90
+ },
91
+ "fi": {
92
+ "accuracy": 0.4571620712844654,
93
+ "accuracy_stderr": 0.017153367751714842,
94
+ "f1": 0.42200836560817534,
95
+ "f1_stderr": 0.01296679026779231,
96
+ "main_score": 0.4571620712844654
97
+ },
98
+ "fr": {
99
+ "accuracy": 0.5702084734364491,
100
+ "accuracy_stderr": 0.01792536535927726,
101
+ "f1": 0.5391065067115182,
102
+ "f1_stderr": 0.015644457381657793,
103
+ "main_score": 0.5702084734364491
104
+ },
105
+ "he": {
106
+ "accuracy": 0.467350369872226,
107
+ "accuracy_stderr": 0.015376071802839134,
108
+ "f1": 0.42509857120773864,
109
+ "f1_stderr": 0.01503573866426814,
110
+ "main_score": 0.467350369872226
111
+ },
112
+ "hi": {
113
+ "accuracy": 0.48550773369199735,
114
+ "accuracy_stderr": 0.025697074177017942,
115
+ "f1": 0.4399327544348294,
116
+ "f1_stderr": 0.01557781435915338,
117
+ "main_score": 0.48550773369199735
118
+ },
119
+ "hu": {
120
+ "accuracy": 0.506455951580363,
121
+ "accuracy_stderr": 0.01624099685011615,
122
+ "f1": 0.4528464736653043,
123
+ "f1_stderr": 0.017640921589272128,
124
+ "main_score": 0.506455951580363
125
+ },
126
+ "hy": {
127
+ "accuracy": 0.4079354404841964,
128
+ "accuracy_stderr": 0.017714722185914075,
129
+ "f1": 0.3690100598587695,
130
+ "f1_stderr": 0.015501526953292766,
131
+ "main_score": 0.4079354404841964
132
+ },
133
+ "id": {
134
+ "accuracy": 0.55995292535306,
135
+ "accuracy_stderr": 0.015307239433639182,
136
+ "f1": 0.5244999289764702,
137
+ "f1_stderr": 0.020131500580668243,
138
+ "main_score": 0.55995292535306
139
+ },
140
+ "is": {
141
+ "accuracy": 0.16079354404841964,
142
+ "accuracy_stderr": 0.007161045103622155,
143
+ "f1": 0.14926428149458182,
144
+ "f1_stderr": 0.004459678207251405,
145
+ "main_score": 0.16079354404841964
146
+ },
147
+ "it": {
148
+ "accuracy": 0.5764626765299259,
149
+ "accuracy_stderr": 0.02105252952503758,
150
+ "f1": 0.537737970315679,
151
+ "f1_stderr": 0.017787669402883663,
152
+ "main_score": 0.5764626765299259
153
+ },
154
+ "ja": {
155
+ "accuracy": 0.5532952252858104,
156
+ "accuracy_stderr": 0.018399989226628034,
157
+ "f1": 0.5089055472943818,
158
+ "f1_stderr": 0.01535183000717043,
159
+ "main_score": 0.5532952252858104
160
+ },
161
+ "jv": {
162
+ "accuracy": 0.2816408876933423,
163
+ "accuracy_stderr": 0.014997692973435263,
164
+ "f1": 0.25896264320477325,
165
+ "f1_stderr": 0.012251397477138987,
166
+ "main_score": 0.2816408876933423
167
+ },
168
+ "ka": {
169
+ "accuracy": 0.2941156691324815,
170
+ "accuracy_stderr": 0.010098682352619555,
171
+ "f1": 0.26845594782987,
172
+ "f1_stderr": 0.008974618076973445,
173
+ "main_score": 0.2941156691324815
174
+ },
175
+ "km": {
176
+ "accuracy": 0.047915265635507735,
177
+ "accuracy_stderr": 0.012472243464149484,
178
+ "f1": 0.014491239093711444,
179
+ "f1_stderr": 0.0041633995297136114,
180
+ "main_score": 0.047915265635507735
181
+ },
182
+ "kn": {
183
+ "accuracy": 0.03365837256220579,
184
+ "accuracy_stderr": 0.019705048968397948,
185
+ "f1": 0.013064783225018712,
186
+ "f1_stderr": 0.004541200395152722,
187
+ "main_score": 0.03365837256220579
188
+ },
189
+ "ko": {
190
+ "accuracy": 0.499663752521856,
191
+ "accuracy_stderr": 0.02604605175112272,
192
+ "f1": 0.4628463081207797,
193
+ "f1_stderr": 0.02180767402572388,
194
+ "main_score": 0.499663752521856
195
+ },
196
+ "lv": {
197
+ "accuracy": 0.4431405514458642,
198
+ "accuracy_stderr": 0.021649977423000316,
199
+ "f1": 0.4159880687298492,
200
+ "f1_stderr": 0.016865284059111217,
201
+ "main_score": 0.4431405514458642
202
+ },
203
+ "ml": {
204
+ "accuracy": 0.032447881640887695,
205
+ "accuracy_stderr": 0.017969116826796732,
206
+ "f1": 0.011130430676330431,
207
+ "f1_stderr": 0.002741915892112217,
208
+ "main_score": 0.032447881640887695
209
+ },
210
+ "mn": {
211
+ "accuracy": 0.4036650975117687,
212
+ "accuracy_stderr": 0.012931546437213906,
213
+ "f1": 0.36405182949755555,
214
+ "f1_stderr": 0.01775983664020273,
215
+ "main_score": 0.4036650975117687
216
+ },
217
+ "ms": {
218
+ "accuracy": 0.47969065232010755,
219
+ "accuracy_stderr": 0.011198807949105707,
220
+ "f1": 0.43564724873023736,
221
+ "f1_stderr": 0.017839068381719857,
222
+ "main_score": 0.47969065232010755
223
+ },
224
+ "my": {
225
+ "accuracy": 0.38483523873570946,
226
+ "accuracy_stderr": 0.021052019322452955,
227
+ "f1": 0.33325537301233815,
228
+ "f1_stderr": 0.01907728575472844,
229
+ "main_score": 0.38483523873570946
230
+ },
231
+ "nb": {
232
+ "accuracy": 0.46008742434431743,
233
+ "accuracy_stderr": 0.0152361668946242,
234
+ "f1": 0.431074675107609,
235
+ "f1_stderr": 0.012629106265396966,
236
+ "main_score": 0.46008742434431743
237
+ },
238
+ "nl": {
239
+ "accuracy": 0.5829186281102892,
240
+ "accuracy_stderr": 0.016563183531898014,
241
+ "f1": 0.5338326950257227,
242
+ "f1_stderr": 0.010833144067549285,
243
+ "main_score": 0.5829186281102892
244
+ },
245
+ "pl": {
246
+ "accuracy": 0.5310020174848689,
247
+ "accuracy_stderr": 0.01972245530891041,
248
+ "f1": 0.48491009241597,
249
+ "f1_stderr": 0.019935790515877495,
250
+ "main_score": 0.5310020174848689
251
+ },
252
+ "pt": {
253
+ "accuracy": 0.5862811028917283,
254
+ "accuracy_stderr": 0.01736691370178251,
255
+ "f1": 0.5639037901287144,
256
+ "f1_stderr": 0.013980722069024997,
257
+ "main_score": 0.5862811028917283
258
+ },
259
+ "ro": {
260
+ "accuracy": 0.5063214525891055,
261
+ "accuracy_stderr": 0.01702076630395707,
262
+ "f1": 0.47522720473016566,
263
+ "f1_stderr": 0.020584965674963217,
264
+ "main_score": 0.5063214525891055
265
+ },
266
+ "ru": {
267
+ "accuracy": 0.5795897780766645,
268
+ "accuracy_stderr": 0.026075140999766558,
269
+ "f1": 0.5379707075942384,
270
+ "f1_stderr": 0.01941934315504064,
271
+ "main_score": 0.5795897780766645
272
+ },
273
+ "sl": {
274
+ "accuracy": 0.5065904505716207,
275
+ "accuracy_stderr": 0.014787397993498005,
276
+ "f1": 0.4869839976207718,
277
+ "f1_stderr": 0.012648733132223209,
278
+ "main_score": 0.5065904505716207
279
+ },
280
+ "sq": {
281
+ "accuracy": 0.5025218560860794,
282
+ "accuracy_stderr": 0.02342428906124986,
283
+ "f1": 0.46925456055473524,
284
+ "f1_stderr": 0.0185376827548175,
285
+ "main_score": 0.5025218560860794
286
+ },
287
+ "sv": {
288
+ "accuracy": 0.5241089441829188,
289
+ "accuracy_stderr": 0.012996780611425651,
290
+ "f1": 0.4764228703598475,
291
+ "f1_stderr": 0.012919477781569308,
292
+ "main_score": 0.5241089441829188
293
+ },
294
+ "sw": {
295
+ "accuracy": 0.19293880295897783,
296
+ "accuracy_stderr": 0.010498031835835437,
297
+ "f1": 0.17665029718291048,
298
+ "f1_stderr": 0.014431535650922677,
299
+ "main_score": 0.19293880295897783
300
+ },
301
+ "ta": {
302
+ "accuracy": 0.037861466039004706,
303
+ "accuracy_stderr": 0.01833224211197537,
304
+ "f1": 0.012869466371674323,
305
+ "f1_stderr": 0.003941552331604661,
306
+ "main_score": 0.037861466039004706
307
+ },
308
+ "te": {
309
+ "accuracy": 0.033591123066577,
310
+ "accuracy_stderr": 0.020711561851646545,
311
+ "f1": 0.013191646312270083,
312
+ "f1_stderr": 0.0027599921687235835,
313
+ "main_score": 0.033591123066577
314
+ },
315
+ "th": {
316
+ "accuracy": 0.45279085406859443,
317
+ "accuracy_stderr": 0.01940949066478539,
318
+ "f1": 0.42542426590317606,
319
+ "f1_stderr": 0.011877154683154497,
320
+ "main_score": 0.45279085406859443
321
+ },
322
+ "tl": {
323
+ "accuracy": 0.2843981170141224,
324
+ "accuracy_stderr": 0.015648915728262866,
325
+ "f1": 0.2528322629101539,
326
+ "f1_stderr": 0.009620009795902596,
327
+ "main_score": 0.2843981170141224
328
+ },
329
+ "tr": {
330
+ "accuracy": 0.5047410894418292,
331
+ "accuracy_stderr": 0.016718263837633297,
332
+ "f1": 0.4718657479743079,
333
+ "f1_stderr": 0.015846704267558968,
334
+ "main_score": 0.5047410894418292
335
+ },
336
+ "ur": {
337
+ "accuracy": 0.4602891728312038,
338
+ "accuracy_stderr": 0.02104299474887776,
339
+ "f1": 0.4142008348263186,
340
+ "f1_stderr": 0.018044890348411873,
341
+ "main_score": 0.4602891728312038
342
+ },
343
+ "vi": {
344
+ "accuracy": 0.4525218560860794,
345
+ "accuracy_stderr": 0.018715306278617536,
346
+ "f1": 0.4169045540062304,
347
+ "f1_stderr": 0.0190581910778906,
348
+ "main_score": 0.4525218560860794
349
+ },
350
+ "zh-CN": {
351
+ "accuracy": 0.5921990585070611,
352
+ "accuracy_stderr": 0.01785901491028706,
353
+ "f1": 0.5621401131609249,
354
+ "f1_stderr": 0.014409990861033293,
355
+ "main_score": 0.5921990585070611
356
+ },
357
+ "zh-TW": {
358
+ "accuracy": 0.5496301277740416,
359
+ "accuracy_stderr": 0.014749885702127947,
360
+ "f1": 0.5302026835629304,
361
+ "f1_stderr": 0.015492526123127626,
362
+ "main_score": 0.5496301277740416
363
+ }
364
+ },
365
+ "validation": {
366
+ "af": {
367
+ "accuracy": 0.4120019675356616,
368
+ "accuracy_stderr": 0.020871417611446966,
369
+ "f1": 0.36145392548498007,
370
+ "f1_stderr": 0.013888398467252245,
371
+ "main_score": 0.4120019675356616
372
+ },
373
+ "am": {
374
+ "accuracy": 0.02311854402361043,
375
+ "accuracy_stderr": 0.014974613914007163,
376
+ "f1": 0.007014656918090101,
377
+ "f1_stderr": 0.0034908338077239715,
378
+ "main_score": 0.02311854402361043
379
+ },
380
+ "ar": {
381
+ "accuracy": 0.4417609444171176,
382
+ "accuracy_stderr": 0.014072436907924597,
383
+ "f1": 0.3881170569323368,
384
+ "f1_stderr": 0.01318644175976311,
385
+ "main_score": 0.4417609444171176
386
+ },
387
+ "az": {
388
+ "accuracy": 0.2530250860796852,
389
+ "accuracy_stderr": 0.008741705810732142,
390
+ "f1": 0.2459114891382888,
391
+ "f1_stderr": 0.010232094269761323,
392
+ "main_score": 0.2530250860796852
393
+ },
394
+ "bn": {
395
+ "accuracy": 0.04874569601574029,
396
+ "accuracy_stderr": 0.01074166809918121,
397
+ "f1": 0.017390724599598215,
398
+ "f1_stderr": 0.003920663445927572,
399
+ "main_score": 0.04874569601574029
400
+ },
401
+ "cy": {
402
+ "accuracy": 0.13910477127397933,
403
+ "accuracy_stderr": 0.007004122975436327,
404
+ "f1": 0.129285630296109,
405
+ "f1_stderr": 0.007442036390530482,
406
+ "main_score": 0.13910477127397933
407
+ },
408
+ "da": {
409
+ "accuracy": 0.5210034431874078,
410
+ "accuracy_stderr": 0.016661437714777846,
411
+ "f1": 0.4677736003573324,
412
+ "f1_stderr": 0.008626693012212788,
413
+ "main_score": 0.5210034431874078
414
+ },
415
+ "de": {
416
+ "accuracy": 0.5094441711756026,
417
+ "accuracy_stderr": 0.02184103060800012,
418
+ "f1": 0.44977656924608994,
419
+ "f1_stderr": 0.01959505466045921,
420
+ "main_score": 0.5094441711756026
421
+ },
422
+ "el": {
423
+ "accuracy": 0.5062961141170683,
424
+ "accuracy_stderr": 0.01840599114882109,
425
+ "f1": 0.4491963187895525,
426
+ "f1_stderr": 0.01596968429163176,
427
+ "main_score": 0.5062961141170683
428
+ },
429
+ "en": {
430
+ "accuracy": 0.6721101819970486,
431
+ "accuracy_stderr": 0.017233074177416555,
432
+ "f1": 0.6220559777901397,
433
+ "f1_stderr": 0.013446950242525011,
434
+ "main_score": 0.6721101819970486
435
+ },
436
+ "es": {
437
+ "accuracy": 0.5662567634038367,
438
+ "accuracy_stderr": 0.017408521598180578,
439
+ "f1": 0.5328709936798957,
440
+ "f1_stderr": 0.018839069406111273,
441
+ "main_score": 0.5662567634038367
442
+ },
443
+ "evaluation_time": 601.6,
444
+ "fa": {
445
+ "accuracy": 0.5536153467781604,
446
+ "accuracy_stderr": 0.02112153198353419,
447
+ "f1": 0.49142434977081856,
448
+ "f1_stderr": 0.019739996433365627,
449
+ "main_score": 0.5536153467781604
450
+ },
451
+ "fi": {
452
+ "accuracy": 0.44323659616330546,
453
+ "accuracy_stderr": 0.017847253964341434,
454
+ "f1": 0.4059362005111408,
455
+ "f1_stderr": 0.014461237890046785,
456
+ "main_score": 0.44323659616330546
457
+ },
458
+ "fr": {
459
+ "accuracy": 0.5809149040826365,
460
+ "accuracy_stderr": 0.014794180498404636,
461
+ "f1": 0.5239557490784114,
462
+ "f1_stderr": 0.01370522230554907,
463
+ "main_score": 0.5809149040826365
464
+ },
465
+ "he": {
466
+ "accuracy": 0.4782587309394983,
467
+ "accuracy_stderr": 0.013549683290970102,
468
+ "f1": 0.4196084670367144,
469
+ "f1_stderr": 0.013157712863369832,
470
+ "main_score": 0.4782587309394983
471
+ },
472
+ "hi": {
473
+ "accuracy": 0.4850467289719626,
474
+ "accuracy_stderr": 0.02479158370193768,
475
+ "f1": 0.4265658759964584,
476
+ "f1_stderr": 0.013901034065135496,
477
+ "main_score": 0.4850467289719626
478
+ },
479
+ "hu": {
480
+ "accuracy": 0.5049680275454993,
481
+ "accuracy_stderr": 0.017569553644234785,
482
+ "f1": 0.4481298212309654,
483
+ "f1_stderr": 0.01569835867865692,
484
+ "main_score": 0.5049680275454993
485
+ },
486
+ "hy": {
487
+ "accuracy": 0.40875553369404816,
488
+ "accuracy_stderr": 0.015450166959934376,
489
+ "f1": 0.3715471169379082,
490
+ "f1_stderr": 0.011091342062183296,
491
+ "main_score": 0.40875553369404816
492
+ },
493
+ "id": {
494
+ "accuracy": 0.5586325627151993,
495
+ "accuracy_stderr": 0.012385418199022624,
496
+ "f1": 0.5056957364850152,
497
+ "f1_stderr": 0.014391377325523784,
498
+ "main_score": 0.5586325627151993
499
+ },
500
+ "is": {
501
+ "accuracy": 0.1573536645351697,
502
+ "accuracy_stderr": 0.012041515161595732,
503
+ "f1": 0.14521472498886528,
504
+ "f1_stderr": 0.011861503949666357,
505
+ "main_score": 0.1573536645351697
506
+ },
507
+ "it": {
508
+ "accuracy": 0.5760452533202164,
509
+ "accuracy_stderr": 0.02101380820655486,
510
+ "f1": 0.5258911027334193,
511
+ "f1_stderr": 0.014509539362938024,
512
+ "main_score": 0.5760452533202164
513
+ },
514
+ "ja": {
515
+ "accuracy": 0.5591736350221349,
516
+ "accuracy_stderr": 0.01783281023908015,
517
+ "f1": 0.49724192784617377,
518
+ "f1_stderr": 0.014568198366260159,
519
+ "main_score": 0.5591736350221349
520
+ },
521
+ "jv": {
522
+ "accuracy": 0.29424495818986723,
523
+ "accuracy_stderr": 0.018384617846592517,
524
+ "f1": 0.2603663242964646,
525
+ "f1_stderr": 0.013204560186932114,
526
+ "main_score": 0.29424495818986723
527
+ },
528
+ "ka": {
529
+ "accuracy": 0.2925233644859813,
530
+ "accuracy_stderr": 0.009011146861252551,
531
+ "f1": 0.2563575480372773,
532
+ "f1_stderr": 0.00995838950989548,
533
+ "main_score": 0.2925233644859813
534
+ },
535
+ "km": {
536
+ "accuracy": 0.04328578455484505,
537
+ "accuracy_stderr": 0.013248032870121065,
538
+ "f1": 0.01244013357776507,
539
+ "f1_stderr": 0.0027584375918762293,
540
+ "main_score": 0.04328578455484505
541
+ },
542
+ "kn": {
543
+ "accuracy": 0.03128381701918347,
544
+ "accuracy_stderr": 0.016340913746472473,
545
+ "f1": 0.009704820794413562,
546
+ "f1_stderr": 0.003075494218871469,
547
+ "main_score": 0.03128381701918347
548
+ },
549
+ "ko": {
550
+ "accuracy": 0.5062469257255288,
551
+ "accuracy_stderr": 0.018807469512666777,
552
+ "f1": 0.45215767607746465,
553
+ "f1_stderr": 0.013931361419488378,
554
+ "main_score": 0.5062469257255288
555
+ },
556
+ "lv": {
557
+ "accuracy": 0.43861288735858334,
558
+ "accuracy_stderr": 0.01821173211024853,
559
+ "f1": 0.4027134346413036,
560
+ "f1_stderr": 0.014912739939672895,
561
+ "main_score": 0.43861288735858334
562
+ },
563
+ "ml": {
564
+ "accuracy": 0.03349729463846532,
565
+ "accuracy_stderr": 0.015348748722605358,
566
+ "f1": 0.011119396716472664,
567
+ "f1_stderr": 0.002591802705914599,
568
+ "main_score": 0.03349729463846532
569
+ },
570
+ "mn": {
571
+ "accuracy": 0.4052631578947368,
572
+ "accuracy_stderr": 0.014615575485092116,
573
+ "f1": 0.36149197901906255,
574
+ "f1_stderr": 0.02176874964688678,
575
+ "main_score": 0.4052631578947368
576
+ },
577
+ "ms": {
578
+ "accuracy": 0.4799803246433842,
579
+ "accuracy_stderr": 0.01757175686120304,
580
+ "f1": 0.42357843419648705,
581
+ "f1_stderr": 0.0172923493039438,
582
+ "main_score": 0.4799803246433842
583
+ },
584
+ "my": {
585
+ "accuracy": 0.3902115100836203,
586
+ "accuracy_stderr": 0.015130608289960105,
587
+ "f1": 0.3249161981054664,
588
+ "f1_stderr": 0.014391440834175201,
589
+ "main_score": 0.3902115100836203
590
+ },
591
+ "nb": {
592
+ "accuracy": 0.45966551893753077,
593
+ "accuracy_stderr": 0.01765753640854775,
594
+ "f1": 0.41905279718242844,
595
+ "f1_stderr": 0.016582591669814493,
596
+ "main_score": 0.45966551893753077
597
+ },
598
+ "nl": {
599
+ "accuracy": 0.572306935563207,
600
+ "accuracy_stderr": 0.016090319100326968,
601
+ "f1": 0.5075167280579722,
602
+ "f1_stderr": 0.01097151787552597,
603
+ "main_score": 0.572306935563207
604
+ },
605
+ "pl": {
606
+ "accuracy": 0.5271028037383177,
607
+ "accuracy_stderr": 0.021523167487578203,
608
+ "f1": 0.4695836687300233,
609
+ "f1_stderr": 0.01676894589885832,
610
+ "main_score": 0.5271028037383177
611
+ },
612
+ "pt": {
613
+ "accuracy": 0.5801278898180029,
614
+ "accuracy_stderr": 0.019279339708369262,
615
+ "f1": 0.5422489228364059,
616
+ "f1_stderr": 0.014525351146668216,
617
+ "main_score": 0.5801278898180029
618
+ },
619
+ "ro": {
620
+ "accuracy": 0.5144121987211019,
621
+ "accuracy_stderr": 0.01893695580476549,
622
+ "f1": 0.47275824704580466,
623
+ "f1_stderr": 0.015554742076960719,
624
+ "main_score": 0.5144121987211019
625
+ },
626
+ "ru": {
627
+ "accuracy": 0.5799311362518446,
628
+ "accuracy_stderr": 0.02329162533346036,
629
+ "f1": 0.5166613789510148,
630
+ "f1_stderr": 0.01675166420516213,
631
+ "main_score": 0.5799311362518446
632
+ },
633
+ "sl": {
634
+ "accuracy": 0.5054599114608952,
635
+ "accuracy_stderr": 0.015677444652719225,
636
+ "f1": 0.47305701895571695,
637
+ "f1_stderr": 0.009895492206845696,
638
+ "main_score": 0.5054599114608952
639
+ },
640
+ "sq": {
641
+ "accuracy": 0.5120019675356615,
642
+ "accuracy_stderr": 0.021249669798242804,
643
+ "f1": 0.4705432464565664,
644
+ "f1_stderr": 0.015240500781190793,
645
+ "main_score": 0.5120019675356615
646
+ },
647
+ "sv": {
648
+ "accuracy": 0.5299065420560748,
649
+ "accuracy_stderr": 0.013872608963761628,
650
+ "f1": 0.46812732983244476,
651
+ "f1_stderr": 0.012439929628433453,
652
+ "main_score": 0.5299065420560748
653
+ },
654
+ "sw": {
655
+ "accuracy": 0.19665518937530743,
656
+ "accuracy_stderr": 0.013788200173494919,
657
+ "f1": 0.17847054930985967,
658
+ "f1_stderr": 0.013167579039331419,
659
+ "main_score": 0.19665518937530743
660
+ },
661
+ "ta": {
662
+ "accuracy": 0.032562715199212985,
663
+ "accuracy_stderr": 0.016704945538395907,
664
+ "f1": 0.010751302470988326,
665
+ "f1_stderr": 0.00390288052147719,
666
+ "main_score": 0.032562715199212985
667
+ },
668
+ "te": {
669
+ "accuracy": 0.031529758976881456,
670
+ "accuracy_stderr": 0.017871639356760533,
671
+ "f1": 0.010151512907443582,
672
+ "f1_stderr": 0.002330514976502465,
673
+ "main_score": 0.031529758976881456
674
+ },
675
+ "th": {
676
+ "accuracy": 0.4613379242498771,
677
+ "accuracy_stderr": 0.017219379882016703,
678
+ "f1": 0.41250222768854944,
679
+ "f1_stderr": 0.011269402501639412,
680
+ "main_score": 0.4613379242498771
681
+ },
682
+ "tl": {
683
+ "accuracy": 0.27142154451549433,
684
+ "accuracy_stderr": 0.013093354823259564,
685
+ "f1": 0.23722399647556558,
686
+ "f1_stderr": 0.00852693664126537,
687
+ "main_score": 0.27142154451549433
688
+ },
689
+ "tr": {
690
+ "accuracy": 0.5085587801278898,
691
+ "accuracy_stderr": 0.013243374967765318,
692
+ "f1": 0.46137293259500256,
693
+ "f1_stderr": 0.008538476359620625,
694
+ "main_score": 0.5085587801278898
695
+ },
696
+ "ur": {
697
+ "accuracy": 0.46606000983767826,
698
+ "accuracy_stderr": 0.017645199981977652,
699
+ "f1": 0.4118480128911691,
700
+ "f1_stderr": 0.009361097824299693,
701
+ "main_score": 0.46606000983767826
702
+ },
703
+ "vi": {
704
+ "accuracy": 0.44220363994097395,
705
+ "accuracy_stderr": 0.0175568797811852,
706
+ "f1": 0.4058691699496233,
707
+ "f1_stderr": 0.016619806733593402,
708
+ "main_score": 0.44220363994097395
709
+ },
710
+ "zh-CN": {
711
+ "accuracy": 0.6034923757993113,
712
+ "accuracy_stderr": 0.018642919524976046,
713
+ "f1": 0.5556735374033064,
714
+ "f1_stderr": 0.014692692007771837,
715
+ "main_score": 0.6034923757993113
716
+ },
717
+ "zh-TW": {
718
+ "accuracy": 0.5613871126414167,
719
+ "accuracy_stderr": 0.011741610097811076,
720
+ "f1": 0.5280178788819753,
721
+ "f1_stderr": 0.013419322355917092,
722
+ "main_score": 0.5613871126414167
723
+ }
724
+ }
725
+ }
results/distiluse-base-multilingual-cased-v2/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,725 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "af": {
7
+ "accuracy": 0.5366509751176867,
8
+ "accuracy_stderr": 0.015901341973891016,
9
+ "f1": 0.48816628257216454,
10
+ "f1_stderr": 0.014395801425278739,
11
+ "main_score": 0.5366509751176867
12
+ },
13
+ "am": {
14
+ "accuracy": 0.07720242098184263,
15
+ "accuracy_stderr": 0.030644747333378176,
16
+ "f1": 0.03017236016204755,
17
+ "f1_stderr": 0.00973073374650384,
18
+ "main_score": 0.07720242098184263
19
+ },
20
+ "ar": {
21
+ "accuracy": 0.5218897108271687,
22
+ "accuracy_stderr": 0.012756548654062672,
23
+ "f1": 0.5236066873405811,
24
+ "f1_stderr": 0.012534635824162184,
25
+ "main_score": 0.5218897108271687
26
+ },
27
+ "az": {
28
+ "accuracy": 0.3474781439139206,
29
+ "accuracy_stderr": 0.03171197010724359,
30
+ "f1": 0.32559538526453335,
31
+ "f1_stderr": 0.030841772076578934,
32
+ "main_score": 0.3474781439139206
33
+ },
34
+ "bn": {
35
+ "accuracy": 0.10652320107599192,
36
+ "accuracy_stderr": 0.024095973828173443,
37
+ "f1": 0.06439785272600618,
38
+ "f1_stderr": 0.01103147994993239,
39
+ "main_score": 0.10652320107599192
40
+ },
41
+ "cy": {
42
+ "accuracy": 0.21237390719569604,
43
+ "accuracy_stderr": 0.01839012434279269,
44
+ "f1": 0.18428497244325157,
45
+ "f1_stderr": 0.012556052702419744,
46
+ "main_score": 0.21237390719569604
47
+ },
48
+ "da": {
49
+ "accuracy": 0.6254875588433088,
50
+ "accuracy_stderr": 0.009576620972832873,
51
+ "f1": 0.6069001958508912,
52
+ "f1_stderr": 0.010669671665209384,
53
+ "main_score": 0.6254875588433088
54
+ },
55
+ "de": {
56
+ "accuracy": 0.6140215198386013,
57
+ "accuracy_stderr": 0.019085384348152396,
58
+ "f1": 0.5807492599013545,
59
+ "f1_stderr": 0.015541602133835949,
60
+ "main_score": 0.6140215198386013
61
+ },
62
+ "el": {
63
+ "accuracy": 0.6067585743106927,
64
+ "accuracy_stderr": 0.017268362362388955,
65
+ "f1": 0.5805582762779206,
66
+ "f1_stderr": 0.016262604616914596,
67
+ "main_score": 0.6067585743106927
68
+ },
69
+ "en": {
70
+ "accuracy": 0.7400470746469401,
71
+ "accuracy_stderr": 0.01392198564741959,
72
+ "f1": 0.7222931856264793,
73
+ "f1_stderr": 0.013758791507299395,
74
+ "main_score": 0.7400470746469401
75
+ },
76
+ "es": {
77
+ "accuracy": 0.646133154001345,
78
+ "accuracy_stderr": 0.01809823743941933,
79
+ "f1": 0.6334590750295819,
80
+ "f1_stderr": 0.019074041482075442,
81
+ "main_score": 0.646133154001345
82
+ },
83
+ "evaluation_time": 82.19,
84
+ "fa": {
85
+ "accuracy": 0.592434431741762,
86
+ "accuracy_stderr": 0.017902423295707747,
87
+ "f1": 0.5740580117369346,
88
+ "f1_stderr": 0.017928770223828584,
89
+ "main_score": 0.592434431741762
90
+ },
91
+ "fi": {
92
+ "accuracy": 0.5466039004707464,
93
+ "accuracy_stderr": 0.020614575752124848,
94
+ "f1": 0.5145432689446743,
95
+ "f1_stderr": 0.019285744926834963,
96
+ "main_score": 0.5466039004707464
97
+ },
98
+ "fr": {
99
+ "accuracy": 0.6519502353732347,
100
+ "accuracy_stderr": 0.01624075319435103,
101
+ "f1": 0.6350200684075783,
102
+ "f1_stderr": 0.015770617934000787,
103
+ "main_score": 0.6519502353732347
104
+ },
105
+ "he": {
106
+ "accuracy": 0.5474445191661064,
107
+ "accuracy_stderr": 0.018020763159236206,
108
+ "f1": 0.5262150844808929,
109
+ "f1_stderr": 0.015758009400296105,
110
+ "main_score": 0.5474445191661064
111
+ },
112
+ "hi": {
113
+ "accuracy": 0.5598520511096167,
114
+ "accuracy_stderr": 0.018547942445541317,
115
+ "f1": 0.5370079438430524,
116
+ "f1_stderr": 0.017796424170267083,
117
+ "main_score": 0.5598520511096167
118
+ },
119
+ "hu": {
120
+ "accuracy": 0.6120040349697378,
121
+ "accuracy_stderr": 0.02150956483913394,
122
+ "f1": 0.585060672562612,
123
+ "f1_stderr": 0.023651019097465215,
124
+ "main_score": 0.6120040349697378
125
+ },
126
+ "hy": {
127
+ "accuracy": 0.4963349024882314,
128
+ "accuracy_stderr": 0.02435158466583094,
129
+ "f1": 0.4739478501763526,
130
+ "f1_stderr": 0.023771863635092094,
131
+ "main_score": 0.4963349024882314
132
+ },
133
+ "id": {
134
+ "accuracy": 0.6525218560860793,
135
+ "accuracy_stderr": 0.016652365679240923,
136
+ "f1": 0.6345266636240826,
137
+ "f1_stderr": 0.019474010101350445,
138
+ "main_score": 0.6525218560860793
139
+ },
140
+ "is": {
141
+ "accuracy": 0.22599193006052457,
142
+ "accuracy_stderr": 0.01937407081933763,
143
+ "f1": 0.2193829297740852,
144
+ "f1_stderr": 0.01808277976437254,
145
+ "main_score": 0.22599193006052457
146
+ },
147
+ "it": {
148
+ "accuracy": 0.6463349024882313,
149
+ "accuracy_stderr": 0.01840954175149463,
150
+ "f1": 0.6315345402734339,
151
+ "f1_stderr": 0.016259565883979928,
152
+ "main_score": 0.6463349024882313
153
+ },
154
+ "ja": {
155
+ "accuracy": 0.6232010759919301,
156
+ "accuracy_stderr": 0.020091848995397416,
157
+ "f1": 0.6002914271738089,
158
+ "f1_stderr": 0.02082671892733581,
159
+ "main_score": 0.6232010759919301
160
+ },
161
+ "jv": {
162
+ "accuracy": 0.3576664425016812,
163
+ "accuracy_stderr": 0.02286979751378726,
164
+ "f1": 0.33528305250648593,
165
+ "f1_stderr": 0.02005783697388505,
166
+ "main_score": 0.3576664425016812
167
+ },
168
+ "ka": {
169
+ "accuracy": 0.3908204438466712,
170
+ "accuracy_stderr": 0.023950664992710625,
171
+ "f1": 0.37312566552928733,
172
+ "f1_stderr": 0.019980991429658132,
173
+ "main_score": 0.3908204438466712
174
+ },
175
+ "km": {
176
+ "accuracy": 0.09236718224613315,
177
+ "accuracy_stderr": 0.04011646614810752,
178
+ "f1": 0.0341684484979606,
179
+ "f1_stderr": 0.009919306743092654,
180
+ "main_score": 0.09236718224613315
181
+ },
182
+ "kn": {
183
+ "accuracy": 0.08278412911903162,
184
+ "accuracy_stderr": 0.0301873839867578,
185
+ "f1": 0.03941809480667743,
186
+ "f1_stderr": 0.007011730949554586,
187
+ "main_score": 0.08278412911903162
188
+ },
189
+ "ko": {
190
+ "accuracy": 0.5759583053127101,
191
+ "accuracy_stderr": 0.01768447572060904,
192
+ "f1": 0.5642188880877947,
193
+ "f1_stderr": 0.016231082520527726,
194
+ "main_score": 0.5759583053127101
195
+ },
196
+ "lv": {
197
+ "accuracy": 0.5172158708809683,
198
+ "accuracy_stderr": 0.023125320999401264,
199
+ "f1": 0.4990313684327526,
200
+ "f1_stderr": 0.02128846436844353,
201
+ "main_score": 0.5172158708809683
202
+ },
203
+ "ml": {
204
+ "accuracy": 0.08254875588433089,
205
+ "accuracy_stderr": 0.017756445829680056,
206
+ "f1": 0.040681340980956396,
207
+ "f1_stderr": 0.005018277082524067,
208
+ "main_score": 0.08254875588433089
209
+ },
210
+ "mn": {
211
+ "accuracy": 0.4720914593140552,
212
+ "accuracy_stderr": 0.022499454216212092,
213
+ "f1": 0.4478012101794023,
214
+ "f1_stderr": 0.019510541046287914,
215
+ "main_score": 0.4720914593140552
216
+ },
217
+ "ms": {
218
+ "accuracy": 0.5564559515803632,
219
+ "accuracy_stderr": 0.019409053776817852,
220
+ "f1": 0.5310457083056076,
221
+ "f1_stderr": 0.013315243908621295,
222
+ "main_score": 0.5564559515803632
223
+ },
224
+ "my": {
225
+ "accuracy": 0.43308675184936113,
226
+ "accuracy_stderr": 0.02576498358255783,
227
+ "f1": 0.4040654924373442,
228
+ "f1_stderr": 0.020089215344942325,
229
+ "main_score": 0.43308675184936113
230
+ },
231
+ "nb": {
232
+ "accuracy": 0.5498318762609281,
233
+ "accuracy_stderr": 0.022741767017734164,
234
+ "f1": 0.5422408282419106,
235
+ "f1_stderr": 0.02042799367334959,
236
+ "main_score": 0.5498318762609281
237
+ },
238
+ "nl": {
239
+ "accuracy": 0.6749159381304641,
240
+ "accuracy_stderr": 0.016135535169786802,
241
+ "f1": 0.6466608521628295,
242
+ "f1_stderr": 0.015738734557601008,
243
+ "main_score": 0.6749159381304641
244
+ },
245
+ "pl": {
246
+ "accuracy": 0.6128782784129119,
247
+ "accuracy_stderr": 0.015000595073118924,
248
+ "f1": 0.5936495517929654,
249
+ "f1_stderr": 0.015505031155806273,
250
+ "main_score": 0.6128782784129119
251
+ },
252
+ "pt": {
253
+ "accuracy": 0.6426361802286483,
254
+ "accuracy_stderr": 0.018739455382559184,
255
+ "f1": 0.6301306314842477,
256
+ "f1_stderr": 0.017048422475100907,
257
+ "main_score": 0.6426361802286483
258
+ },
259
+ "ro": {
260
+ "accuracy": 0.5802622730329523,
261
+ "accuracy_stderr": 0.019783271183030637,
262
+ "f1": 0.558928740774695,
263
+ "f1_stderr": 0.01934895845853177,
264
+ "main_score": 0.5802622730329523
265
+ },
266
+ "ru": {
267
+ "accuracy": 0.6541358439811701,
268
+ "accuracy_stderr": 0.018503512663978917,
269
+ "f1": 0.6415512608670189,
270
+ "f1_stderr": 0.018387117791678623,
271
+ "main_score": 0.6541358439811701
272
+ },
273
+ "sl": {
274
+ "accuracy": 0.5935776731674512,
275
+ "accuracy_stderr": 0.017431797009640954,
276
+ "f1": 0.5828447907816511,
277
+ "f1_stderr": 0.014626508146847813,
278
+ "main_score": 0.5935776731674512
279
+ },
280
+ "sq": {
281
+ "accuracy": 0.6268661735036987,
282
+ "accuracy_stderr": 0.014487584612425898,
283
+ "f1": 0.5949767603465277,
284
+ "f1_stderr": 0.014500363072857713,
285
+ "main_score": 0.6268661735036987
286
+ },
287
+ "sv": {
288
+ "accuracy": 0.6435104236718224,
289
+ "accuracy_stderr": 0.015096575638642027,
290
+ "f1": 0.6162298238070602,
291
+ "f1_stderr": 0.016182321593427144,
292
+ "main_score": 0.6435104236718224
293
+ },
294
+ "sw": {
295
+ "accuracy": 0.2512104909213181,
296
+ "accuracy_stderr": 0.022214458820611765,
297
+ "f1": 0.22063961287382483,
298
+ "f1_stderr": 0.022712298156057933,
299
+ "main_score": 0.2512104909213181
300
+ },
301
+ "ta": {
302
+ "accuracy": 0.0867182246133154,
303
+ "accuracy_stderr": 0.023271848044933815,
304
+ "f1": 0.04160922973001201,
305
+ "f1_stderr": 0.006326533657124594,
306
+ "main_score": 0.0867182246133154
307
+ },
308
+ "te": {
309
+ "accuracy": 0.07821116341627439,
310
+ "accuracy_stderr": 0.03038464386596996,
311
+ "f1": 0.0359600077788794,
312
+ "f1_stderr": 0.008230758197652708,
313
+ "main_score": 0.07821116341627439
314
+ },
315
+ "th": {
316
+ "accuracy": 0.5464694014794889,
317
+ "accuracy_stderr": 0.01484584975249916,
318
+ "f1": 0.515915867779775,
319
+ "f1_stderr": 0.01470123732412852,
320
+ "main_score": 0.5464694014794889
321
+ },
322
+ "tl": {
323
+ "accuracy": 0.3608607935440485,
324
+ "accuracy_stderr": 0.010065881502121135,
325
+ "f1": 0.3246731674317254,
326
+ "f1_stderr": 0.008370980952137726,
327
+ "main_score": 0.3608607935440485
328
+ },
329
+ "tr": {
330
+ "accuracy": 0.6089441829186282,
331
+ "accuracy_stderr": 0.021857637935698466,
332
+ "f1": 0.6011999627480401,
333
+ "f1_stderr": 0.020648940428186456,
334
+ "main_score": 0.6089441829186282
335
+ },
336
+ "ur": {
337
+ "accuracy": 0.547074646940148,
338
+ "accuracy_stderr": 0.01693833112729649,
339
+ "f1": 0.5246709289947394,
340
+ "f1_stderr": 0.015481476455338466,
341
+ "main_score": 0.547074646940148
342
+ },
343
+ "vi": {
344
+ "accuracy": 0.551546738399462,
345
+ "accuracy_stderr": 0.018859287455847015,
346
+ "f1": 0.5411090226223558,
347
+ "f1_stderr": 0.019093711420780545,
348
+ "main_score": 0.551546738399462
349
+ },
350
+ "zh-CN": {
351
+ "accuracy": 0.664357767316745,
352
+ "accuracy_stderr": 0.010951348590480296,
353
+ "f1": 0.6494684758602547,
354
+ "f1_stderr": 0.013318705749811505,
355
+ "main_score": 0.664357767316745
356
+ },
357
+ "zh-TW": {
358
+ "accuracy": 0.6288836583725623,
359
+ "accuracy_stderr": 0.015212402259172282,
360
+ "f1": 0.617106895387137,
361
+ "f1_stderr": 0.018893417788500195,
362
+ "main_score": 0.6288836583725623
363
+ }
364
+ },
365
+ "validation": {
366
+ "af": {
367
+ "accuracy": 0.5516478111165765,
368
+ "accuracy_stderr": 0.017372975013701834,
369
+ "f1": 0.5050450891974643,
370
+ "f1_stderr": 0.01936042284803411,
371
+ "main_score": 0.5516478111165765
372
+ },
373
+ "am": {
374
+ "accuracy": 0.07761928184948352,
375
+ "accuracy_stderr": 0.027552349650600525,
376
+ "f1": 0.029958740401687178,
377
+ "f1_stderr": 0.006791595031284069,
378
+ "main_score": 0.07761928184948352
379
+ },
380
+ "ar": {
381
+ "accuracy": 0.5229709788489917,
382
+ "accuracy_stderr": 0.018427996145927928,
383
+ "f1": 0.5202049451997024,
384
+ "f1_stderr": 0.018184461406029976,
385
+ "main_score": 0.5229709788489917
386
+ },
387
+ "az": {
388
+ "accuracy": 0.3485489424495819,
389
+ "accuracy_stderr": 0.027436542291176154,
390
+ "f1": 0.3313501601388159,
391
+ "f1_stderr": 0.027081612149106836,
392
+ "main_score": 0.3485489424495819
393
+ },
394
+ "bn": {
395
+ "accuracy": 0.10560747663551402,
396
+ "accuracy_stderr": 0.02282579876270162,
397
+ "f1": 0.06341165812684066,
398
+ "f1_stderr": 0.010400847066568866,
399
+ "main_score": 0.10560747663551402
400
+ },
401
+ "cy": {
402
+ "accuracy": 0.20659124446630592,
403
+ "accuracy_stderr": 0.02050748932504314,
404
+ "f1": 0.18541360422093703,
405
+ "f1_stderr": 0.014839105223388362,
406
+ "main_score": 0.20659124446630592
407
+ },
408
+ "da": {
409
+ "accuracy": 0.6151500245941957,
410
+ "accuracy_stderr": 0.012091142990997788,
411
+ "f1": 0.5982682809688111,
412
+ "f1_stderr": 0.012459918653656378,
413
+ "main_score": 0.6151500245941957
414
+ },
415
+ "de": {
416
+ "accuracy": 0.6047712739793408,
417
+ "accuracy_stderr": 0.016043636747188894,
418
+ "f1": 0.5797106547989015,
419
+ "f1_stderr": 0.015171814058353493,
420
+ "main_score": 0.6047712739793408
421
+ },
422
+ "el": {
423
+ "accuracy": 0.6089522872602067,
424
+ "accuracy_stderr": 0.019640894561031696,
425
+ "f1": 0.5849657704595443,
426
+ "f1_stderr": 0.019310550074404298,
427
+ "main_score": 0.6089522872602067
428
+ },
429
+ "en": {
430
+ "accuracy": 0.7377766847024103,
431
+ "accuracy_stderr": 0.01270632412817561,
432
+ "f1": 0.7206578045634918,
433
+ "f1_stderr": 0.014864932240730423,
434
+ "main_score": 0.7377766847024103
435
+ },
436
+ "es": {
437
+ "accuracy": 0.6441219872110182,
438
+ "accuracy_stderr": 0.01852418106999561,
439
+ "f1": 0.6307675172530458,
440
+ "f1_stderr": 0.02016317381783843,
441
+ "main_score": 0.6441219872110182
442
+ },
443
+ "evaluation_time": 100.79,
444
+ "fa": {
445
+ "accuracy": 0.5853910477127398,
446
+ "accuracy_stderr": 0.02108851541865383,
447
+ "f1": 0.5666241218934962,
448
+ "f1_stderr": 0.020157721860729564,
449
+ "main_score": 0.5853910477127398
450
+ },
451
+ "fi": {
452
+ "accuracy": 0.5398425971470732,
453
+ "accuracy_stderr": 0.016918590207041517,
454
+ "f1": 0.5161164378056002,
455
+ "f1_stderr": 0.01257048575064173,
456
+ "main_score": 0.5398425971470732
457
+ },
458
+ "fr": {
459
+ "accuracy": 0.6535169699950811,
460
+ "accuracy_stderr": 0.012275830108491429,
461
+ "f1": 0.6392304539407139,
462
+ "f1_stderr": 0.012755726012855972,
463
+ "main_score": 0.6535169699950811
464
+ },
465
+ "he": {
466
+ "accuracy": 0.5482538121003444,
467
+ "accuracy_stderr": 0.01953986797946493,
468
+ "f1": 0.5295793461894392,
469
+ "f1_stderr": 0.018246697868480273,
470
+ "main_score": 0.5482538121003444
471
+ },
472
+ "hi": {
473
+ "accuracy": 0.5594687653713724,
474
+ "accuracy_stderr": 0.01195429720795376,
475
+ "f1": 0.5370596588499968,
476
+ "f1_stderr": 0.01419271347976376,
477
+ "main_score": 0.5594687653713724
478
+ },
479
+ "hu": {
480
+ "accuracy": 0.6138219380226266,
481
+ "accuracy_stderr": 0.020042624842369013,
482
+ "f1": 0.5941229272254975,
483
+ "f1_stderr": 0.020465535639181735,
484
+ "main_score": 0.6138219380226266
485
+ },
486
+ "hy": {
487
+ "accuracy": 0.4934579439252336,
488
+ "accuracy_stderr": 0.024650010448977192,
489
+ "f1": 0.473288097937573,
490
+ "f1_stderr": 0.02185209921806471,
491
+ "main_score": 0.4934579439252336
492
+ },
493
+ "id": {
494
+ "accuracy": 0.6541564190850959,
495
+ "accuracy_stderr": 0.014537562339330085,
496
+ "f1": 0.6346738899637925,
497
+ "f1_stderr": 0.016167154430061066,
498
+ "main_score": 0.6541564190850959
499
+ },
500
+ "is": {
501
+ "accuracy": 0.23536645351697,
502
+ "accuracy_stderr": 0.019662132791740616,
503
+ "f1": 0.2270510831887765,
504
+ "f1_stderr": 0.02167220950670803,
505
+ "main_score": 0.23536645351697
506
+ },
507
+ "it": {
508
+ "accuracy": 0.6546974913920314,
509
+ "accuracy_stderr": 0.017917002438010747,
510
+ "f1": 0.6399762846530412,
511
+ "f1_stderr": 0.0156868314683275,
512
+ "main_score": 0.6546974913920314
513
+ },
514
+ "ja": {
515
+ "accuracy": 0.628430890309887,
516
+ "accuracy_stderr": 0.021493920066148326,
517
+ "f1": 0.6070556854224135,
518
+ "f1_stderr": 0.020529348132386892,
519
+ "main_score": 0.628430890309887
520
+ },
521
+ "jv": {
522
+ "accuracy": 0.36571569109690116,
523
+ "accuracy_stderr": 0.021611847788159288,
524
+ "f1": 0.3513567150725679,
525
+ "f1_stderr": 0.015646766183479204,
526
+ "main_score": 0.36571569109690116
527
+ },
528
+ "ka": {
529
+ "accuracy": 0.3947860304968028,
530
+ "accuracy_stderr": 0.03170511409604041,
531
+ "f1": 0.3810831812807749,
532
+ "f1_stderr": 0.029009260581427092,
533
+ "main_score": 0.3947860304968028
534
+ },
535
+ "km": {
536
+ "accuracy": 0.08863748155435316,
537
+ "accuracy_stderr": 0.037725796804044545,
538
+ "f1": 0.031998040803703526,
539
+ "f1_stderr": 0.007698745921074696,
540
+ "main_score": 0.08863748155435316
541
+ },
542
+ "kn": {
543
+ "accuracy": 0.0835710772257747,
544
+ "accuracy_stderr": 0.031189047693394246,
545
+ "f1": 0.03989273869929425,
546
+ "f1_stderr": 0.007820389417730708,
547
+ "main_score": 0.0835710772257747
548
+ },
549
+ "ko": {
550
+ "accuracy": 0.5884899163797344,
551
+ "accuracy_stderr": 0.017903763704391364,
552
+ "f1": 0.5753967741432244,
553
+ "f1_stderr": 0.015249469540768773,
554
+ "main_score": 0.5884899163797344
555
+ },
556
+ "lv": {
557
+ "accuracy": 0.5254795868175111,
558
+ "accuracy_stderr": 0.024488755779659522,
559
+ "f1": 0.5086248897101241,
560
+ "f1_stderr": 0.020856520346833204,
561
+ "main_score": 0.5254795868175111
562
+ },
563
+ "ml": {
564
+ "accuracy": 0.08775209050664043,
565
+ "accuracy_stderr": 0.02030140250937279,
566
+ "f1": 0.04439924108869044,
567
+ "f1_stderr": 0.008749944966388916,
568
+ "main_score": 0.08775209050664043
569
+ },
570
+ "mn": {
571
+ "accuracy": 0.47555336940482046,
572
+ "accuracy_stderr": 0.024882520470382558,
573
+ "f1": 0.4564698976791247,
574
+ "f1_stderr": 0.01894586373188271,
575
+ "main_score": 0.47555336940482046
576
+ },
577
+ "ms": {
578
+ "accuracy": 0.5584849975405805,
579
+ "accuracy_stderr": 0.018101328086571545,
580
+ "f1": 0.5319025578863811,
581
+ "f1_stderr": 0.01308304082900312,
582
+ "main_score": 0.5584849975405805
583
+ },
584
+ "my": {
585
+ "accuracy": 0.43974422036399413,
586
+ "accuracy_stderr": 0.020265378534307408,
587
+ "f1": 0.4196374455137207,
588
+ "f1_stderr": 0.014502814089256877,
589
+ "main_score": 0.43974422036399413
590
+ },
591
+ "nb": {
592
+ "accuracy": 0.5441711756025578,
593
+ "accuracy_stderr": 0.020866606211572873,
594
+ "f1": 0.5358789456153541,
595
+ "f1_stderr": 0.017683513546565128,
596
+ "main_score": 0.5441711756025578
597
+ },
598
+ "nl": {
599
+ "accuracy": 0.6698475159862273,
600
+ "accuracy_stderr": 0.011081698132348518,
601
+ "f1": 0.6416430798811394,
602
+ "f1_stderr": 0.0123689013190242,
603
+ "main_score": 0.6698475159862273
604
+ },
605
+ "pl": {
606
+ "accuracy": 0.610231185440236,
607
+ "accuracy_stderr": 0.012091142990997786,
608
+ "f1": 0.5917296041546735,
609
+ "f1_stderr": 0.015659079401564028,
610
+ "main_score": 0.610231185440236
611
+ },
612
+ "pt": {
613
+ "accuracy": 0.6450073782587309,
614
+ "accuracy_stderr": 0.0161545491710674,
615
+ "f1": 0.633366932977432,
616
+ "f1_stderr": 0.016056965049386068,
617
+ "main_score": 0.6450073782587309
618
+ },
619
+ "ro": {
620
+ "accuracy": 0.5941957697983276,
621
+ "accuracy_stderr": 0.020519284051574743,
622
+ "f1": 0.5742419794904731,
623
+ "f1_stderr": 0.021512805285723153,
624
+ "main_score": 0.5941957697983276
625
+ },
626
+ "ru": {
627
+ "accuracy": 0.654156419085096,
628
+ "accuracy_stderr": 0.019736072085047567,
629
+ "f1": 0.6414328615917462,
630
+ "f1_stderr": 0.018766742209807574,
631
+ "main_score": 0.654156419085096
632
+ },
633
+ "sl": {
634
+ "accuracy": 0.5965568125922281,
635
+ "accuracy_stderr": 0.020016472416427237,
636
+ "f1": 0.5848114214201292,
637
+ "f1_stderr": 0.019105688554454107,
638
+ "main_score": 0.5965568125922281
639
+ },
640
+ "sq": {
641
+ "accuracy": 0.6297097884899163,
642
+ "accuracy_stderr": 0.01582276716615828,
643
+ "f1": 0.6015269875641813,
644
+ "f1_stderr": 0.014823041132842616,
645
+ "main_score": 0.6297097884899163
646
+ },
647
+ "sv": {
648
+ "accuracy": 0.6479586817511067,
649
+ "accuracy_stderr": 0.01606594072698441,
650
+ "f1": 0.6182276155982869,
651
+ "f1_stderr": 0.016873004397551076,
652
+ "main_score": 0.6479586817511067
653
+ },
654
+ "sw": {
655
+ "accuracy": 0.2534677816035416,
656
+ "accuracy_stderr": 0.018721401652220998,
657
+ "f1": 0.22846166322400446,
658
+ "f1_stderr": 0.024032447392137623,
659
+ "main_score": 0.2534677816035416
660
+ },
661
+ "ta": {
662
+ "accuracy": 0.08544023610427938,
663
+ "accuracy_stderr": 0.023123828555317152,
664
+ "f1": 0.040249289724332546,
665
+ "f1_stderr": 0.0076222169607151925,
666
+ "main_score": 0.08544023610427938
667
+ },
668
+ "te": {
669
+ "accuracy": 0.07909493359567141,
670
+ "accuracy_stderr": 0.03096353573002915,
671
+ "f1": 0.03772687007380544,
672
+ "f1_stderr": 0.010114598175420026,
673
+ "main_score": 0.07909493359567141
674
+ },
675
+ "th": {
676
+ "accuracy": 0.5497786522380719,
677
+ "accuracy_stderr": 0.013425908171728488,
678
+ "f1": 0.5248955075737486,
679
+ "f1_stderr": 0.013322593137736103,
680
+ "main_score": 0.5497786522380719
681
+ },
682
+ "tl": {
683
+ "accuracy": 0.362075750122971,
684
+ "accuracy_stderr": 0.016046954165251693,
685
+ "f1": 0.3368318836191356,
686
+ "f1_stderr": 0.014374028977183335,
687
+ "main_score": 0.362075750122971
688
+ },
689
+ "tr": {
690
+ "accuracy": 0.6099852434825381,
691
+ "accuracy_stderr": 0.024320613127284566,
692
+ "f1": 0.5984689694013847,
693
+ "f1_stderr": 0.02218444797652521,
694
+ "main_score": 0.6099852434825381
695
+ },
696
+ "ur": {
697
+ "accuracy": 0.5557796360059026,
698
+ "accuracy_stderr": 0.013361598997042295,
699
+ "f1": 0.5347249852040503,
700
+ "f1_stderr": 0.011427106920460835,
701
+ "main_score": 0.5557796360059026
702
+ },
703
+ "vi": {
704
+ "accuracy": 0.5476635514018692,
705
+ "accuracy_stderr": 0.01773539963136706,
706
+ "f1": 0.5384612852637958,
707
+ "f1_stderr": 0.01967526491677906,
708
+ "main_score": 0.5476635514018692
709
+ },
710
+ "zh-CN": {
711
+ "accuracy": 0.6641908509591736,
712
+ "accuracy_stderr": 0.01279475067587787,
713
+ "f1": 0.6477127265966637,
714
+ "f1_stderr": 0.014522861971480186,
715
+ "main_score": 0.6641908509591736
716
+ },
717
+ "zh-TW": {
718
+ "accuracy": 0.6355632070831285,
719
+ "accuracy_stderr": 0.01647106176894887,
720
+ "f1": 0.6243568830105254,
721
+ "f1_stderr": 0.018043841978706926,
722
+ "main_score": 0.6355632070831285
723
+ }
724
+ }
725
+ }
results/distiluse-base-multilingual-cased-v2/MindSmallReranking.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "3bdac13927fdc888b903db93b2ffdbd90b295a69",
3
+ "mteb_dataset_name": "MindSmallReranking",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "evaluation_time": 873.24,
7
+ "map": 0.3038952232360689,
8
+ "mrr": 0.3150719866263721
9
+ }
10
+ }
results/distiluse-base-multilingual-cased-v2/SICK-R.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "a6ea5a8cab320b040a23452cc28066d9beae2cee",
3
+ "mteb_dataset_name": "SICK-R",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.8118466748223793,
8
+ "spearman": 0.7524738784985722
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.7851159752223624,
12
+ "spearman": 0.7546087065937311
13
+ },
14
+ "evaluation_time": 3.49,
15
+ "manhattan": {
16
+ "pearson": 0.7716743820738003,
17
+ "spearman": 0.7349433694282184
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/STS12.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "a0d554a64d88156834ff5ae9920b964011b16384",
3
+ "mteb_dataset_name": "STS12",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.7935237266605724,
8
+ "spearman": 0.7295904349793416
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.7307895490202788,
12
+ "spearman": 0.7166451640969629
13
+ },
14
+ "evaluation_time": 1.4,
15
+ "manhattan": {
16
+ "pearson": 0.7308359981539324,
17
+ "spearman": 0.7191126963073746
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/STS13.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7e90230a92c190f1bf69ae9002b8cea547a64cca",
3
+ "mteb_dataset_name": "STS13",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.6826126180159084,
8
+ "spearman": 0.705821267642011
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.6932005598610408,
12
+ "spearman": 0.6991767420734865
13
+ },
14
+ "evaluation_time": 0.61,
15
+ "manhattan": {
16
+ "pearson": 0.6965574245013868,
17
+ "spearman": 0.7022188522513176
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/STS14.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "6031580fec1f6af667f0bd2da0a551cf4f0b2375",
3
+ "mteb_dataset_name": "STS14",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.738304467062826,
8
+ "spearman": 0.7028565248557119
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.7280361711138981,
12
+ "spearman": 0.7063777081958187
13
+ },
14
+ "evaluation_time": 1.55,
15
+ "manhattan": {
16
+ "pearson": 0.7288892597106383,
17
+ "spearman": 0.7086449280993048
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/STS15.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ae752c7c21bf194d8b67fd573edf7ae58183cbe3",
3
+ "mteb_dataset_name": "STS15",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.8141478503988436,
8
+ "spearman": 0.8194087130039842
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.8123351470401855,
12
+ "spearman": 0.8143266713211875
13
+ },
14
+ "evaluation_time": 1.28,
15
+ "manhattan": {
16
+ "pearson": 0.8116667353510841,
17
+ "spearman": 0.8124163241523068
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/STS16.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "4d8694f8f0e0100860b497b999b3dbed754a0513",
3
+ "mteb_dataset_name": "STS16",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.7508475719822001,
8
+ "spearman": 0.7680438358515593
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.7590649123881406,
12
+ "spearman": 0.7594823191640231
13
+ },
14
+ "evaluation_time": 0.54,
15
+ "manhattan": {
16
+ "pearson": 0.7564396465387331,
17
+ "spearman": 0.7556185817375638
18
+ }
19
+ }
20
+ }
results/distiluse-base-multilingual-cased-v2/STS17.json ADDED
@@ -0,0 +1,162 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "af5e6fb845001ecf41f4c1e033ce921939a2a68d",
3
+ "mteb_dataset_name": "STS17",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "ar-ar": {
7
+ "cos_sim": {
8
+ "pearson": 0.7654105158056127,
9
+ "spearman": 0.7734104635434047
10
+ },
11
+ "euclidean": {
12
+ "pearson": 0.7528125389103582,
13
+ "spearman": 0.7542418151345
14
+ },
15
+ "manhattan": {
16
+ "pearson": 0.742691880967768,
17
+ "spearman": 0.7414253657856801
18
+ }
19
+ },
20
+ "en-ar": {
21
+ "cos_sim": {
22
+ "pearson": 0.7702928931510961,
23
+ "spearman": 0.7745907270306684
24
+ },
25
+ "euclidean": {
26
+ "pearson": 0.7747937379735675,
27
+ "spearman": 0.7721301895586583
28
+ },
29
+ "manhattan": {
30
+ "pearson": 0.766676288138473,
31
+ "spearman": 0.767187203876331
32
+ }
33
+ },
34
+ "en-de": {
35
+ "cos_sim": {
36
+ "pearson": 0.7985147526701459,
37
+ "spearman": 0.8024439450219447
38
+ },
39
+ "euclidean": {
40
+ "pearson": 0.8016905693851314,
41
+ "spearman": 0.7930869641757035
42
+ },
43
+ "manhattan": {
44
+ "pearson": 0.794830024429918,
45
+ "spearman": 0.7864845690144578
46
+ }
47
+ },
48
+ "en-en": {
49
+ "cos_sim": {
50
+ "pearson": 0.8523328074603815,
51
+ "spearman": 0.8618847213007086
52
+ },
53
+ "euclidean": {
54
+ "pearson": 0.8591331577309407,
55
+ "spearman": 0.8589967500124904
56
+ },
57
+ "manhattan": {
58
+ "pearson": 0.8513857617716476,
59
+ "spearman": 0.8482259586513993
60
+ }
61
+ },
62
+ "en-tr": {
63
+ "cos_sim": {
64
+ "pearson": 0.7538182956463326,
65
+ "spearman": 0.7434143229429068
66
+ },
67
+ "euclidean": {
68
+ "pearson": 0.7666151217728661,
69
+ "spearman": 0.7568846427284615
70
+ },
71
+ "manhattan": {
72
+ "pearson": 0.7555942040372382,
73
+ "spearman": 0.7467284614447758
74
+ }
75
+ },
76
+ "es-en": {
77
+ "cos_sim": {
78
+ "pearson": 0.7694108940753875,
79
+ "spearman": 0.7739619379750976
80
+ },
81
+ "euclidean": {
82
+ "pearson": 0.767736720732895,
83
+ "spearman": 0.7629160645031078
84
+ },
85
+ "manhattan": {
86
+ "pearson": 0.7469337188827635,
87
+ "spearman": 0.7447874230344613
88
+ }
89
+ },
90
+ "es-es": {
91
+ "cos_sim": {
92
+ "pearson": 0.8399450399002905,
93
+ "spearman": 0.8371182297187156
94
+ },
95
+ "euclidean": {
96
+ "pearson": 0.8514304799861979,
97
+ "spearman": 0.8369127569618827
98
+ },
99
+ "manhattan": {
100
+ "pearson": 0.8490116866712871,
101
+ "spearman": 0.8331690582990805
102
+ }
103
+ },
104
+ "evaluation_time": 2.45,
105
+ "fr-en": {
106
+ "cos_sim": {
107
+ "pearson": 0.7912525161262888,
108
+ "spearman": 0.7927905944348255
109
+ },
110
+ "euclidean": {
111
+ "pearson": 0.8037847361563627,
112
+ "spearman": 0.7945430583111714
113
+ },
114
+ "manhattan": {
115
+ "pearson": 0.7939311209355259,
116
+ "spearman": 0.7835224091918822
117
+ }
118
+ },
119
+ "it-en": {
120
+ "cos_sim": {
121
+ "pearson": 0.8035229136945712,
122
+ "spearman": 0.8082110464777067
123
+ },
124
+ "euclidean": {
125
+ "pearson": 0.808820546236635,
126
+ "spearman": 0.8052608029482143
127
+ },
128
+ "manhattan": {
129
+ "pearson": 0.7987881836256757,
130
+ "spearman": 0.7921409642635105
131
+ }
132
+ },
133
+ "ko-ko": {
134
+ "cos_sim": {
135
+ "pearson": 0.7657756740555968,
136
+ "spearman": 0.7639843364267264
137
+ },
138
+ "euclidean": {
139
+ "pearson": 0.7540424583472578,
140
+ "spearman": 0.7531307938562327
141
+ },
142
+ "manhattan": {
143
+ "pearson": 0.7473109587053861,
144
+ "spearman": 0.7454667368714957
145
+ }
146
+ },
147
+ "nl-en": {
148
+ "cos_sim": {
149
+ "pearson": 0.8008711291606405,
150
+ "spearman": 0.8050747550174946
151
+ },
152
+ "euclidean": {
153
+ "pearson": 0.8019128295947303,
154
+ "spearman": 0.7980068556328985
155
+ },
156
+ "manhattan": {
157
+ "pearson": 0.7928055314670001,
158
+ "spearman": 0.7867459586691882
159
+ }
160
+ }
161
+ }
162
+ }
results/distiluse-base-multilingual-cased-v2/STS22.json ADDED
@@ -0,0 +1,260 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "6d1ba47164174a496b7fa5d3569dae26a6813b80",
3
+ "mteb_dataset_name": "STS22",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "ar": {
7
+ "cos_sim": {
8
+ "pearson": 0.4319525519197726,
9
+ "spearman": 0.4904013064287781
10
+ },
11
+ "euclidean": {
12
+ "pearson": 0.4151101650799975,
13
+ "spearman": 0.45694919819202556
14
+ },
15
+ "manhattan": {
16
+ "pearson": 0.41798306097489685,
17
+ "spearman": 0.45889699163278647
18
+ }
19
+ },
20
+ "de": {
21
+ "cos_sim": {
22
+ "pearson": 0.30728588031668386,
23
+ "spearman": 0.35729106419179457
24
+ },
25
+ "euclidean": {
26
+ "pearson": 0.27727483814940634,
27
+ "spearman": 0.36697908777201876
28
+ },
29
+ "manhattan": {
30
+ "pearson": 0.26887457740598375,
31
+ "spearman": 0.35651935891649017
32
+ }
33
+ },
34
+ "de-en": {
35
+ "cos_sim": {
36
+ "pearson": 0.5239033873441029,
37
+ "spearman": 0.4750888019756861
38
+ },
39
+ "euclidean": {
40
+ "pearson": 0.5409329593694967,
41
+ "spearman": 0.4674591134379504
42
+ },
43
+ "manhattan": {
44
+ "pearson": 0.5507151796287579,
45
+ "spearman": 0.4782505012490346
46
+ }
47
+ },
48
+ "de-fr": {
49
+ "cos_sim": {
50
+ "pearson": 0.562735220514599,
51
+ "spearman": 0.6076242915296164
52
+ },
53
+ "euclidean": {
54
+ "pearson": 0.5473358313453174,
55
+ "spearman": 0.5901153256838316
56
+ },
57
+ "manhattan": {
58
+ "pearson": 0.5330971466711619,
59
+ "spearman": 0.5742760292614851
60
+ }
61
+ },
62
+ "de-pl": {
63
+ "cos_sim": {
64
+ "pearson": 0.33210422466959244,
65
+ "spearman": 0.3609068930156353
66
+ },
67
+ "euclidean": {
68
+ "pearson": 0.3672425141682268,
69
+ "spearman": 0.333808081935963
70
+ },
71
+ "manhattan": {
72
+ "pearson": 0.3547249118003641,
73
+ "spearman": 0.31964279432613435
74
+ }
75
+ },
76
+ "en": {
77
+ "cos_sim": {
78
+ "pearson": 0.6374947679318765,
79
+ "spearman": 0.6287618960301087
80
+ },
81
+ "euclidean": {
82
+ "pearson": 0.6200259194547161,
83
+ "spearman": 0.6014134804263503
84
+ },
85
+ "manhattan": {
86
+ "pearson": 0.6185663435862556,
87
+ "spearman": 0.6049194043559385
88
+ }
89
+ },
90
+ "es": {
91
+ "cos_sim": {
92
+ "pearson": 0.5851573251701789,
93
+ "spearman": 0.5934352724163223
94
+ },
95
+ "euclidean": {
96
+ "pearson": 0.5937822334487575,
97
+ "spearman": 0.599529665367923
98
+ },
99
+ "manhattan": {
100
+ "pearson": 0.5934905346132588,
101
+ "spearman": 0.5958363163864109
102
+ }
103
+ },
104
+ "es-en": {
105
+ "cos_sim": {
106
+ "pearson": 0.6622856680218524,
107
+ "spearman": 0.6895835518547431
108
+ },
109
+ "euclidean": {
110
+ "pearson": 0.6945990476537346,
111
+ "spearman": 0.6951326488176927
112
+ },
113
+ "manhattan": {
114
+ "pearson": 0.692654378415376,
115
+ "spearman": 0.6925549968332008
116
+ }
117
+ },
118
+ "es-it": {
119
+ "cos_sim": {
120
+ "pearson": 0.6054093974085284,
121
+ "spearman": 0.6327724621350164
122
+ },
123
+ "euclidean": {
124
+ "pearson": 0.5921790717375445,
125
+ "spearman": 0.6077632900198519
126
+ },
127
+ "manhattan": {
128
+ "pearson": 0.5957257324550282,
129
+ "spearman": 0.6086391917522135
130
+ }
131
+ },
132
+ "evaluation_time": 14.69,
133
+ "fr": {
134
+ "cos_sim": {
135
+ "pearson": 0.7666955782585071,
136
+ "spearman": 0.7641254265129491
137
+ },
138
+ "euclidean": {
139
+ "pearson": 0.6810573760855496,
140
+ "spearman": 0.7153756176277793
141
+ },
142
+ "manhattan": {
143
+ "pearson": 0.677124757126929,
144
+ "spearman": 0.7152537846395397
145
+ }
146
+ },
147
+ "fr-pl": {
148
+ "cos_sim": {
149
+ "pearson": 0.6272171062751704,
150
+ "spearman": 0.6197797868009122
151
+ },
152
+ "euclidean": {
153
+ "pearson": 0.6359898515445168,
154
+ "spearman": 0.8451542547285167
155
+ },
156
+ "manhattan": {
157
+ "pearson": 0.6215380605376377,
158
+ "spearman": 0.7324670207647144
159
+ }
160
+ },
161
+ "it": {
162
+ "cos_sim": {
163
+ "pearson": 0.6386370050619784,
164
+ "spearman": 0.6510152541505574
165
+ },
166
+ "euclidean": {
167
+ "pearson": 0.6123738658178195,
168
+ "spearman": 0.6277231926242124
169
+ },
170
+ "manhattan": {
171
+ "pearson": 0.6120141239111747,
172
+ "spearman": 0.6258683030963466
173
+ }
174
+ },
175
+ "pl": {
176
+ "cos_sim": {
177
+ "pearson": 0.2673251862968695,
178
+ "spearman": 0.3457702083368428
179
+ },
180
+ "euclidean": {
181
+ "pearson": 0.11555722679629112,
182
+ "spearman": 0.3383302978677857
183
+ },
184
+ "manhattan": {
185
+ "pearson": 0.11309586078967969,
186
+ "spearman": 0.3345113736058396
187
+ }
188
+ },
189
+ "pl-en": {
190
+ "cos_sim": {
191
+ "pearson": 0.7024310112698742,
192
+ "spearman": 0.7132608389737901
193
+ },
194
+ "euclidean": {
195
+ "pearson": 0.6953167907457565,
196
+ "spearman": 0.6924756304760876
197
+ },
198
+ "manhattan": {
199
+ "pearson": 0.6944320012141271,
200
+ "spearman": 0.6992998467998947
201
+ }
202
+ },
203
+ "ru": {
204
+ "cos_sim": {
205
+ "pearson": 0.4672887212606245,
206
+ "spearman": 0.5240251410115027
207
+ },
208
+ "euclidean": {
209
+ "pearson": 0.4261087105318375,
210
+ "spearman": 0.4931647979068464
211
+ },
212
+ "manhattan": {
213
+ "pearson": 0.41971488569524223,
214
+ "spearman": 0.4860394808010442
215
+ }
216
+ },
217
+ "tr": {
218
+ "cos_sim": {
219
+ "pearson": 0.5059069907623683,
220
+ "spearman": 0.5407437321160808
221
+ },
222
+ "euclidean": {
223
+ "pearson": 0.5531327716542195,
224
+ "spearman": 0.55862881519289
225
+ },
226
+ "manhattan": {
227
+ "pearson": 0.5576874086920313,
228
+ "spearman": 0.5638920793992543
229
+ }
230
+ },
231
+ "zh": {
232
+ "cos_sim": {
233
+ "pearson": 0.502828997035562,
234
+ "spearman": 0.5431518993723914
235
+ },
236
+ "euclidean": {
237
+ "pearson": 0.46926861345873216,
238
+ "spearman": 0.504258942374202
239
+ },
240
+ "manhattan": {
241
+ "pearson": 0.4711937333538452,
242
+ "spearman": 0.5029054521403065
243
+ }
244
+ },
245
+ "zh-en": {
246
+ "cos_sim": {
247
+ "pearson": 0.6296457033320131,
248
+ "spearman": 0.6175062747584529
249
+ },
250
+ "euclidean": {
251
+ "pearson": 0.5958377101704754,
252
+ "spearman": 0.5591175172327044
253
+ },
254
+ "manhattan": {
255
+ "pearson": 0.5964672089274813,
256
+ "spearman": 0.5593114256617111
257
+ }
258
+ }
259
+ }
260
+ }
results/distiluse-base-multilingual-cased-v2/STSBenchmark.json ADDED
@@ -0,0 +1,35 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "b0fddb56ed78048fa8b90373c8a3cfc37b684831",
3
+ "mteb_dataset_name": "STSBenchmark",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "pearson": 0.8168394886293749,
8
+ "spearman": 0.807547875467642
9
+ },
10
+ "euclidean": {
11
+ "pearson": 0.8067588249670364,
12
+ "spearman": 0.802296669116562
13
+ },
14
+ "evaluation_time": 0.56,
15
+ "manhattan": {
16
+ "pearson": 0.7979275882752755,
17
+ "spearman": 0.7941562131296505
18
+ }
19
+ },
20
+ "validation": {
21
+ "cos_sim": {
22
+ "pearson": 0.8190973141367347,
23
+ "spearman": 0.8193330529749685
24
+ },
25
+ "euclidean": {
26
+ "pearson": 0.8114127619720831,
27
+ "spearman": 0.8098557154737026
28
+ },
29
+ "evaluation_time": 0.7,
30
+ "manhattan": {
31
+ "pearson": 0.8037859722167895,
32
+ "spearman": 0.802001005426749
33
+ }
34
+ }
35
+ }
results/distiluse-base-multilingual-cased-v2/SciDocsRR.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d3c5e1fc0b855ab6097bf1cda04dd73947d7caab",
3
+ "mteb_dataset_name": "SciDocsRR",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "evaluation_time": 33.01,
7
+ "map": 0.6921586199162861,
8
+ "mrr": 0.8886282290694054
9
+ },
10
+ "validation": {
11
+ "evaluation_time": 33.16,
12
+ "map": 0.6930483858264198,
13
+ "mrr": 0.8874754289223016
14
+ }
15
+ }
results/distiluse-base-multilingual-cased-v2/SprintDuplicateQuestions.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46",
3
+ "mteb_dataset_name": "SprintDuplicateQuestions",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.9962079207920792,
8
+ "accuracy_threshold": 0.6402517557144165,
9
+ "ap": 0.8714976457350163,
10
+ "f1": 0.8107317073170732,
11
+ "f1_threshold": 0.63487708568573,
12
+ "precision": 0.7914285714285715,
13
+ "recall": 0.831
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.9957722772277228,
17
+ "accuracy_threshold": 0.5670002102851868,
18
+ "ap": 0.8407833605976549,
19
+ "f1": 0.7788461538461539,
20
+ "f1_threshold": 0.5344296097755432,
21
+ "precision": 0.75,
22
+ "recall": 0.81
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.9961287128712871,
26
+ "accuracy_threshold": 0.7639181017875671,
27
+ "ap": 0.8694165408325188,
28
+ "f1": 0.8033596837944663,
29
+ "f1_threshold": 0.7781314849853516,
30
+ "precision": 0.7939453125,
31
+ "recall": 0.813
32
+ },
33
+ "evaluation_time": 3.96,
34
+ "manhattan": {
35
+ "accuracy": 0.9964653465346535,
36
+ "accuracy_threshold": 13.177294731140137,
37
+ "ap": 0.8843495903247095,
38
+ "f1": 0.817193675889328,
39
+ "f1_threshold": 13.655448913574219,
40
+ "precision": 0.8076171875,
41
+ "recall": 0.827
42
+ },
43
+ "max": {
44
+ "accuracy": 0.9964653465346535,
45
+ "ap": 0.8843495903247095,
46
+ "f1": 0.817193675889328
47
+ }
48
+ },
49
+ "validation": {
50
+ "cos_sim": {
51
+ "accuracy": 0.9962673267326733,
52
+ "accuracy_threshold": 0.6435202360153198,
53
+ "ap": 0.8654947393156018,
54
+ "f1": 0.808780487804878,
55
+ "f1_threshold": 0.625939130783081,
56
+ "precision": 0.7895238095238095,
57
+ "recall": 0.829
58
+ },
59
+ "dot": {
60
+ "accuracy": 0.995990099009901,
61
+ "accuracy_threshold": 0.5402953624725342,
62
+ "ap": 0.8619092379785963,
63
+ "f1": 0.7924161400097229,
64
+ "f1_threshold": 0.5242564678192139,
65
+ "precision": 0.771050141911069,
66
+ "recall": 0.815
67
+ },
68
+ "euclidean": {
69
+ "accuracy": 0.996009900990099,
70
+ "accuracy_threshold": 0.7580994367599487,
71
+ "ap": 0.8482742572952514,
72
+ "f1": 0.7965200579990334,
73
+ "f1_threshold": 0.7885181307792664,
74
+ "precision": 0.7708138447146866,
75
+ "recall": 0.824
76
+ },
77
+ "evaluation_time": 5.48,
78
+ "manhattan": {
79
+ "accuracy": 0.9963861386138614,
80
+ "accuracy_threshold": 13.252960205078125,
81
+ "ap": 0.8715435403229772,
82
+ "f1": 0.8145985401459853,
83
+ "f1_threshold": 13.848575592041016,
84
+ "precision": 0.7933649289099526,
85
+ "recall": 0.837
86
+ },
87
+ "max": {
88
+ "accuracy": 0.9963861386138614,
89
+ "ap": 0.8715435403229772,
90
+ "f1": 0.8145985401459853
91
+ }
92
+ }
93
+ }
results/distiluse-base-multilingual-cased-v2/StackOverflowDupQuestions.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "e185fbe320c72810689fc5848eb6114e1ef5ec69",
3
+ "mteb_dataset_name": "StackOverflowDupQuestions",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "evaluation_time": 17.24,
7
+ "map": 0.41920314992536173,
8
+ "mrr": 0.4211711389101095
9
+ }
10
+ }
results/distiluse-base-multilingual-cased-v2/ToxicConversationsClassification.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d7c0de2777da35d6aae2200a62c6e0e5af397c4c",
3
+ "mteb_dataset_name": "ToxicConversationsClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "accuracy": 0.690936,
7
+ "accuracy_stderr": 0.044703511092530526,
8
+ "ap": 0.13464419132094954,
9
+ "ap_stderr": 0.011043199478743274,
10
+ "evaluation_time": 36.86,
11
+ "f1": 0.5317756829624628,
12
+ "f1_stderr": 0.027320019113534114,
13
+ "main_score": 0.690936
14
+ }
15
+ }
results/distiluse-base-multilingual-cased-v2/TweetSentimentExtractionClassification.json ADDED
@@ -0,0 +1,13 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d604517c81ca91fe16a244d1248fc021f9ecee7a",
3
+ "mteb_dataset_name": "TweetSentimentExtractionClassification",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "accuracy": 0.599688737973967,
7
+ "accuracy_stderr": 0.021314288948809584,
8
+ "evaluation_time": 2.06,
9
+ "f1": 0.6023697658216021,
10
+ "f1_stderr": 0.020896006423816196,
11
+ "main_score": 0.599688737973967
12
+ }
13
+ }
results/distiluse-base-multilingual-cased-v2/TwitterSemEval2015.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "70970daeab8776df92f5ea462b6173c0b46fd2d1",
3
+ "mteb_dataset_name": "TwitterSemEval2015",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.827978780473267,
8
+ "accuracy_threshold": 0.7602837085723877,
9
+ "ap": 0.6166929108121391,
10
+ "f1": 0.5768693665100927,
11
+ "f1_threshold": 0.6814744472503662,
12
+ "precision": 0.5559089796917054,
13
+ "recall": 0.5994722955145119
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.8168921738093818,
17
+ "accuracy_threshold": 0.701004147529602,
18
+ "ap": 0.5739705387908134,
19
+ "f1": 0.5472479298587434,
20
+ "f1_threshold": 0.6023658514022827,
21
+ "precision": 0.5081411126187245,
22
+ "recall": 0.5928759894459102
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.8285152291828098,
26
+ "accuracy_threshold": 0.6665006875991821,
27
+ "ap": 0.6245681717082225,
28
+ "f1": 0.5832305795314425,
29
+ "f1_threshold": 0.7631351351737976,
30
+ "precision": 0.5474537037037037,
31
+ "recall": 0.6240105540897097
32
+ },
33
+ "evaluation_time": 4.46,
34
+ "manhattan": {
35
+ "accuracy": 0.8276807534124099,
36
+ "accuracy_threshold": 12.007987976074219,
37
+ "ap": 0.6185267667234617,
38
+ "f1": 0.5762629336579428,
39
+ "f1_threshold": 13.591428756713867,
40
+ "precision": 0.5349152542372881,
41
+ "recall": 0.6245382585751978
42
+ },
43
+ "max": {
44
+ "accuracy": 0.8285152291828098,
45
+ "ap": 0.6245681717082225,
46
+ "f1": 0.5832305795314425
47
+ }
48
+ }
49
+ }
results/distiluse-base-multilingual-cased-v2/TwitterURLCorpus.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "8b6510b0b1fa4e4c4f879467980e9be563ec1cdf",
3
+ "mteb_dataset_name": "TwitterURLCorpus",
4
+ "mteb_version": "1.0.2",
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.8803896456708192,
8
+ "accuracy_threshold": 0.6452248096466064,
9
+ "ap": 0.8402495588793271,
10
+ "f1": 0.7626290458870641,
11
+ "f1_threshold": 0.5941753387451172,
12
+ "precision": 0.7293233082706767,
13
+ "recall": 0.7991222667077302
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.8787402491558971,
17
+ "accuracy_threshold": 0.5848736763000488,
18
+ "ap": 0.8320076543059168,
19
+ "f1": 0.7602826329490517,
20
+ "f1_threshold": 0.554627001285553,
21
+ "precision": 0.7352898863472882,
22
+ "recall": 0.7870341854019095
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.8796328637404432,
26
+ "accuracy_threshold": 0.8007122278213501,
27
+ "ap": 0.8378378095020464,
28
+ "f1": 0.7594917787742901,
29
+ "f1_threshold": 0.8566794395446777,
30
+ "precision": 0.7378739471391229,
31
+ "recall": 0.7824145364952264
32
+ },
33
+ "evaluation_time": 16.04,
34
+ "manhattan": {
35
+ "accuracy": 0.8799239337136647,
36
+ "accuracy_threshold": 14.142051696777344,
37
+ "ap": 0.8372045889779074,
38
+ "f1": 0.7593527315914489,
39
+ "f1_threshold": 15.342501640319824,
40
+ "precision": 0.733018056749785,
41
+ "recall": 0.7876501385894672
42
+ },
43
+ "max": {
44
+ "accuracy": 0.8803896456708192,
45
+ "ap": 0.8402495588793271,
46
+ "f1": 0.7626290458870641
47
+ }
48
+ }
49
+ }