Amanda commited on
Commit
351f330
1 Parent(s): 1cecd71

Upload 14 files

Browse files
results/Amanda/bge_portuguese_v6/FloresBitextMining.json ADDED
@@ -0,0 +1,193 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "80dc3040d19756742c9a18267ab30f54fb8e226b",
3
+ "dev": {
4
+ "arb_Arab-por_Latn": {
5
+ "accuracy": 0.9979939819458375,
6
+ "f1": 0.9973253092611166,
7
+ "main_score": 0.9973253092611166,
8
+ "precision": 0.9969909729187563,
9
+ "recall": 0.9979939819458375
10
+ },
11
+ "deu_Latn-por_Latn": {
12
+ "accuracy": 0.9969909729187563,
13
+ "f1": 0.995987963891675,
14
+ "main_score": 0.995987963891675,
15
+ "precision": 0.9954864593781344,
16
+ "recall": 0.9969909729187563
17
+ },
18
+ "eng_Latn-por_Latn": {
19
+ "accuracy": 0.9989969909729187,
20
+ "f1": 0.9986626546305583,
21
+ "main_score": 0.9986626546305583,
22
+ "precision": 0.9984954864593781,
23
+ "recall": 0.9989969909729187
24
+ },
25
+ "evaluation_time": 180.36,
26
+ "fra_Latn-por_Latn": {
27
+ "accuracy": 0.9979939819458375,
28
+ "f1": 0.9973253092611166,
29
+ "main_score": 0.9973253092611166,
30
+ "precision": 0.9969909729187563,
31
+ "recall": 0.9979939819458375
32
+ },
33
+ "ita_Latn-por_Latn": {
34
+ "accuracy": 0.9979939819458375,
35
+ "f1": 0.9973253092611166,
36
+ "main_score": 0.9973253092611166,
37
+ "precision": 0.9969909729187563,
38
+ "recall": 0.9979939819458375
39
+ },
40
+ "jpn_Jpan-por_Latn": {
41
+ "accuracy": 0.9979939819458375,
42
+ "f1": 0.9973253092611166,
43
+ "main_score": 0.9973253092611166,
44
+ "precision": 0.9969909729187563,
45
+ "recall": 0.9979939819458375
46
+ },
47
+ "kor_Hang-por_Latn": {
48
+ "accuracy": 0.9929789368104313,
49
+ "f1": 0.9906385824139085,
50
+ "main_score": 0.9906385824139085,
51
+ "precision": 0.9894684052156469,
52
+ "recall": 0.9929789368104313
53
+ },
54
+ "pol_Latn-por_Latn": {
55
+ "accuracy": 0.995987963891675,
56
+ "f1": 0.9946506185222335,
57
+ "main_score": 0.9946506185222335,
58
+ "precision": 0.9939819458375125,
59
+ "recall": 0.995987963891675
60
+ },
61
+ "rus_Cyrl-por_Latn": {
62
+ "accuracy": 0.995987963891675,
63
+ "f1": 0.9946506185222334,
64
+ "main_score": 0.9946506185222334,
65
+ "precision": 0.9939819458375125,
66
+ "recall": 0.995987963891675
67
+ },
68
+ "spa_Latn-por_Latn": {
69
+ "accuracy": 0.9989969909729187,
70
+ "f1": 0.9986626546305583,
71
+ "main_score": 0.9986626546305583,
72
+ "precision": 0.9984954864593781,
73
+ "recall": 0.9989969909729187
74
+ },
75
+ "swe_Latn-por_Latn": {
76
+ "accuracy": 1.0,
77
+ "f1": 1.0,
78
+ "main_score": 1.0,
79
+ "precision": 1.0,
80
+ "recall": 1.0
81
+ },
82
+ "zho_Hans-por_Latn": {
83
+ "accuracy": 0.9969909729187563,
84
+ "f1": 0.995987963891675,
85
+ "main_score": 0.995987963891675,
86
+ "precision": 0.9954864593781344,
87
+ "recall": 0.9969909729187563
88
+ },
89
+ "zho_Hant-por_Latn": {
90
+ "accuracy": 0.9939819458375125,
91
+ "f1": 0.9919759277833501,
92
+ "main_score": 0.9919759277833501,
93
+ "precision": 0.9909729187562688,
94
+ "recall": 0.9939819458375125
95
+ }
96
+ },
97
+ "devtest": {
98
+ "arb_Arab-por_Latn": {
99
+ "accuracy": 0.9970355731225297,
100
+ "f1": 0.9960474308300395,
101
+ "main_score": 0.9960474308300395,
102
+ "precision": 0.9955533596837944,
103
+ "recall": 0.9970355731225297
104
+ },
105
+ "deu_Latn-por_Latn": {
106
+ "accuracy": 0.9970355731225297,
107
+ "f1": 0.9960474308300395,
108
+ "main_score": 0.9960474308300395,
109
+ "precision": 0.9955533596837944,
110
+ "recall": 0.9970355731225297
111
+ },
112
+ "eng_Latn-por_Latn": {
113
+ "accuracy": 0.9990118577075099,
114
+ "f1": 0.9986824769433464,
115
+ "main_score": 0.9986824769433464,
116
+ "precision": 0.9985177865612648,
117
+ "recall": 0.9990118577075099
118
+ },
119
+ "evaluation_time": 196.89,
120
+ "fra_Latn-por_Latn": {
121
+ "accuracy": 0.9980237154150198,
122
+ "f1": 0.997364953886693,
123
+ "main_score": 0.997364953886693,
124
+ "precision": 0.9970355731225297,
125
+ "recall": 0.9980237154150198
126
+ },
127
+ "ita_Latn-por_Latn": {
128
+ "accuracy": 0.9980237154150198,
129
+ "f1": 0.997364953886693,
130
+ "main_score": 0.997364953886693,
131
+ "precision": 0.9970355731225297,
132
+ "recall": 0.9980237154150198
133
+ },
134
+ "jpn_Jpan-por_Latn": {
135
+ "accuracy": 0.9960474308300395,
136
+ "f1": 0.994729907773386,
137
+ "main_score": 0.994729907773386,
138
+ "precision": 0.9940711462450593,
139
+ "recall": 0.9960474308300395
140
+ },
141
+ "kor_Hang-por_Latn": {
142
+ "accuracy": 0.9940711462450593,
143
+ "f1": 0.9920948616600791,
144
+ "main_score": 0.9920948616600791,
145
+ "precision": 0.991106719367589,
146
+ "recall": 0.9940711462450593
147
+ },
148
+ "pol_Latn-por_Latn": {
149
+ "accuracy": 0.9960474308300395,
150
+ "f1": 0.994729907773386,
151
+ "main_score": 0.994729907773386,
152
+ "precision": 0.9940711462450593,
153
+ "recall": 0.9960474308300395
154
+ },
155
+ "rus_Cyrl-por_Latn": {
156
+ "accuracy": 0.9980237154150198,
157
+ "f1": 0.997364953886693,
158
+ "main_score": 0.997364953886693,
159
+ "precision": 0.9970355731225297,
160
+ "recall": 0.9980237154150198
161
+ },
162
+ "spa_Latn-por_Latn": {
163
+ "accuracy": 0.9980237154150198,
164
+ "f1": 0.997364953886693,
165
+ "main_score": 0.997364953886693,
166
+ "precision": 0.9970355731225297,
167
+ "recall": 0.9980237154150198
168
+ },
169
+ "swe_Latn-por_Latn": {
170
+ "accuracy": 1.0,
171
+ "f1": 1.0,
172
+ "main_score": 1.0,
173
+ "precision": 1.0,
174
+ "recall": 1.0
175
+ },
176
+ "zho_Hans-por_Latn": {
177
+ "accuracy": 0.9960474308300395,
178
+ "f1": 0.994729907773386,
179
+ "main_score": 0.994729907773386,
180
+ "precision": 0.9940711462450593,
181
+ "recall": 0.9960474308300395
182
+ },
183
+ "zho_Hant-por_Latn": {
184
+ "accuracy": 0.9960474308300395,
185
+ "f1": 0.994729907773386,
186
+ "main_score": 0.994729907773386,
187
+ "precision": 0.9940711462450593,
188
+ "recall": 0.9960474308300395
189
+ }
190
+ },
191
+ "mteb_dataset_name": "FloresBitextMining",
192
+ "mteb_version": 2
193
+ }
results/Amanda/bge_portuguese_v6/MassiveIntentClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "31efe3c427b0bae9c22cbb560b8f15491cc6bed7",
3
+ "mteb_dataset_name": "MassiveIntentClassification",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 96.89,
7
+ "pt": {
8
+ "accuracy": 0.6429388029589779,
9
+ "accuracy_stderr": 0.01574159007730203,
10
+ "f1": 0.624149449605407,
11
+ "f1_stderr": 0.011597207566151069,
12
+ "main_score": 0.6429388029589779
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 105.97,
17
+ "pt": {
18
+ "accuracy": 0.6440727988194788,
19
+ "accuracy_stderr": 0.018382512054495048,
20
+ "f1": 0.6093616344832975,
21
+ "f1_stderr": 0.01076436730695683,
22
+ "main_score": 0.6440727988194788
23
+ }
24
+ }
25
+ }
results/Amanda/bge_portuguese_v6/MassiveScenarioClassification.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "7d571f92784cd94a019292a1f45445077d0ef634",
3
+ "mteb_dataset_name": "MassiveScenarioClassification",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 62.45,
7
+ "pt": {
8
+ "accuracy": 0.6876933422999327,
9
+ "accuracy_stderr": 0.014407566848919572,
10
+ "f1": 0.6829025521195558,
11
+ "f1_stderr": 0.013884840252469367,
12
+ "main_score": 0.6876933422999327
13
+ }
14
+ },
15
+ "validation": {
16
+ "evaluation_time": 47.48,
17
+ "pt": {
18
+ "accuracy": 0.683915395966552,
19
+ "accuracy_stderr": 0.017521288958123687,
20
+ "f1": 0.6775990739298102,
21
+ "f1_stderr": 0.017684324125881695,
22
+ "main_score": 0.683915395966552
23
+ }
24
+ }
25
+ }
results/Amanda/bge_portuguese_v6/MintakaRetrieval.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "efa78cc2f74bbcd21eff2261f9e13aebe40b814e",
3
+ "mteb_dataset_name": "MintakaRetrieval",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 53.13,
7
+ "pt": {
8
+ "map_at_1": 0.15081,
9
+ "map_at_10": 0.21799,
10
+ "map_at_100": 0.2268,
11
+ "map_at_1000": 0.22794,
12
+ "map_at_20": 0.22234,
13
+ "map_at_3": 0.19867,
14
+ "map_at_5": 0.20946,
15
+ "mrr_at_1": 0.15081,
16
+ "mrr_at_10": 0.21803,
17
+ "mrr_at_100": 0.22683,
18
+ "mrr_at_1000": 0.22797,
19
+ "mrr_at_20": 0.22237,
20
+ "mrr_at_3": 0.19867,
21
+ "mrr_at_5": 0.20946,
22
+ "ndcg_at_1": 0.15081,
23
+ "ndcg_at_10": 0.25454,
24
+ "ndcg_at_100": 0.30478,
25
+ "ndcg_at_1000": 0.34213,
26
+ "ndcg_at_20": 0.27041,
27
+ "ndcg_at_3": 0.214,
28
+ "ndcg_at_5": 0.23352,
29
+ "precision_at_1": 0.15081,
30
+ "precision_at_10": 0.03717,
31
+ "precision_at_100": 0.00624,
32
+ "precision_at_1000": 0.00093,
33
+ "precision_at_20": 0.02173,
34
+ "precision_at_3": 0.08609,
35
+ "precision_at_5": 0.06117,
36
+ "recall_at_1": 0.15081,
37
+ "recall_at_10": 0.3715,
38
+ "recall_at_100": 0.62383,
39
+ "recall_at_1000": 0.93373,
40
+ "recall_at_20": 0.43437,
41
+ "recall_at_3": 0.25828,
42
+ "recall_at_5": 0.30586
43
+ }
44
+ }
45
+ }
results/Amanda/bge_portuguese_v6/MultiHateClassification.json ADDED
@@ -0,0 +1,17 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "ef137ea2b7c719183f8f60edf536b50f56d1365b",
3
+ "mteb_dataset_name": "MultiHateClassification",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 21.52,
7
+ "por": {
8
+ "accuracy": 0.6023,
9
+ "accuracy_stderr": 0.040573513527916216,
10
+ "ap": 0.3465812754549832,
11
+ "ap_stderr": 0.024717781473017272,
12
+ "f1": 0.5641933623962817,
13
+ "f1_stderr": 0.037581938438066446,
14
+ "main_score": 0.6023
15
+ }
16
+ }
17
+ }
results/Amanda/bge_portuguese_v6/MultiLongDocRetrieval.json ADDED
@@ -0,0 +1,85 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "d67138e705d963e346253a80e59676ddb418810a",
3
+ "dev": {
4
+ "evaluation_time": 9073.93,
5
+ "pt": {
6
+ "map_at_1": 0.55,
7
+ "map_at_10": 0.5978,
8
+ "map_at_100": 0.60393,
9
+ "map_at_1000": 0.60447,
10
+ "map_at_20": 0.60064,
11
+ "map_at_3": 0.585,
12
+ "map_at_5": 0.59275,
13
+ "mrr_at_1": 0.55,
14
+ "mrr_at_10": 0.5978,
15
+ "mrr_at_100": 0.60393,
16
+ "mrr_at_1000": 0.60447,
17
+ "mrr_at_20": 0.60064,
18
+ "mrr_at_3": 0.585,
19
+ "mrr_at_5": 0.59275,
20
+ "ndcg_at_1": 0.55,
21
+ "ndcg_at_10": 0.62244,
22
+ "ndcg_at_100": 0.65668,
23
+ "ndcg_at_1000": 0.66971,
24
+ "ndcg_at_20": 0.63264,
25
+ "ndcg_at_3": 0.59655,
26
+ "ndcg_at_5": 0.61074,
27
+ "precision_at_1": 0.55,
28
+ "precision_at_10": 0.07,
29
+ "precision_at_100": 0.0087,
30
+ "precision_at_1000": 0.00097,
31
+ "precision_at_20": 0.037,
32
+ "precision_at_3": 0.21,
33
+ "precision_at_5": 0.133,
34
+ "recall_at_1": 0.55,
35
+ "recall_at_10": 0.7,
36
+ "recall_at_100": 0.87,
37
+ "recall_at_1000": 0.97,
38
+ "recall_at_20": 0.74,
39
+ "recall_at_3": 0.63,
40
+ "recall_at_5": 0.665
41
+ }
42
+ },
43
+ "mteb_dataset_name": "MultiLongDocRetrieval",
44
+ "mteb_version": 2,
45
+ "test": {
46
+ "evaluation_time": 9078.26,
47
+ "pt": {
48
+ "map_at_1": 0.525,
49
+ "map_at_10": 0.59302,
50
+ "map_at_100": 0.59871,
51
+ "map_at_1000": 0.59907,
52
+ "map_at_20": 0.59621,
53
+ "map_at_3": 0.58083,
54
+ "map_at_5": 0.58958,
55
+ "mrr_at_1": 0.525,
56
+ "mrr_at_10": 0.59302,
57
+ "mrr_at_100": 0.59871,
58
+ "mrr_at_1000": 0.59907,
59
+ "mrr_at_20": 0.59621,
60
+ "mrr_at_3": 0.58083,
61
+ "mrr_at_5": 0.58958,
62
+ "ndcg_at_1": 0.525,
63
+ "ndcg_at_10": 0.62175,
64
+ "ndcg_at_100": 0.65232,
65
+ "ndcg_at_1000": 0.66342,
66
+ "ndcg_at_20": 0.63321,
67
+ "ndcg_at_3": 0.59744,
68
+ "ndcg_at_5": 0.61357,
69
+ "precision_at_1": 0.525,
70
+ "precision_at_10": 0.071,
71
+ "precision_at_100": 0.0086,
72
+ "precision_at_1000": 0.00095,
73
+ "precision_at_20": 0.03775,
74
+ "precision_at_3": 0.215,
75
+ "precision_at_5": 0.137,
76
+ "recall_at_1": 0.525,
77
+ "recall_at_10": 0.71,
78
+ "recall_at_100": 0.86,
79
+ "recall_at_1000": 0.95,
80
+ "recall_at_20": 0.755,
81
+ "recall_at_3": 0.645,
82
+ "recall_at_5": 0.685
83
+ }
84
+ }
85
+ }
results/Amanda/bge_portuguese_v6/NTREXBitextMining.json ADDED
@@ -0,0 +1,92 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "fd20d54141b6da952d5c68a2989472892489da0f",
3
+ "mteb_dataset_name": "NTREXBitextMining",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "arb_Arab-por_Latn": {
7
+ "accuracy": 0.9694541812719079,
8
+ "f1": 0.9597729928225672,
9
+ "main_score": 0.9597729928225672,
10
+ "precision": 0.9550158571190119,
11
+ "recall": 0.9694541812719079
12
+ },
13
+ "deu_Latn-por_Latn": {
14
+ "accuracy": 0.9844767150726089,
15
+ "f1": 0.9798030378901686,
16
+ "main_score": 0.9798030378901686,
17
+ "precision": 0.9775496578200633,
18
+ "recall": 0.9844767150726089
19
+ },
20
+ "eng_Latn-por_Latn": {
21
+ "accuracy": 0.9904857285928893,
22
+ "f1": 0.9873977633116341,
23
+ "main_score": 0.9873977633116341,
24
+ "precision": 0.9858955099315639,
25
+ "recall": 0.9904857285928893
26
+ },
27
+ "evaluation_time": 315.58,
28
+ "fra_Latn-por_Latn": {
29
+ "accuracy": 0.9779669504256384,
30
+ "f1": 0.971123351694208,
31
+ "main_score": 0.971123351694208,
32
+ "precision": 0.9677850108496078,
33
+ "recall": 0.9779669504256384
34
+ },
35
+ "ita_Latn-por_Latn": {
36
+ "accuracy": 0.9869804707060591,
37
+ "f1": 0.9829744616925388,
38
+ "main_score": 0.9829744616925388,
39
+ "precision": 0.9810549157068936,
40
+ "recall": 0.9869804707060591
41
+ },
42
+ "jpn_Jpan-por_Latn": {
43
+ "accuracy": 0.9679519278918377,
44
+ "f1": 0.957686529794692,
45
+ "main_score": 0.957686529794692,
46
+ "precision": 0.9525955600066767,
47
+ "recall": 0.9679519278918377
48
+ },
49
+ "kor_Hang-por_Latn": {
50
+ "accuracy": 0.9689534301452178,
51
+ "f1": 0.9593557002169921,
52
+ "main_score": 0.9593557002169921,
53
+ "precision": 0.9546820230345519,
54
+ "recall": 0.9689534301452178
55
+ },
56
+ "pol_Latn-por_Latn": {
57
+ "accuracy": 0.9799699549323986,
58
+ "f1": 0.9741612418627943,
59
+ "main_score": 0.9741612418627943,
60
+ "precision": 0.9713319979969955,
61
+ "recall": 0.9799699549323986
62
+ },
63
+ "rus_Cyrl-por_Latn": {
64
+ "accuracy": 0.9669504256384577,
65
+ "f1": 0.956685027541312,
66
+ "main_score": 0.956685027541312,
67
+ "precision": 0.9516775162744117,
68
+ "recall": 0.9669504256384577
69
+ },
70
+ "spa_Latn-por_Latn": {
71
+ "accuracy": 0.9889834752128193,
72
+ "f1": 0.985478217325989,
73
+ "main_score": 0.985478217325989,
74
+ "precision": 0.9837255883825738,
75
+ "recall": 0.9889834752128193
76
+ },
77
+ "swe_Latn-por_Latn": {
78
+ "accuracy": 0.9884827240861291,
79
+ "f1": 0.984727090635954,
80
+ "main_score": 0.984727090635954,
81
+ "precision": 0.9828910031714237,
82
+ "recall": 0.9884827240861291
83
+ },
84
+ "zho_Hant-por_Latn": {
85
+ "accuracy": 0.9684526790185278,
86
+ "f1": 0.959522617259222,
87
+ "main_score": 0.959522617259222,
88
+ "precision": 0.9552662326823568,
89
+ "recall": 0.9684526790185278
90
+ }
91
+ }
92
+ }
results/Amanda/bge_portuguese_v6/PTT_AskUbuntuDupQuestions.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_AskUbuntuDupQuestions",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 12.01,
7
+ "map": 0.5871499931945899,
8
+ "mrr": 0.7170887745679989
9
+ }
10
+ }
results/Amanda/bge_portuguese_v6/PTT_CQADupstackStatsRetrieval.json ADDED
@@ -0,0 +1,43 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_CQADupstackStatsRetrieval",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 1154.06,
7
+ "map_at_1": 0.1641,
8
+ "map_at_10": 0.21003,
9
+ "map_at_100": 0.2176,
10
+ "map_at_1000": 0.21862,
11
+ "map_at_20": 0.21333,
12
+ "map_at_3": 0.19249,
13
+ "map_at_5": 0.20368,
14
+ "mrr_at_1": 0.18712,
15
+ "mrr_at_10": 0.23483,
16
+ "mrr_at_100": 0.24201,
17
+ "mrr_at_1000": 0.24281,
18
+ "mrr_at_20": 0.2382,
19
+ "mrr_at_3": 0.21677,
20
+ "mrr_at_5": 0.22866,
21
+ "ndcg_at_1": 0.18712,
22
+ "ndcg_at_10": 0.24199,
23
+ "ndcg_at_100": 0.28338,
24
+ "ndcg_at_1000": 0.31107,
25
+ "ndcg_at_20": 0.25378,
26
+ "ndcg_at_3": 0.20857,
27
+ "ndcg_at_5": 0.22757,
28
+ "precision_at_1": 0.18712,
29
+ "precision_at_10": 0.03834,
30
+ "precision_at_100": 0.00658,
31
+ "precision_at_1000": 0.00097,
32
+ "precision_at_20": 0.02201,
33
+ "precision_at_3": 0.08947,
34
+ "precision_at_5": 0.06564,
35
+ "recall_at_1": 0.1641,
36
+ "recall_at_10": 0.3173,
37
+ "recall_at_100": 0.50974,
38
+ "recall_at_1000": 0.7187,
39
+ "recall_at_20": 0.36146,
40
+ "recall_at_3": 0.22689,
41
+ "recall_at_5": 0.27367
42
+ }
43
+ }
results/Amanda/bge_portuguese_v6/PTT_MedrxivClusteringS2S.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_MedrxivClusteringS2S",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 133.69,
7
+ "main_score": 0.24574152331605165,
8
+ "v_measure": 0.24574152331605165,
9
+ "v_measure_std": 0.010843354122673225,
10
+ "v_measures": [
11
+ 0.23696843618653413,
12
+ 0.23483748484109662,
13
+ 0.2360348323630235,
14
+ 0.235950205875761,
15
+ 0.2353593295033998,
16
+ 0.2553759269570152,
17
+ 0.2527774690663667,
18
+ 0.2662060045385906,
19
+ 0.2566302713792394,
20
+ 0.2472752724494892
21
+ ]
22
+ }
23
+ }
results/Amanda/bge_portuguese_v6/PTT_SprintDuplicateQuestions.json ADDED
@@ -0,0 +1,93 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_SprintDuplicateQuestions",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.9975544554455446,
8
+ "accuracy_threshold": 0.8014061450958252,
9
+ "ap": 0.9340558806075424,
10
+ "f1": 0.8720870015535991,
11
+ "f1_threshold": 0.7990845441818237,
12
+ "precision": 0.9044038668098818,
13
+ "recall": 0.842
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.9975544554455446,
17
+ "accuracy_threshold": 0.8014061450958252,
18
+ "ap": 0.9340558806075424,
19
+ "f1": 0.8720870015535991,
20
+ "f1_threshold": 0.7990845441818237,
21
+ "precision": 0.9044038668098818,
22
+ "recall": 0.842
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.9975544554455446,
26
+ "accuracy_threshold": 0.6302282810211182,
27
+ "ap": 0.9340558806075424,
28
+ "f1": 0.8720870015535991,
29
+ "f1_threshold": 0.6339012384414673,
30
+ "precision": 0.9044038668098818,
31
+ "recall": 0.842
32
+ },
33
+ "evaluation_time": 22.38,
34
+ "manhattan": {
35
+ "accuracy": 0.9975148514851485,
36
+ "accuracy_threshold": 15.930070877075195,
37
+ "ap": 0.9344473861455143,
38
+ "f1": 0.8712160082093381,
39
+ "f1_threshold": 16.173168182373047,
40
+ "precision": 0.8946259220231823,
41
+ "recall": 0.849
42
+ },
43
+ "max": {
44
+ "accuracy": 0.9975544554455446,
45
+ "ap": 0.9344473861455143,
46
+ "f1": 0.8720870015535991
47
+ }
48
+ },
49
+ "validation": {
50
+ "cos_sim": {
51
+ "accuracy": 0.9975148514851485,
52
+ "accuracy_threshold": 0.7869253158569336,
53
+ "ap": 0.93019369369862,
54
+ "f1": 0.8700155359917141,
55
+ "f1_threshold": 0.7869253158569336,
56
+ "precision": 0.9022556390977443,
57
+ "recall": 0.84
58
+ },
59
+ "dot": {
60
+ "accuracy": 0.9975148514851485,
61
+ "accuracy_threshold": 0.7869253158569336,
62
+ "ap": 0.9301936945331359,
63
+ "f1": 0.8700155359917141,
64
+ "f1_threshold": 0.7869253158569336,
65
+ "precision": 0.9022556390977443,
66
+ "recall": 0.84
67
+ },
68
+ "euclidean": {
69
+ "accuracy": 0.9975148514851485,
70
+ "accuracy_threshold": 0.6528010368347168,
71
+ "ap": 0.9301936936986199,
72
+ "f1": 0.8700155359917141,
73
+ "f1_threshold": 0.6528010368347168,
74
+ "precision": 0.9022556390977443,
75
+ "recall": 0.84
76
+ },
77
+ "evaluation_time": 23.04,
78
+ "manhattan": {
79
+ "accuracy": 0.9975247524752475,
80
+ "accuracy_threshold": 16.37934112548828,
81
+ "ap": 0.9315169831294592,
82
+ "f1": 0.8722772277227723,
83
+ "f1_threshold": 17.05251693725586,
84
+ "precision": 0.8637254901960785,
85
+ "recall": 0.881
86
+ },
87
+ "max": {
88
+ "accuracy": 0.9975247524752475,
89
+ "ap": 0.9315169831294592,
90
+ "f1": 0.8722772277227723
91
+ }
92
+ }
93
+ }
results/Amanda/bge_portuguese_v6/PTT_StackOverflowDupQuestions.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_StackOverflowDupQuestions",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 192.59,
7
+ "map": 0.45365620296066006,
8
+ "mrr": 0.4591718126644597
9
+ }
10
+ }
results/Amanda/bge_portuguese_v6/PTT_TwentyNewsgroupsClustering.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_TwentyNewsgroupsClustering",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "evaluation_time": 48.28,
7
+ "main_score": 0.3203252082759369,
8
+ "v_measure": 0.3203252082759369,
9
+ "v_measure_std": 0.013038394244103977,
10
+ "v_measures": [
11
+ 0.34464477221611034,
12
+ 0.3008189919274136,
13
+ 0.30303205031385755,
14
+ 0.3348185647197237,
15
+ 0.32259668239864503,
16
+ 0.3115956277457805,
17
+ 0.33027729830180125,
18
+ 0.32003375415418994,
19
+ 0.3212424329192006,
20
+ 0.3141919080626463
21
+ ]
22
+ }
23
+ }
results/Amanda/bge_portuguese_v6/PTT_TwitterSemEval2015.json ADDED
@@ -0,0 +1,49 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "dataset_revision": "main",
3
+ "mteb_dataset_name": "PTT_TwitterSemEval2015",
4
+ "mteb_version": 2,
5
+ "test": {
6
+ "cos_sim": {
7
+ "accuracy": 0.8337670580768011,
8
+ "accuracy_threshold": 0.7939179539680481,
9
+ "ap": 0.6487893788087082,
10
+ "f1": 0.6073446327683616,
11
+ "f1_threshold": 0.7287165522575378,
12
+ "precision": 0.5526992287917738,
13
+ "recall": 0.6739811912225705
14
+ },
15
+ "dot": {
16
+ "accuracy": 0.8337670580768011,
17
+ "accuracy_threshold": 0.7939180135726929,
18
+ "ap": 0.6487893418943692,
19
+ "f1": 0.6073446327683616,
20
+ "f1_threshold": 0.7287165522575378,
21
+ "precision": 0.5526992287917738,
22
+ "recall": 0.6739811912225705
23
+ },
24
+ "euclidean": {
25
+ "accuracy": 0.8337670580768011,
26
+ "accuracy_threshold": 0.6420000791549683,
27
+ "ap": 0.6487893162168841,
28
+ "f1": 0.6073446327683616,
29
+ "f1_threshold": 0.7365913391113281,
30
+ "precision": 0.5526992287917738,
31
+ "recall": 0.6739811912225705
32
+ },
33
+ "evaluation_time": 23.59,
34
+ "manhattan": {
35
+ "accuracy": 0.8335131704220883,
36
+ "accuracy_threshold": 16.44284439086914,
37
+ "ap": 0.6470422250419247,
38
+ "f1": 0.6064184852374839,
39
+ "f1_threshold": 18.61992835998535,
40
+ "precision": 0.5517402476056996,
41
+ "recall": 0.6731262467939584
42
+ },
43
+ "max": {
44
+ "accuracy": 0.8337670580768011,
45
+ "ap": 0.6487893788087082,
46
+ "f1": 0.6073446327683616
47
+ }
48
+ }
49
+ }