lmilliken commited on
Commit
03e1db8
1 Parent(s): b3b71a4

add mteb benchmark results

Browse files
Files changed (1) hide show
  1. README.md +1700 -0
README.md CHANGED
@@ -4,10 +4,1710 @@ tags:
4
  - finetuner
5
  - feature-extraction
6
  - sentence-similarity
 
7
  datasets:
8
  - jinaai/negation-dataset
9
  language: en
10
  license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <br><br>
 
4
  - finetuner
5
  - feature-extraction
6
  - sentence-similarity
7
+ - mteb
8
  datasets:
9
  - jinaai/negation-dataset
10
  language: en
11
  license: apache-2.0
12
+ model-index:
13
+ - name: jina-embedding-s-en-v1
14
+ results:
15
+ - task:
16
+ type: Classification
17
+ dataset:
18
+ type: mteb/amazon_counterfactual
19
+ name: MTEB AmazonCounterfactualClassification (en)
20
+ config: en
21
+ split: test
22
+ revision: e8379541af4e31359cca9fbcf4b00f2671dba205
23
+ metrics:
24
+ - type: accuracy
25
+ value: 64.58208955223881
26
+ - type: ap
27
+ value: 27.24359671025387
28
+ - type: f1
29
+ value: 58.201387941715495
30
+ - task:
31
+ type: Classification
32
+ dataset:
33
+ type: mteb/amazon_polarity
34
+ name: MTEB AmazonPolarityClassification
35
+ config: default
36
+ split: test
37
+ revision: e2d317d38cd51312af73b3d32a06d1a08b442046
38
+ metrics:
39
+ - type: accuracy
40
+ value: 61.926550000000006
41
+ - type: ap
42
+ value: 58.40954250092862
43
+ - type: f1
44
+ value: 59.921771639047904
45
+ - task:
46
+ type: Classification
47
+ dataset:
48
+ type: mteb/amazon_reviews_multi
49
+ name: MTEB AmazonReviewsClassification (en)
50
+ config: en
51
+ split: test
52
+ revision: 1399c76144fd37290681b995c656ef9b2e06e26d
53
+ metrics:
54
+ - type: accuracy
55
+ value: 28.499999999999996
56
+ - type: f1
57
+ value: 27.160929516206465
58
+ - task:
59
+ type: Retrieval
60
+ dataset:
61
+ type: arguana
62
+ name: MTEB ArguAna
63
+ config: default
64
+ split: test
65
+ revision: None
66
+ metrics:
67
+ - type: map_at_1
68
+ value: 22.262
69
+ - type: map_at_10
70
+ value: 36.677
71
+ - type: map_at_100
72
+ value: 37.839
73
+ - type: map_at_1000
74
+ value: 37.857
75
+ - type: map_at_3
76
+ value: 31.685999999999996
77
+ - type: map_at_5
78
+ value: 34.544999999999995
79
+ - type: mrr_at_1
80
+ value: 22.404
81
+ - type: mrr_at_10
82
+ value: 36.713
83
+ - type: mrr_at_100
84
+ value: 37.881
85
+ - type: mrr_at_1000
86
+ value: 37.899
87
+ - type: mrr_at_3
88
+ value: 31.709
89
+ - type: mrr_at_5
90
+ value: 34.629
91
+ - type: ndcg_at_1
92
+ value: 22.262
93
+ - type: ndcg_at_10
94
+ value: 45.18
95
+ - type: ndcg_at_100
96
+ value: 50.4
97
+ - type: ndcg_at_1000
98
+ value: 50.841
99
+ - type: ndcg_at_3
100
+ value: 34.882000000000005
101
+ - type: ndcg_at_5
102
+ value: 40.036
103
+ - type: precision_at_1
104
+ value: 22.262
105
+ - type: precision_at_10
106
+ value: 7.255000000000001
107
+ - type: precision_at_100
108
+ value: 0.959
109
+ - type: precision_at_1000
110
+ value: 0.099
111
+ - type: precision_at_3
112
+ value: 14.723
113
+ - type: precision_at_5
114
+ value: 11.337
115
+ - type: recall_at_1
116
+ value: 22.262
117
+ - type: recall_at_10
118
+ value: 72.54599999999999
119
+ - type: recall_at_100
120
+ value: 95.946
121
+ - type: recall_at_1000
122
+ value: 99.36
123
+ - type: recall_at_3
124
+ value: 44.168
125
+ - type: recall_at_5
126
+ value: 56.686
127
+ - task:
128
+ type: Clustering
129
+ dataset:
130
+ type: mteb/arxiv-clustering-p2p
131
+ name: MTEB ArxivClusteringP2P
132
+ config: default
133
+ split: test
134
+ revision: a122ad7f3f0291bf49cc6f4d32aa80929df69d5d
135
+ metrics:
136
+ - type: v_measure
137
+ value: 34.97570470844357
138
+ - task:
139
+ type: Clustering
140
+ dataset:
141
+ type: mteb/arxiv-clustering-s2s
142
+ name: MTEB ArxivClusteringS2S
143
+ config: default
144
+ split: test
145
+ revision: f910caf1a6075f7329cdf8c1a6135696f37dbd53
146
+ metrics:
147
+ - type: v_measure
148
+ value: 24.372872291698265
149
+ - task:
150
+ type: Reranking
151
+ dataset:
152
+ type: mteb/askubuntudupquestions-reranking
153
+ name: MTEB AskUbuntuDupQuestions
154
+ config: default
155
+ split: test
156
+ revision: 2000358ca161889fa9c082cb41daa8dcfb161a54
157
+ metrics:
158
+ - type: map
159
+ value: 60.58753030525579
160
+ - type: mrr
161
+ value: 75.03484588664644
162
+ - task:
163
+ type: STS
164
+ dataset:
165
+ type: mteb/biosses-sts
166
+ name: MTEB BIOSSES
167
+ config: default
168
+ split: test
169
+ revision: d3fb88f8f02e40887cd149695127462bbcf29b4a
170
+ metrics:
171
+ - type: cos_sim_pearson
172
+ value: 85.21378425036666
173
+ - type: cos_sim_spearman
174
+ value: 80.45665253651644
175
+ - type: euclidean_pearson
176
+ value: 46.71436482437946
177
+ - type: euclidean_spearman
178
+ value: 45.13476336596072
179
+ - type: manhattan_pearson
180
+ value: 47.06449770246884
181
+ - type: manhattan_spearman
182
+ value: 45.498627078529
183
+ - task:
184
+ type: Classification
185
+ dataset:
186
+ type: mteb/banking77
187
+ name: MTEB Banking77Classification
188
+ config: default
189
+ split: test
190
+ revision: 0fd18e25b25c072e09e0d92ab615fda904d66300
191
+ metrics:
192
+ - type: accuracy
193
+ value: 74.48701298701299
194
+ - type: f1
195
+ value: 73.30813366682357
196
+ - task:
197
+ type: Clustering
198
+ dataset:
199
+ type: mteb/biorxiv-clustering-p2p
200
+ name: MTEB BiorxivClusteringP2P
201
+ config: default
202
+ split: test
203
+ revision: 65b79d1d13f80053f67aca9498d9402c2d9f1f40
204
+ metrics:
205
+ - type: v_measure
206
+ value: 29.66289767477026
207
+ - task:
208
+ type: Clustering
209
+ dataset:
210
+ type: mteb/biorxiv-clustering-s2s
211
+ name: MTEB BiorxivClusteringS2S
212
+ config: default
213
+ split: test
214
+ revision: 258694dd0231531bc1fd9de6ceb52a0853c6d908
215
+ metrics:
216
+ - type: v_measure
217
+ value: 22.324367934720776
218
+ - task:
219
+ type: Retrieval
220
+ dataset:
221
+ type: climate-fever
222
+ name: MTEB ClimateFEVER
223
+ config: default
224
+ split: test
225
+ revision: None
226
+ metrics:
227
+ - type: map_at_1
228
+ value: 6.524000000000001
229
+ - type: map_at_10
230
+ value: 11.187
231
+ - type: map_at_100
232
+ value: 12.389999999999999
233
+ - type: map_at_1000
234
+ value: 12.559000000000001
235
+ - type: map_at_3
236
+ value: 9.386
237
+ - type: map_at_5
238
+ value: 10.295
239
+ - type: mrr_at_1
240
+ value: 13.941
241
+ - type: mrr_at_10
242
+ value: 22.742
243
+ - type: mrr_at_100
244
+ value: 23.896
245
+ - type: mrr_at_1000
246
+ value: 23.965
247
+ - type: mrr_at_3
248
+ value: 19.881
249
+ - type: mrr_at_5
250
+ value: 21.555
251
+ - type: ndcg_at_1
252
+ value: 13.941
253
+ - type: ndcg_at_10
254
+ value: 16.619999999999997
255
+ - type: ndcg_at_100
256
+ value: 22.415
257
+ - type: ndcg_at_1000
258
+ value: 26.05
259
+ - type: ndcg_at_3
260
+ value: 13.148000000000001
261
+ - type: ndcg_at_5
262
+ value: 14.433000000000002
263
+ - type: precision_at_1
264
+ value: 13.941
265
+ - type: precision_at_10
266
+ value: 5.153
267
+ - type: precision_at_100
268
+ value: 1.124
269
+ - type: precision_at_1000
270
+ value: 0.178
271
+ - type: precision_at_3
272
+ value: 9.685
273
+ - type: precision_at_5
274
+ value: 7.582999999999999
275
+ - type: recall_at_1
276
+ value: 6.524000000000001
277
+ - type: recall_at_10
278
+ value: 21.041999999999998
279
+ - type: recall_at_100
280
+ value: 41.515
281
+ - type: recall_at_1000
282
+ value: 62.507999999999996
283
+ - type: recall_at_3
284
+ value: 12.549
285
+ - type: recall_at_5
286
+ value: 15.939999999999998
287
+ - task:
288
+ type: Retrieval
289
+ dataset:
290
+ type: dbpedia-entity
291
+ name: MTEB DBPedia
292
+ config: default
293
+ split: test
294
+ revision: None
295
+ metrics:
296
+ - type: map_at_1
297
+ value: 6.483
298
+ - type: map_at_10
299
+ value: 11.955
300
+ - type: map_at_100
301
+ value: 15.470999999999998
302
+ - type: map_at_1000
303
+ value: 16.308
304
+ - type: map_at_3
305
+ value: 9.292
306
+ - type: map_at_5
307
+ value: 10.459
308
+ - type: mrr_at_1
309
+ value: 50.74999999999999
310
+ - type: mrr_at_10
311
+ value: 58.743
312
+ - type: mrr_at_100
313
+ value: 59.41499999999999
314
+ - type: mrr_at_1000
315
+ value: 59.431999999999995
316
+ - type: mrr_at_3
317
+ value: 56.708000000000006
318
+ - type: mrr_at_5
319
+ value: 57.80800000000001
320
+ - type: ndcg_at_1
321
+ value: 39.0
322
+ - type: ndcg_at_10
323
+ value: 26.721
324
+ - type: ndcg_at_100
325
+ value: 29.366999999999997
326
+ - type: ndcg_at_1000
327
+ value: 35.618
328
+ - type: ndcg_at_3
329
+ value: 31.244
330
+ - type: ndcg_at_5
331
+ value: 28.614
332
+ - type: precision_at_1
333
+ value: 50.74999999999999
334
+ - type: precision_at_10
335
+ value: 20.45
336
+ - type: precision_at_100
337
+ value: 6.0600000000000005
338
+ - type: precision_at_1000
339
+ value: 1.346
340
+ - type: precision_at_3
341
+ value: 33.917
342
+ - type: precision_at_5
343
+ value: 26.950000000000003
344
+ - type: recall_at_1
345
+ value: 6.483
346
+ - type: recall_at_10
347
+ value: 16.215
348
+ - type: recall_at_100
349
+ value: 33.382
350
+ - type: recall_at_1000
351
+ value: 54.445
352
+ - type: recall_at_3
353
+ value: 10.6
354
+ - type: recall_at_5
355
+ value: 12.889999999999999
356
+ - task:
357
+ type: Classification
358
+ dataset:
359
+ type: mteb/emotion
360
+ name: MTEB EmotionClassification
361
+ config: default
362
+ split: test
363
+ revision: 4f58c6b202a23cf9a4da393831edf4f9183cad37
364
+ metrics:
365
+ - type: accuracy
366
+ value: 34.39
367
+ - type: f1
368
+ value: 31.334865751249474
369
+ - task:
370
+ type: Retrieval
371
+ dataset:
372
+ type: fever
373
+ name: MTEB FEVER
374
+ config: default
375
+ split: test
376
+ revision: None
377
+ metrics:
378
+ - type: map_at_1
379
+ value: 44.698
380
+ - type: map_at_10
381
+ value: 55.30500000000001
382
+ - type: map_at_100
383
+ value: 55.838
384
+ - type: map_at_1000
385
+ value: 55.87
386
+ - type: map_at_3
387
+ value: 52.884
388
+ - type: map_at_5
389
+ value: 54.352000000000004
390
+ - type: mrr_at_1
391
+ value: 48.32
392
+ - type: mrr_at_10
393
+ value: 59.39
394
+ - type: mrr_at_100
395
+ value: 59.89
396
+ - type: mrr_at_1000
397
+ value: 59.913000000000004
398
+ - type: mrr_at_3
399
+ value: 56.977999999999994
400
+ - type: mrr_at_5
401
+ value: 58.44200000000001
402
+ - type: ndcg_at_1
403
+ value: 48.32
404
+ - type: ndcg_at_10
405
+ value: 61.23800000000001
406
+ - type: ndcg_at_100
407
+ value: 63.79
408
+ - type: ndcg_at_1000
409
+ value: 64.575
410
+ - type: ndcg_at_3
411
+ value: 56.489999999999995
412
+ - type: ndcg_at_5
413
+ value: 59.016999999999996
414
+ - type: precision_at_1
415
+ value: 48.32
416
+ - type: precision_at_10
417
+ value: 8.288
418
+ - type: precision_at_100
419
+ value: 0.964
420
+ - type: precision_at_1000
421
+ value: 0.104
422
+ - type: precision_at_3
423
+ value: 22.867
424
+ - type: precision_at_5
425
+ value: 15.098
426
+ - type: recall_at_1
427
+ value: 44.698
428
+ - type: recall_at_10
429
+ value: 75.752
430
+ - type: recall_at_100
431
+ value: 87.402
432
+ - type: recall_at_1000
433
+ value: 93.316
434
+ - type: recall_at_3
435
+ value: 62.82600000000001
436
+ - type: recall_at_5
437
+ value: 69.01899999999999
438
+ - task:
439
+ type: Retrieval
440
+ dataset:
441
+ type: fiqa
442
+ name: MTEB FiQA2018
443
+ config: default
444
+ split: test
445
+ revision: None
446
+ metrics:
447
+ - type: map_at_1
448
+ value: 12.119
449
+ - type: map_at_10
450
+ value: 20.299
451
+ - type: map_at_100
452
+ value: 21.863
453
+ - type: map_at_1000
454
+ value: 22.064
455
+ - type: map_at_3
456
+ value: 17.485999999999997
457
+ - type: map_at_5
458
+ value: 19.148
459
+ - type: mrr_at_1
460
+ value: 24.383
461
+ - type: mrr_at_10
462
+ value: 33.074
463
+ - type: mrr_at_100
464
+ value: 34.03
465
+ - type: mrr_at_1000
466
+ value: 34.102
467
+ - type: mrr_at_3
468
+ value: 30.736
469
+ - type: mrr_at_5
470
+ value: 32.202
471
+ - type: ndcg_at_1
472
+ value: 24.383
473
+ - type: ndcg_at_10
474
+ value: 26.645999999999997
475
+ - type: ndcg_at_100
476
+ value: 33.348
477
+ - type: ndcg_at_1000
478
+ value: 37.294
479
+ - type: ndcg_at_3
480
+ value: 23.677
481
+ - type: ndcg_at_5
482
+ value: 24.935
483
+ - type: precision_at_1
484
+ value: 24.383
485
+ - type: precision_at_10
486
+ value: 7.654
487
+ - type: precision_at_100
488
+ value: 1.461
489
+ - type: precision_at_1000
490
+ value: 0.214
491
+ - type: precision_at_3
492
+ value: 16.101
493
+ - type: precision_at_5
494
+ value: 12.222
495
+ - type: recall_at_1
496
+ value: 12.119
497
+ - type: recall_at_10
498
+ value: 32.531
499
+ - type: recall_at_100
500
+ value: 58.028999999999996
501
+ - type: recall_at_1000
502
+ value: 82.513
503
+ - type: recall_at_3
504
+ value: 21.787
505
+ - type: recall_at_5
506
+ value: 27.229999999999997
507
+ - task:
508
+ type: Retrieval
509
+ dataset:
510
+ type: hotpotqa
511
+ name: MTEB HotpotQA
512
+ config: default
513
+ split: test
514
+ revision: None
515
+ metrics:
516
+ - type: map_at_1
517
+ value: 26.057000000000002
518
+ - type: map_at_10
519
+ value: 34.892
520
+ - type: map_at_100
521
+ value: 35.687000000000005
522
+ - type: map_at_1000
523
+ value: 35.763
524
+ - type: map_at_3
525
+ value: 32.879000000000005
526
+ - type: map_at_5
527
+ value: 34.105000000000004
528
+ - type: mrr_at_1
529
+ value: 52.113
530
+ - type: mrr_at_10
531
+ value: 58.940000000000005
532
+ - type: mrr_at_100
533
+ value: 59.438
534
+ - type: mrr_at_1000
535
+ value: 59.473
536
+ - type: mrr_at_3
537
+ value: 57.299
538
+ - type: mrr_at_5
539
+ value: 58.353
540
+ - type: ndcg_at_1
541
+ value: 52.113
542
+ - type: ndcg_at_10
543
+ value: 43.105
544
+ - type: ndcg_at_100
545
+ value: 46.44
546
+ - type: ndcg_at_1000
547
+ value: 48.241
548
+ - type: ndcg_at_3
549
+ value: 39.566
550
+ - type: ndcg_at_5
551
+ value: 41.508
552
+ - type: precision_at_1
553
+ value: 52.113
554
+ - type: precision_at_10
555
+ value: 8.892999999999999
556
+ - type: precision_at_100
557
+ value: 1.1520000000000001
558
+ - type: precision_at_1000
559
+ value: 0.13899999999999998
560
+ - type: precision_at_3
561
+ value: 24.398
562
+ - type: precision_at_5
563
+ value: 16.181
564
+ - type: recall_at_1
565
+ value: 26.057000000000002
566
+ - type: recall_at_10
567
+ value: 44.463
568
+ - type: recall_at_100
569
+ value: 57.616
570
+ - type: recall_at_1000
571
+ value: 69.65599999999999
572
+ - type: recall_at_3
573
+ value: 36.597
574
+ - type: recall_at_5
575
+ value: 40.452
576
+ - task:
577
+ type: Classification
578
+ dataset:
579
+ type: mteb/imdb
580
+ name: MTEB ImdbClassification
581
+ config: default
582
+ split: test
583
+ revision: 3d86128a09e091d6018b6d26cad27f2739fc2db7
584
+ metrics:
585
+ - type: accuracy
586
+ value: 58.268399999999986
587
+ - type: ap
588
+ value: 55.03852332714837
589
+ - type: f1
590
+ value: 57.23656436062262
591
+ - task:
592
+ type: Retrieval
593
+ dataset:
594
+ type: msmarco
595
+ name: MTEB MSMARCO
596
+ config: default
597
+ split: dev
598
+ revision: None
599
+ metrics:
600
+ - type: map_at_1
601
+ value: 14.273
602
+ - type: map_at_10
603
+ value: 23.953
604
+ - type: map_at_100
605
+ value: 25.207
606
+ - type: map_at_1000
607
+ value: 25.285999999999998
608
+ - type: map_at_3
609
+ value: 20.727
610
+ - type: map_at_5
611
+ value: 22.492
612
+ - type: mrr_at_1
613
+ value: 14.685
614
+ - type: mrr_at_10
615
+ value: 24.423000000000002
616
+ - type: mrr_at_100
617
+ value: 25.64
618
+ - type: mrr_at_1000
619
+ value: 25.713
620
+ - type: mrr_at_3
621
+ value: 21.213
622
+ - type: mrr_at_5
623
+ value: 22.979
624
+ - type: ndcg_at_1
625
+ value: 14.685
626
+ - type: ndcg_at_10
627
+ value: 29.698
628
+ - type: ndcg_at_100
629
+ value: 36.010999999999996
630
+ - type: ndcg_at_1000
631
+ value: 38.102999999999994
632
+ - type: ndcg_at_3
633
+ value: 23.0
634
+ - type: ndcg_at_5
635
+ value: 26.186
636
+ - type: precision_at_1
637
+ value: 14.685
638
+ - type: precision_at_10
639
+ value: 4.954
640
+ - type: precision_at_100
641
+ value: 0.815
642
+ - type: precision_at_1000
643
+ value: 0.099
644
+ - type: precision_at_3
645
+ value: 10.038
646
+ - type: precision_at_5
647
+ value: 7.636
648
+ - type: recall_at_1
649
+ value: 14.273
650
+ - type: recall_at_10
651
+ value: 47.559000000000005
652
+ - type: recall_at_100
653
+ value: 77.375
654
+ - type: recall_at_1000
655
+ value: 93.616
656
+ - type: recall_at_3
657
+ value: 29.110999999999997
658
+ - type: recall_at_5
659
+ value: 36.825
660
+ - task:
661
+ type: Classification
662
+ dataset:
663
+ type: mteb/mtop_domain
664
+ name: MTEB MTOPDomainClassification (en)
665
+ config: en
666
+ split: test
667
+ revision: d80d48c1eb48d3562165c59d59d0034df9fff0bf
668
+ metrics:
669
+ - type: accuracy
670
+ value: 89.85636114911081
671
+ - type: f1
672
+ value: 89.65403786390279
673
+ - task:
674
+ type: Classification
675
+ dataset:
676
+ type: mteb/mtop_intent
677
+ name: MTEB MTOPIntentClassification (en)
678
+ config: en
679
+ split: test
680
+ revision: ae001d0e6b1228650b7bd1c2c65fb50ad11a8aba
681
+ metrics:
682
+ - type: accuracy
683
+ value: 59.03784769721842
684
+ - type: f1
685
+ value: 42.57604111096128
686
+ - task:
687
+ type: Classification
688
+ dataset:
689
+ type: mteb/amazon_massive_intent
690
+ name: MTEB MassiveIntentClassification (en)
691
+ config: en
692
+ split: test
693
+ revision: 31efe3c427b0bae9c22cbb560b8f15491cc6bed7
694
+ metrics:
695
+ - type: accuracy
696
+ value: 65.00336247478144
697
+ - type: f1
698
+ value: 63.12578076844032
699
+ - task:
700
+ type: Classification
701
+ dataset:
702
+ type: mteb/amazon_massive_scenario
703
+ name: MTEB MassiveScenarioClassification (en)
704
+ config: en
705
+ split: test
706
+ revision: 7d571f92784cd94a019292a1f45445077d0ef634
707
+ metrics:
708
+ - type: accuracy
709
+ value: 72.14862138533962
710
+ - type: f1
711
+ value: 71.91174720216141
712
+ - task:
713
+ type: Clustering
714
+ dataset:
715
+ type: mteb/medrxiv-clustering-p2p
716
+ name: MTEB MedrxivClusteringP2P
717
+ config: default
718
+ split: test
719
+ revision: e7a26af6f3ae46b30dde8737f02c07b1505bcc73
720
+ metrics:
721
+ - type: v_measure
722
+ value: 28.259326082067094
723
+ - task:
724
+ type: Clustering
725
+ dataset:
726
+ type: mteb/medrxiv-clustering-s2s
727
+ name: MTEB MedrxivClusteringS2S
728
+ config: default
729
+ split: test
730
+ revision: 35191c8c0dca72d8ff3efcd72aa802307d469663
731
+ metrics:
732
+ - type: v_measure
733
+ value: 23.874256261395775
734
+ - task:
735
+ type: Reranking
736
+ dataset:
737
+ type: mteb/mind_small
738
+ name: MTEB MindSmallReranking
739
+ config: default
740
+ split: test
741
+ revision: 3bdac13927fdc888b903db93b2ffdbd90b295a69
742
+ metrics:
743
+ - type: map
744
+ value: 29.251614283788385
745
+ - type: mrr
746
+ value: 29.9695581475798
747
+ - task:
748
+ type: Retrieval
749
+ dataset:
750
+ type: nfcorpus
751
+ name: MTEB NFCorpus
752
+ config: default
753
+ split: test
754
+ revision: None
755
+ metrics:
756
+ - type: map_at_1
757
+ value: 3.9309999999999996
758
+ - type: map_at_10
759
+ value: 8.472
760
+ - type: map_at_100
761
+ value: 10.461
762
+ - type: map_at_1000
763
+ value: 11.588
764
+ - type: map_at_3
765
+ value: 6.343999999999999
766
+ - type: map_at_5
767
+ value: 7.379
768
+ - type: mrr_at_1
769
+ value: 35.913000000000004
770
+ - type: mrr_at_10
771
+ value: 43.91
772
+ - type: mrr_at_100
773
+ value: 44.519999999999996
774
+ - type: mrr_at_1000
775
+ value: 44.59
776
+ - type: mrr_at_3
777
+ value: 41.589
778
+ - type: mrr_at_5
779
+ value: 42.626
780
+ - type: ndcg_at_1
781
+ value: 34.52
782
+ - type: ndcg_at_10
783
+ value: 25.128
784
+ - type: ndcg_at_100
785
+ value: 22.917
786
+ - type: ndcg_at_1000
787
+ value: 31.64
788
+ - type: ndcg_at_3
789
+ value: 29.866999999999997
790
+ - type: ndcg_at_5
791
+ value: 27.494000000000003
792
+ - type: precision_at_1
793
+ value: 35.913000000000004
794
+ - type: precision_at_10
795
+ value: 18.607000000000003
796
+ - type: precision_at_100
797
+ value: 6.006
798
+ - type: precision_at_1000
799
+ value: 1.814
800
+ - type: precision_at_3
801
+ value: 28.277
802
+ - type: precision_at_5
803
+ value: 23.777
804
+ - type: recall_at_1
805
+ value: 3.9309999999999996
806
+ - type: recall_at_10
807
+ value: 11.684
808
+ - type: recall_at_100
809
+ value: 24.212
810
+ - type: recall_at_1000
811
+ value: 55.36
812
+ - type: recall_at_3
813
+ value: 7.329
814
+ - type: recall_at_5
815
+ value: 9.059000000000001
816
+ - task:
817
+ type: Retrieval
818
+ dataset:
819
+ type: nq
820
+ name: MTEB NQ
821
+ config: default
822
+ split: test
823
+ revision: None
824
+ metrics:
825
+ - type: map_at_1
826
+ value: 19.03
827
+ - type: map_at_10
828
+ value: 30.990000000000002
829
+ - type: map_at_100
830
+ value: 32.211
831
+ - type: map_at_1000
832
+ value: 32.267
833
+ - type: map_at_3
834
+ value: 26.833000000000002
835
+ - type: map_at_5
836
+ value: 29.128
837
+ - type: mrr_at_1
838
+ value: 21.523999999999997
839
+ - type: mrr_at_10
840
+ value: 33.085
841
+ - type: mrr_at_100
842
+ value: 34.096
843
+ - type: mrr_at_1000
844
+ value: 34.139
845
+ - type: mrr_at_3
846
+ value: 29.354999999999997
847
+ - type: mrr_at_5
848
+ value: 31.441999999999997
849
+ - type: ndcg_at_1
850
+ value: 21.495
851
+ - type: ndcg_at_10
852
+ value: 37.971
853
+ - type: ndcg_at_100
854
+ value: 43.492999999999995
855
+ - type: ndcg_at_1000
856
+ value: 44.925
857
+ - type: ndcg_at_3
858
+ value: 29.808
859
+ - type: ndcg_at_5
860
+ value: 33.748
861
+ - type: precision_at_1
862
+ value: 21.495
863
+ - type: precision_at_10
864
+ value: 6.819
865
+ - type: precision_at_100
866
+ value: 0.991
867
+ - type: precision_at_1000
868
+ value: 0.11299999999999999
869
+ - type: precision_at_3
870
+ value: 13.886000000000001
871
+ - type: precision_at_5
872
+ value: 10.574
873
+ - type: recall_at_1
874
+ value: 19.03
875
+ - type: recall_at_10
876
+ value: 57.493
877
+ - type: recall_at_100
878
+ value: 82.03200000000001
879
+ - type: recall_at_1000
880
+ value: 92.879
881
+ - type: recall_at_3
882
+ value: 35.899
883
+ - type: recall_at_5
884
+ value: 45.092
885
+ - task:
886
+ type: Retrieval
887
+ dataset:
888
+ type: quora
889
+ name: MTEB QuoraRetrieval
890
+ config: default
891
+ split: test
892
+ revision: None
893
+ metrics:
894
+ - type: map_at_1
895
+ value: 67.97
896
+ - type: map_at_10
897
+ value: 81.478
898
+ - type: map_at_100
899
+ value: 82.147
900
+ - type: map_at_1000
901
+ value: 82.172
902
+ - type: map_at_3
903
+ value: 78.456
904
+ - type: map_at_5
905
+ value: 80.337
906
+ - type: mrr_at_1
907
+ value: 78.24
908
+ - type: mrr_at_10
909
+ value: 84.941
910
+ - type: mrr_at_100
911
+ value: 85.08099999999999
912
+ - type: mrr_at_1000
913
+ value: 85.083
914
+ - type: mrr_at_3
915
+ value: 83.743
916
+ - type: mrr_at_5
917
+ value: 84.553
918
+ - type: ndcg_at_1
919
+ value: 78.24
920
+ - type: ndcg_at_10
921
+ value: 85.61999999999999
922
+ - type: ndcg_at_100
923
+ value: 87.113
924
+ - type: ndcg_at_1000
925
+ value: 87.318
926
+ - type: ndcg_at_3
927
+ value: 82.403
928
+ - type: ndcg_at_5
929
+ value: 84.15700000000001
930
+ - type: precision_at_1
931
+ value: 78.24
932
+ - type: precision_at_10
933
+ value: 12.979
934
+ - type: precision_at_100
935
+ value: 1.503
936
+ - type: precision_at_1000
937
+ value: 0.156
938
+ - type: precision_at_3
939
+ value: 35.9
940
+ - type: precision_at_5
941
+ value: 23.704
942
+ - type: recall_at_1
943
+ value: 67.97
944
+ - type: recall_at_10
945
+ value: 93.563
946
+ - type: recall_at_100
947
+ value: 98.834
948
+ - type: recall_at_1000
949
+ value: 99.901
950
+ - type: recall_at_3
951
+ value: 84.319
952
+ - type: recall_at_5
953
+ value: 89.227
954
+ - task:
955
+ type: Clustering
956
+ dataset:
957
+ type: mteb/reddit-clustering
958
+ name: MTEB RedditClustering
959
+ config: default
960
+ split: test
961
+ revision: 24640382cdbf8abc73003fb0fa6d111a705499eb
962
+ metrics:
963
+ - type: v_measure
964
+ value: 35.853649010160694
965
+ - task:
966
+ type: Clustering
967
+ dataset:
968
+ type: mteb/reddit-clustering-p2p
969
+ name: MTEB RedditClusteringP2P
970
+ config: default
971
+ split: test
972
+ revision: 282350215ef01743dc01b456c7f5241fa8937f16
973
+ metrics:
974
+ - type: v_measure
975
+ value: 47.270443152349415
976
+ - task:
977
+ type: Retrieval
978
+ dataset:
979
+ type: scidocs
980
+ name: MTEB SCIDOCS
981
+ config: default
982
+ split: test
983
+ revision: None
984
+ metrics:
985
+ - type: map_at_1
986
+ value: 3.803
987
+ - type: map_at_10
988
+ value: 8.790000000000001
989
+ - type: map_at_100
990
+ value: 10.313
991
+ - type: map_at_1000
992
+ value: 10.562000000000001
993
+ - type: map_at_3
994
+ value: 6.483
995
+ - type: map_at_5
996
+ value: 7.591
997
+ - type: mrr_at_1
998
+ value: 18.7
999
+ - type: mrr_at_10
1000
+ value: 27.349
1001
+ - type: mrr_at_100
1002
+ value: 28.474
1003
+ - type: mrr_at_1000
1004
+ value: 28.544999999999998
1005
+ - type: mrr_at_3
1006
+ value: 24.567
1007
+ - type: mrr_at_5
1008
+ value: 26.172
1009
+ - type: ndcg_at_1
1010
+ value: 18.7
1011
+ - type: ndcg_at_10
1012
+ value: 15.155
1013
+ - type: ndcg_at_100
1014
+ value: 21.63
1015
+ - type: ndcg_at_1000
1016
+ value: 26.595999999999997
1017
+ - type: ndcg_at_3
1018
+ value: 14.706
1019
+ - type: ndcg_at_5
1020
+ value: 12.681999999999999
1021
+ - type: precision_at_1
1022
+ value: 18.7
1023
+ - type: precision_at_10
1024
+ value: 7.6899999999999995
1025
+ - type: precision_at_100
1026
+ value: 1.7080000000000002
1027
+ - type: precision_at_1000
1028
+ value: 0.291
1029
+ - type: precision_at_3
1030
+ value: 13.567000000000002
1031
+ - type: precision_at_5
1032
+ value: 10.9
1033
+ - type: recall_at_1
1034
+ value: 3.803
1035
+ - type: recall_at_10
1036
+ value: 15.607
1037
+ - type: recall_at_100
1038
+ value: 34.717999999999996
1039
+ - type: recall_at_1000
1040
+ value: 59.150000000000006
1041
+ - type: recall_at_3
1042
+ value: 8.258000000000001
1043
+ - type: recall_at_5
1044
+ value: 11.063
1045
+ - task:
1046
+ type: STS
1047
+ dataset:
1048
+ type: mteb/sickr-sts
1049
+ name: MTEB SICK-R
1050
+ config: default
1051
+ split: test
1052
+ revision: a6ea5a8cab320b040a23452cc28066d9beae2cee
1053
+ metrics:
1054
+ - type: cos_sim_pearson
1055
+ value: 81.05755556071047
1056
+ - type: cos_sim_spearman
1057
+ value: 72.44408263672771
1058
+ - type: euclidean_pearson
1059
+ value: 71.65314814604668
1060
+ - type: euclidean_spearman
1061
+ value: 65.1833695751109
1062
+ - type: manhattan_pearson
1063
+ value: 71.81874115177355
1064
+ - type: manhattan_spearman
1065
+ value: 65.45940792270201
1066
+ - task:
1067
+ type: STS
1068
+ dataset:
1069
+ type: mteb/sts12-sts
1070
+ name: MTEB STS12
1071
+ config: default
1072
+ split: test
1073
+ revision: a0d554a64d88156834ff5ae9920b964011b16384
1074
+ metrics:
1075
+ - type: cos_sim_pearson
1076
+ value: 81.75836272926722
1077
+ - type: cos_sim_spearman
1078
+ value: 73.63905703662927
1079
+ - type: euclidean_pearson
1080
+ value: 67.58539517215293
1081
+ - type: euclidean_spearman
1082
+ value: 58.88440181413321
1083
+ - type: manhattan_pearson
1084
+ value: 66.56872028174024
1085
+ - type: manhattan_spearman
1086
+ value: 58.48195528793699
1087
+ - task:
1088
+ type: STS
1089
+ dataset:
1090
+ type: mteb/sts13-sts
1091
+ name: MTEB STS13
1092
+ config: default
1093
+ split: test
1094
+ revision: 7e90230a92c190f1bf69ae9002b8cea547a64cca
1095
+ metrics:
1096
+ - type: cos_sim_pearson
1097
+ value: 76.58680032464127
1098
+ - type: cos_sim_spearman
1099
+ value: 78.03760988363273
1100
+ - type: euclidean_pearson
1101
+ value: 68.23192805876019
1102
+ - type: euclidean_spearman
1103
+ value: 69.21753515532978
1104
+ - type: manhattan_pearson
1105
+ value: 68.07876685109447
1106
+ - type: manhattan_spearman
1107
+ value: 69.08026107263751
1108
+ - task:
1109
+ type: STS
1110
+ dataset:
1111
+ type: mteb/sts14-sts
1112
+ name: MTEB STS14
1113
+ config: default
1114
+ split: test
1115
+ revision: 6031580fec1f6af667f0bd2da0a551cf4f0b2375
1116
+ metrics:
1117
+ - type: cos_sim_pearson
1118
+ value: 78.72357139489792
1119
+ - type: cos_sim_spearman
1120
+ value: 74.53681843472086
1121
+ - type: euclidean_pearson
1122
+ value: 66.73161230236408
1123
+ - type: euclidean_spearman
1124
+ value: 63.81392957525887
1125
+ - type: manhattan_pearson
1126
+ value: 66.33322201893088
1127
+ - type: manhattan_spearman
1128
+ value: 63.55218357111819
1129
+ - task:
1130
+ type: STS
1131
+ dataset:
1132
+ type: mteb/sts15-sts
1133
+ name: MTEB STS15
1134
+ config: default
1135
+ split: test
1136
+ revision: ae752c7c21bf194d8b67fd573edf7ae58183cbe3
1137
+ metrics:
1138
+ - type: cos_sim_pearson
1139
+ value: 82.62456549757793
1140
+ - type: cos_sim_spearman
1141
+ value: 83.89301877076606
1142
+ - type: euclidean_pearson
1143
+ value: 58.128415035981554
1144
+ - type: euclidean_spearman
1145
+ value: 58.47993973876889
1146
+ - type: manhattan_pearson
1147
+ value: 58.37634990795807
1148
+ - type: manhattan_spearman
1149
+ value: 58.89541748905865
1150
+ - task:
1151
+ type: STS
1152
+ dataset:
1153
+ type: mteb/sts16-sts
1154
+ name: MTEB STS16
1155
+ config: default
1156
+ split: test
1157
+ revision: 4d8694f8f0e0100860b497b999b3dbed754a0513
1158
+ metrics:
1159
+ - type: cos_sim_pearson
1160
+ value: 76.79731685895317
1161
+ - type: cos_sim_spearman
1162
+ value: 79.04240201103201
1163
+ - type: euclidean_pearson
1164
+ value: 64.26869512572189
1165
+ - type: euclidean_spearman
1166
+ value: 65.09728500847595
1167
+ - type: manhattan_pearson
1168
+ value: 64.2772185991136
1169
+ - type: manhattan_spearman
1170
+ value: 65.18852760227209
1171
+ - task:
1172
+ type: STS
1173
+ dataset:
1174
+ type: mteb/sts17-crosslingual-sts
1175
+ name: MTEB STS17 (en-en)
1176
+ config: en-en
1177
+ split: test
1178
+ revision: af5e6fb845001ecf41f4c1e033ce921939a2a68d
1179
+ metrics:
1180
+ - type: cos_sim_pearson
1181
+ value: 86.30962737077412
1182
+ - type: cos_sim_spearman
1183
+ value: 86.77386963770132
1184
+ - type: euclidean_pearson
1185
+ value: 70.0534100015362
1186
+ - type: euclidean_spearman
1187
+ value: 68.17903243639661
1188
+ - type: manhattan_pearson
1189
+ value: 70.03048392176451
1190
+ - type: manhattan_spearman
1191
+ value: 68.19594588464386
1192
+ - task:
1193
+ type: STS
1194
+ dataset:
1195
+ type: mteb/sts22-crosslingual-sts
1196
+ name: MTEB STS22 (en)
1197
+ config: en
1198
+ split: test
1199
+ revision: 6d1ba47164174a496b7fa5d3569dae26a6813b80
1200
+ metrics:
1201
+ - type: cos_sim_pearson
1202
+ value: 64.77791754851359
1203
+ - type: cos_sim_spearman
1204
+ value: 64.28210927783513
1205
+ - type: euclidean_pearson
1206
+ value: 36.337603238543956
1207
+ - type: euclidean_spearman
1208
+ value: 52.70617012481411
1209
+ - type: manhattan_pearson
1210
+ value: 35.49141141164909
1211
+ - type: manhattan_spearman
1212
+ value: 52.084744319382835
1213
+ - task:
1214
+ type: STS
1215
+ dataset:
1216
+ type: mteb/stsbenchmark-sts
1217
+ name: MTEB STSBenchmark
1218
+ config: default
1219
+ split: test
1220
+ revision: b0fddb56ed78048fa8b90373c8a3cfc37b684831
1221
+ metrics:
1222
+ - type: cos_sim_pearson
1223
+ value: 79.741579322503
1224
+ - type: cos_sim_spearman
1225
+ value: 78.83687709048151
1226
+ - type: euclidean_pearson
1227
+ value: 66.59151974274772
1228
+ - type: euclidean_spearman
1229
+ value: 63.76907648545863
1230
+ - type: manhattan_pearson
1231
+ value: 66.91555116739791
1232
+ - type: manhattan_spearman
1233
+ value: 64.2024945118848
1234
+ - task:
1235
+ type: Reranking
1236
+ dataset:
1237
+ type: mteb/scidocs-reranking
1238
+ name: MTEB SciDocsRR
1239
+ config: default
1240
+ split: test
1241
+ revision: d3c5e1fc0b855ab6097bf1cda04dd73947d7caab
1242
+ metrics:
1243
+ - type: map
1244
+ value: 74.31125049985503
1245
+ - type: mrr
1246
+ value: 91.5911222038673
1247
+ - task:
1248
+ type: Retrieval
1249
+ dataset:
1250
+ type: scifact
1251
+ name: MTEB SciFact
1252
+ config: default
1253
+ split: test
1254
+ revision: None
1255
+ metrics:
1256
+ - type: map_at_1
1257
+ value: 39.983000000000004
1258
+ - type: map_at_10
1259
+ value: 48.79
1260
+ - type: map_at_100
1261
+ value: 49.419999999999995
1262
+ - type: map_at_1000
1263
+ value: 49.495
1264
+ - type: map_at_3
1265
+ value: 46.394000000000005
1266
+ - type: map_at_5
1267
+ value: 47.772999999999996
1268
+ - type: mrr_at_1
1269
+ value: 42.667
1270
+ - type: mrr_at_10
1271
+ value: 51.088
1272
+ - type: mrr_at_100
1273
+ value: 51.498999999999995
1274
+ - type: mrr_at_1000
1275
+ value: 51.564
1276
+ - type: mrr_at_3
1277
+ value: 49.111
1278
+ - type: mrr_at_5
1279
+ value: 50.278
1280
+ - type: ndcg_at_1
1281
+ value: 42.667
1282
+ - type: ndcg_at_10
1283
+ value: 53.586999999999996
1284
+ - type: ndcg_at_100
1285
+ value: 56.519
1286
+ - type: ndcg_at_1000
1287
+ value: 58.479000000000006
1288
+ - type: ndcg_at_3
1289
+ value: 49.053000000000004
1290
+ - type: ndcg_at_5
1291
+ value: 51.209
1292
+ - type: precision_at_1
1293
+ value: 42.667
1294
+ - type: precision_at_10
1295
+ value: 7.3999999999999995
1296
+ - type: precision_at_100
1297
+ value: 0.9129999999999999
1298
+ - type: precision_at_1000
1299
+ value: 0.108
1300
+ - type: precision_at_3
1301
+ value: 19.444
1302
+ - type: precision_at_5
1303
+ value: 13.067
1304
+ - type: recall_at_1
1305
+ value: 39.983000000000004
1306
+ - type: recall_at_10
1307
+ value: 66.333
1308
+ - type: recall_at_100
1309
+ value: 80.256
1310
+ - type: recall_at_1000
1311
+ value: 95.667
1312
+ - type: recall_at_3
1313
+ value: 53.449999999999996
1314
+ - type: recall_at_5
1315
+ value: 58.989000000000004
1316
+ - task:
1317
+ type: PairClassification
1318
+ dataset:
1319
+ type: mteb/sprintduplicatequestions-pairclassification
1320
+ name: MTEB SprintDuplicateQuestions
1321
+ config: default
1322
+ split: test
1323
+ revision: d66bd1f72af766a5cc4b0ca5e00c162f89e8cc46
1324
+ metrics:
1325
+ - type: cos_sim_accuracy
1326
+ value: 99.6930693069307
1327
+ - type: cos_sim_ap
1328
+ value: 90.94265768188356
1329
+ - type: cos_sim_f1
1330
+ value: 84.15792103948026
1331
+ - type: cos_sim_precision
1332
+ value: 84.11588411588411
1333
+ - type: cos_sim_recall
1334
+ value: 84.2
1335
+ - type: dot_accuracy
1336
+ value: 99.12178217821783
1337
+ - type: dot_ap
1338
+ value: 42.77306613711772
1339
+ - type: dot_f1
1340
+ value: 44.23963133640553
1341
+ - type: dot_precision
1342
+ value: 38.0677721701514
1343
+ - type: dot_recall
1344
+ value: 52.800000000000004
1345
+ - type: euclidean_accuracy
1346
+ value: 99.55049504950495
1347
+ - type: euclidean_ap
1348
+ value: 78.83886818298362
1349
+ - type: euclidean_f1
1350
+ value: 74.54645409565696
1351
+ - type: euclidean_precision
1352
+ value: 82.78388278388277
1353
+ - type: euclidean_recall
1354
+ value: 67.80000000000001
1355
+ - type: manhattan_accuracy
1356
+ value: 99.54257425742574
1357
+ - type: manhattan_ap
1358
+ value: 77.98046807031727
1359
+ - type: manhattan_f1
1360
+ value: 74.18822234452395
1361
+ - type: manhattan_precision
1362
+ value: 82.4969400244798
1363
+ - type: manhattan_recall
1364
+ value: 67.4
1365
+ - type: max_accuracy
1366
+ value: 99.6930693069307
1367
+ - type: max_ap
1368
+ value: 90.94265768188356
1369
+ - type: max_f1
1370
+ value: 84.15792103948026
1371
+ - task:
1372
+ type: Clustering
1373
+ dataset:
1374
+ type: mteb/stackexchange-clustering
1375
+ name: MTEB StackExchangeClustering
1376
+ config: default
1377
+ split: test
1378
+ revision: 6cbc1f7b2bc0622f2e39d2c77fa502909748c259
1379
+ metrics:
1380
+ - type: v_measure
1381
+ value: 47.81120799399627
1382
+ - task:
1383
+ type: Clustering
1384
+ dataset:
1385
+ type: mteb/stackexchange-clustering-p2p
1386
+ name: MTEB StackExchangeClusteringP2P
1387
+ config: default
1388
+ split: test
1389
+ revision: 815ca46b2622cec33ccafc3735d572c266efdb44
1390
+ metrics:
1391
+ - type: v_measure
1392
+ value: 29.82642033698617
1393
+ - task:
1394
+ type: Reranking
1395
+ dataset:
1396
+ type: mteb/stackoverflowdupquestions-reranking
1397
+ name: MTEB StackOverflowDupQuestions
1398
+ config: default
1399
+ split: test
1400
+ revision: e185fbe320c72810689fc5848eb6114e1ef5ec69
1401
+ metrics:
1402
+ - type: map
1403
+ value: 47.861728758923675
1404
+ - type: mrr
1405
+ value: 48.53185213479331
1406
+ - task:
1407
+ type: Summarization
1408
+ dataset:
1409
+ type: mteb/summeval
1410
+ name: MTEB SummEval
1411
+ config: default
1412
+ split: test
1413
+ revision: cda12ad7615edc362dbf25a00fdd61d3b1eaf93c
1414
+ metrics:
1415
+ - type: cos_sim_pearson
1416
+ value: 30.09237795780992
1417
+ - type: cos_sim_spearman
1418
+ value: 28.95547545518808
1419
+ - type: dot_pearson
1420
+ value: 19.99986205111785
1421
+ - type: dot_spearman
1422
+ value: 21.34033389331779
1423
+ - task:
1424
+ type: Retrieval
1425
+ dataset:
1426
+ type: trec-covid
1427
+ name: MTEB TRECCOVID
1428
+ config: default
1429
+ split: test
1430
+ revision: None
1431
+ metrics:
1432
+ - type: map_at_1
1433
+ value: 0.169
1434
+ - type: map_at_10
1435
+ value: 1.077
1436
+ - type: map_at_100
1437
+ value: 4.9750000000000005
1438
+ - type: map_at_1000
1439
+ value: 11.802
1440
+ - type: map_at_3
1441
+ value: 0.48700000000000004
1442
+ - type: map_at_5
1443
+ value: 0.679
1444
+ - type: mrr_at_1
1445
+ value: 62.0
1446
+ - type: mrr_at_10
1447
+ value: 76.25
1448
+ - type: mrr_at_100
1449
+ value: 76.337
1450
+ - type: mrr_at_1000
1451
+ value: 76.337
1452
+ - type: mrr_at_3
1453
+ value: 74.333
1454
+ - type: mrr_at_5
1455
+ value: 75.333
1456
+ - type: ndcg_at_1
1457
+ value: 56.00000000000001
1458
+ - type: ndcg_at_10
1459
+ value: 50.631
1460
+ - type: ndcg_at_100
1461
+ value: 36.39
1462
+ - type: ndcg_at_1000
1463
+ value: 32.879000000000005
1464
+ - type: ndcg_at_3
1465
+ value: 59.961
1466
+ - type: ndcg_at_5
1467
+ value: 55.913999999999994
1468
+ - type: precision_at_1
1469
+ value: 62.0
1470
+ - type: precision_at_10
1471
+ value: 53.0
1472
+ - type: precision_at_100
1473
+ value: 37.2
1474
+ - type: precision_at_1000
1475
+ value: 14.804
1476
+ - type: precision_at_3
1477
+ value: 67.333
1478
+ - type: precision_at_5
1479
+ value: 60.4
1480
+ - type: recall_at_1
1481
+ value: 0.169
1482
+ - type: recall_at_10
1483
+ value: 1.324
1484
+ - type: recall_at_100
1485
+ value: 8.352
1486
+ - type: recall_at_1000
1487
+ value: 31.041999999999998
1488
+ - type: recall_at_3
1489
+ value: 0.532
1490
+ - type: recall_at_5
1491
+ value: 0.777
1492
+ - task:
1493
+ type: Retrieval
1494
+ dataset:
1495
+ type: webis-touche2020
1496
+ name: MTEB Touche2020
1497
+ config: default
1498
+ split: test
1499
+ revision: None
1500
+ metrics:
1501
+ - type: map_at_1
1502
+ value: 2.018
1503
+ - type: map_at_10
1504
+ value: 8.036
1505
+ - type: map_at_100
1506
+ value: 12.814
1507
+ - type: map_at_1000
1508
+ value: 14.204
1509
+ - type: map_at_3
1510
+ value: 3.9759999999999995
1511
+ - type: map_at_5
1512
+ value: 5.585
1513
+ - type: mrr_at_1
1514
+ value: 24.490000000000002
1515
+ - type: mrr_at_10
1516
+ value: 38.903
1517
+ - type: mrr_at_100
1518
+ value: 39.893
1519
+ - type: mrr_at_1000
1520
+ value: 39.895
1521
+ - type: mrr_at_3
1522
+ value: 35.034
1523
+ - type: mrr_at_5
1524
+ value: 37.789
1525
+ - type: ndcg_at_1
1526
+ value: 21.429000000000002
1527
+ - type: ndcg_at_10
1528
+ value: 20.082
1529
+ - type: ndcg_at_100
1530
+ value: 30.299
1531
+ - type: ndcg_at_1000
1532
+ value: 42.323
1533
+ - type: ndcg_at_3
1534
+ value: 19.826
1535
+ - type: ndcg_at_5
1536
+ value: 19.861
1537
+ - type: precision_at_1
1538
+ value: 24.490000000000002
1539
+ - type: precision_at_10
1540
+ value: 18.776
1541
+ - type: precision_at_100
1542
+ value: 6.551
1543
+ - type: precision_at_1000
1544
+ value: 1.455
1545
+ - type: precision_at_3
1546
+ value: 21.088
1547
+ - type: precision_at_5
1548
+ value: 21.633
1549
+ - type: recall_at_1
1550
+ value: 2.018
1551
+ - type: recall_at_10
1552
+ value: 14.094999999999999
1553
+ - type: recall_at_100
1554
+ value: 40.482
1555
+ - type: recall_at_1000
1556
+ value: 78.214
1557
+ - type: recall_at_3
1558
+ value: 4.884
1559
+ - type: recall_at_5
1560
+ value: 8.203000000000001
1561
+ - task:
1562
+ type: Classification
1563
+ dataset:
1564
+ type: mteb/toxic_conversations_50k
1565
+ name: MTEB ToxicConversationsClassification
1566
+ config: default
1567
+ split: test
1568
+ revision: d7c0de2777da35d6aae2200a62c6e0e5af397c4c
1569
+ metrics:
1570
+ - type: accuracy
1571
+ value: 59.69140000000001
1572
+ - type: ap
1573
+ value: 10.299275820958274
1574
+ - type: f1
1575
+ value: 45.697311005218154
1576
+ - task:
1577
+ type: Classification
1578
+ dataset:
1579
+ type: mteb/tweet_sentiment_extraction
1580
+ name: MTEB TweetSentimentExtractionClassification
1581
+ config: default
1582
+ split: test
1583
+ revision: d604517c81ca91fe16a244d1248fc021f9ecee7a
1584
+ metrics:
1585
+ - type: accuracy
1586
+ value: 53.542727787209955
1587
+ - type: f1
1588
+ value: 53.59495510018717
1589
+ - task:
1590
+ type: Clustering
1591
+ dataset:
1592
+ type: mteb/twentynewsgroups-clustering
1593
+ name: MTEB TwentyNewsgroupsClustering
1594
+ config: default
1595
+ split: test
1596
+ revision: 6125ec4e24fa026cec8a478383ee943acfbd5449
1597
+ metrics:
1598
+ - type: v_measure
1599
+ value: 32.405659957745534
1600
+ - task:
1601
+ type: PairClassification
1602
+ dataset:
1603
+ type: mteb/twittersemeval2015-pairclassification
1604
+ name: MTEB TwitterSemEval2015
1605
+ config: default
1606
+ split: test
1607
+ revision: 70970daeab8776df92f5ea462b6173c0b46fd2d1
1608
+ metrics:
1609
+ - type: cos_sim_accuracy
1610
+ value: 82.34487691482386
1611
+ - type: cos_sim_ap
1612
+ value: 61.4880638625752
1613
+ - type: cos_sim_f1
1614
+ value: 59.350775193798455
1615
+ - type: cos_sim_precision
1616
+ value: 54.858934169278996
1617
+ - type: cos_sim_recall
1618
+ value: 64.64379947229551
1619
+ - type: dot_accuracy
1620
+ value: 77.68373368301842
1621
+ - type: dot_ap
1622
+ value: 36.846940578266626
1623
+ - type: dot_f1
1624
+ value: 42.67407473787974
1625
+ - type: dot_precision
1626
+ value: 32.311032704573215
1627
+ - type: dot_recall
1628
+ value: 62.82321899736147
1629
+ - type: euclidean_accuracy
1630
+ value: 80.40770101925256
1631
+ - type: euclidean_ap
1632
+ value: 53.51906185864526
1633
+ - type: euclidean_f1
1634
+ value: 53.24030024315466
1635
+ - type: euclidean_precision
1636
+ value: 44.41700476274475
1637
+ - type: euclidean_recall
1638
+ value: 66.43799472295514
1639
+ - type: manhattan_accuracy
1640
+ value: 80.31829290099542
1641
+ - type: manhattan_ap
1642
+ value: 53.67183195163967
1643
+ - type: manhattan_f1
1644
+ value: 53.28358208955224
1645
+ - type: manhattan_precision
1646
+ value: 44.70483005366726
1647
+ - type: manhattan_recall
1648
+ value: 65.93667546174143
1649
+ - type: max_accuracy
1650
+ value: 82.34487691482386
1651
+ - type: max_ap
1652
+ value: 61.4880638625752
1653
+ - type: max_f1
1654
+ value: 59.350775193798455
1655
+ - task:
1656
+ type: PairClassification
1657
+ dataset:
1658
+ type: mteb/twitterurlcorpus-pairclassification
1659
+ name: MTEB TwitterURLCorpus
1660
+ config: default
1661
+ split: test
1662
+ revision: 8b6510b0b1fa4e4c4f879467980e9be563ec1cdf
1663
+ metrics:
1664
+ - type: cos_sim_accuracy
1665
+ value: 87.71684713005007
1666
+ - type: cos_sim_ap
1667
+ value: 82.85441942604702
1668
+ - type: cos_sim_f1
1669
+ value: 75.69942543843179
1670
+ - type: cos_sim_precision
1671
+ value: 73.88754490140019
1672
+ - type: cos_sim_recall
1673
+ value: 77.60240221743148
1674
+ - type: dot_accuracy
1675
+ value: 82.23696976753212
1676
+ - type: dot_ap
1677
+ value: 68.47562727147806
1678
+ - type: dot_f1
1679
+ value: 64.99698249849123
1680
+ - type: dot_precision
1681
+ value: 57.566219265946074
1682
+ - type: dot_recall
1683
+ value: 74.63042808746535
1684
+ - type: euclidean_accuracy
1685
+ value: 81.52481856638336
1686
+ - type: euclidean_ap
1687
+ value: 65.96678666430529
1688
+ - type: euclidean_f1
1689
+ value: 59.14671467146715
1690
+ - type: euclidean_precision
1691
+ value: 55.54879285859201
1692
+ - type: euclidean_recall
1693
+ value: 63.24299353249153
1694
+ - type: manhattan_accuracy
1695
+ value: 81.56750882912253
1696
+ - type: manhattan_ap
1697
+ value: 66.07646774834106
1698
+ - type: manhattan_f1
1699
+ value: 59.161485036907756
1700
+ - type: manhattan_precision
1701
+ value: 56.05319368841728
1702
+ - type: manhattan_recall
1703
+ value: 62.634739759778256
1704
+ - type: max_accuracy
1705
+ value: 87.71684713005007
1706
+ - type: max_ap
1707
+ value: 82.85441942604702
1708
+ - type: max_f1
1709
+ value: 75.69942543843179
1710
+ ---
1711
  ---
1712
 
1713
  <br><br>