Pristinenlp commited on
Commit
1d2cb86
1 Parent(s): 4433fab
1_Pooling/config.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "word_embedding_dimension": 1024,
3
+ "pooling_mode_cls_token": true,
4
+ "pooling_mode_mean_tokens": false,
5
+ "pooling_mode_max_tokens": false,
6
+ "pooling_mode_mean_sqrt_len_tokens": false
7
+ }
README.md CHANGED
@@ -1,3 +1,1082 @@
1
  ---
2
- license: mit
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ tags:
3
+ - mteb
4
+ model-index:
5
+ - name: alime-embedding-large-zh
6
+ results:
7
+ - task:
8
+ type: STS
9
+ dataset:
10
+ type: C-MTEB/AFQMC
11
+ name: MTEB AFQMC
12
+ config: default
13
+ split: validation
14
+ revision: None
15
+ metrics:
16
+ - type: cos_sim_pearson
17
+ value: 49.6479989785073
18
+ - type: cos_sim_spearman
19
+ value: 54.733173049795425
20
+ - type: euclidean_pearson
21
+ value: 53.06330391299694
22
+ - type: euclidean_spearman
23
+ value: 54.73321325021156
24
+ - type: manhattan_pearson
25
+ value: 53.0477915350307
26
+ - type: manhattan_spearman
27
+ value: 54.728508847750845
28
+ - task:
29
+ type: STS
30
+ dataset:
31
+ type: C-MTEB/ATEC
32
+ name: MTEB ATEC
33
+ config: default
34
+ split: test
35
+ revision: None
36
+ metrics:
37
+ - type: cos_sim_pearson
38
+ value: 48.658812679136325
39
+ - type: cos_sim_spearman
40
+ value: 55.125070901329146
41
+ - type: euclidean_pearson
42
+ value: 55.73373519622172
43
+ - type: euclidean_spearman
44
+ value: 55.12506864911728
45
+ - type: manhattan_pearson
46
+ value: 55.71155132206361
47
+ - type: manhattan_spearman
48
+ value: 55.121598723227905
49
+ - task:
50
+ type: Classification
51
+ dataset:
52
+ type: mteb/amazon_reviews_multi
53
+ name: MTEB AmazonReviewsClassification (zh)
54
+ config: zh
55
+ split: test
56
+ revision: 1399c76144fd37290681b995c656ef9b2e06e26d
57
+ metrics:
58
+ - type: accuracy
59
+ value: 46.95
60
+ - type: f1
61
+ value: 45.34383964066362
62
+ - task:
63
+ type: STS
64
+ dataset:
65
+ type: C-MTEB/BQ
66
+ name: MTEB BQ
67
+ config: default
68
+ split: test
69
+ revision: None
70
+ metrics:
71
+ - type: cos_sim_pearson
72
+ value: 62.92731050834033
73
+ - type: cos_sim_spearman
74
+ value: 64.8881453551134
75
+ - type: euclidean_pearson
76
+ value: 63.31447523186855
77
+ - type: euclidean_spearman
78
+ value: 64.88814189042776
79
+ - type: manhattan_pearson
80
+ value: 63.222442228527996
81
+ - type: manhattan_spearman
82
+ value: 64.79818263591122
83
+ - task:
84
+ type: Clustering
85
+ dataset:
86
+ type: C-MTEB/CLSClusteringP2P
87
+ name: MTEB CLSClusteringP2P
88
+ config: default
89
+ split: test
90
+ revision: None
91
+ metrics:
92
+ - type: v_measure
93
+ value: 42.518811360488925
94
+ - task:
95
+ type: Clustering
96
+ dataset:
97
+ type: C-MTEB/CLSClusteringS2S
98
+ name: MTEB CLSClusteringS2S
99
+ config: default
100
+ split: test
101
+ revision: None
102
+ metrics:
103
+ - type: v_measure
104
+ value: 39.72890397315954
105
+ - task:
106
+ type: Reranking
107
+ dataset:
108
+ type: C-MTEB/CMedQAv1-reranking
109
+ name: MTEB CMedQAv1
110
+ config: default
111
+ split: test
112
+ revision: None
113
+ metrics:
114
+ - type: map
115
+ value: 86.51852576014969
116
+ - type: mrr
117
+ value: 89.02047619047619
118
+ - task:
119
+ type: Reranking
120
+ dataset:
121
+ type: C-MTEB/CMedQAv2-reranking
122
+ name: MTEB CMedQAv2
123
+ config: default
124
+ split: test
125
+ revision: None
126
+ metrics:
127
+ - type: map
128
+ value: 87.11415162833914
129
+ - type: mrr
130
+ value: 89.6338492063492
131
+ - task:
132
+ type: Retrieval
133
+ dataset:
134
+ type: C-MTEB/CmedqaRetrieval
135
+ name: MTEB CmedqaRetrieval
136
+ config: default
137
+ split: dev
138
+ revision: None
139
+ metrics:
140
+ - type: map_at_1
141
+ value: 24.883
142
+ - type: map_at_10
143
+ value: 37.246
144
+ - type: map_at_100
145
+ value: 39.11
146
+ - type: map_at_1000
147
+ value: 39.222
148
+ - type: map_at_3
149
+ value: 32.956
150
+ - type: map_at_5
151
+ value: 35.411
152
+ - type: mrr_at_1
153
+ value: 37.834
154
+ - type: mrr_at_10
155
+ value: 46.031
156
+ - type: mrr_at_100
157
+ value: 47.033
158
+ - type: mrr_at_1000
159
+ value: 47.077000000000005
160
+ - type: mrr_at_3
161
+ value: 43.415
162
+ - type: mrr_at_5
163
+ value: 44.938
164
+ - type: ndcg_at_1
165
+ value: 37.834
166
+ - type: ndcg_at_10
167
+ value: 43.928
168
+ - type: ndcg_at_100
169
+ value: 51.312999999999995
170
+ - type: ndcg_at_1000
171
+ value: 53.23
172
+ - type: ndcg_at_3
173
+ value: 38.397
174
+ - type: ndcg_at_5
175
+ value: 40.848
176
+ - type: precision_at_1
177
+ value: 37.834
178
+ - type: precision_at_10
179
+ value: 9.782
180
+ - type: precision_at_100
181
+ value: 1.583
182
+ - type: precision_at_1000
183
+ value: 0.183
184
+ - type: precision_at_3
185
+ value: 21.664
186
+ - type: precision_at_5
187
+ value: 15.934000000000001
188
+ - type: recall_at_1
189
+ value: 24.883
190
+ - type: recall_at_10
191
+ value: 54.911
192
+ - type: recall_at_100
193
+ value: 85.419
194
+ - type: recall_at_1000
195
+ value: 98.16
196
+ - type: recall_at_3
197
+ value: 38.416
198
+ - type: recall_at_5
199
+ value: 45.778
200
+ - task:
201
+ type: PairClassification
202
+ dataset:
203
+ type: C-MTEB/CMNLI
204
+ name: MTEB Cmnli
205
+ config: default
206
+ split: validation
207
+ revision: None
208
+ metrics:
209
+ - type: cos_sim_accuracy
210
+ value: 82.5616355983163
211
+ - type: cos_sim_ap
212
+ value: 89.3612977679186
213
+ - type: cos_sim_f1
214
+ value: 83.93428161870108
215
+ - type: cos_sim_precision
216
+ value: 79.42404006677796
217
+ - type: cos_sim_recall
218
+ value: 88.98760813654431
219
+ - type: dot_accuracy
220
+ value: 82.5616355983163
221
+ - type: dot_ap
222
+ value: 89.38168095374776
223
+ - type: dot_f1
224
+ value: 83.93428161870108
225
+ - type: dot_precision
226
+ value: 79.42404006677796
227
+ - type: dot_recall
228
+ value: 88.98760813654431
229
+ - type: euclidean_accuracy
230
+ value: 82.5616355983163
231
+ - type: euclidean_ap
232
+ value: 89.36129603693611
233
+ - type: euclidean_f1
234
+ value: 83.93428161870108
235
+ - type: euclidean_precision
236
+ value: 79.42404006677796
237
+ - type: euclidean_recall
238
+ value: 88.98760813654431
239
+ - type: manhattan_accuracy
240
+ value: 82.42934455802767
241
+ - type: manhattan_ap
242
+ value: 89.36577661305246
243
+ - type: manhattan_f1
244
+ value: 83.94765539803707
245
+ - type: manhattan_precision
246
+ value: 78.66339668914776
247
+ - type: manhattan_recall
248
+ value: 89.99298573766659
249
+ - type: max_accuracy
250
+ value: 82.5616355983163
251
+ - type: max_ap
252
+ value: 89.38168095374776
253
+ - type: max_f1
254
+ value: 83.94765539803707
255
+ - task:
256
+ type: Retrieval
257
+ dataset:
258
+ type: C-MTEB/CovidRetrieval
259
+ name: MTEB CovidRetrieval
260
+ config: default
261
+ split: dev
262
+ revision: None
263
+ metrics:
264
+ - type: map_at_1
265
+ value: 77.608
266
+ - type: map_at_10
267
+ value: 85.1
268
+ - type: map_at_100
269
+ value: 85.215
270
+ - type: map_at_1000
271
+ value: 85.217
272
+ - type: map_at_3
273
+ value: 83.97
274
+ - type: map_at_5
275
+ value: 84.638
276
+ - type: mrr_at_1
277
+ value: 77.97699999999999
278
+ - type: mrr_at_10
279
+ value: 85.173
280
+ - type: mrr_at_100
281
+ value: 85.28
282
+ - type: mrr_at_1000
283
+ value: 85.282
284
+ - type: mrr_at_3
285
+ value: 84.089
286
+ - type: mrr_at_5
287
+ value: 84.726
288
+ - type: ndcg_at_1
289
+ value: 77.871
290
+ - type: ndcg_at_10
291
+ value: 88.141
292
+ - type: ndcg_at_100
293
+ value: 88.612
294
+ - type: ndcg_at_1000
295
+ value: 88.68
296
+ - type: ndcg_at_3
297
+ value: 85.9
298
+ - type: ndcg_at_5
299
+ value: 87.06
300
+ - type: precision_at_1
301
+ value: 77.871
302
+ - type: precision_at_10
303
+ value: 9.841999999999999
304
+ - type: precision_at_100
305
+ value: 1.005
306
+ - type: precision_at_1000
307
+ value: 0.101
308
+ - type: precision_at_3
309
+ value: 30.698999999999998
310
+ - type: precision_at_5
311
+ value: 19.009
312
+ - type: recall_at_1
313
+ value: 77.608
314
+ - type: recall_at_10
315
+ value: 97.418
316
+ - type: recall_at_100
317
+ value: 99.473
318
+ - type: recall_at_1000
319
+ value: 100.0
320
+ - type: recall_at_3
321
+ value: 91.307
322
+ - type: recall_at_5
323
+ value: 94.125
324
+ - task:
325
+ type: Retrieval
326
+ dataset:
327
+ type: C-MTEB/DuRetrieval
328
+ name: MTEB DuRetrieval
329
+ config: default
330
+ split: dev
331
+ revision: None
332
+ metrics:
333
+ - type: map_at_1
334
+ value: 26.104
335
+ - type: map_at_10
336
+ value: 78.62
337
+ - type: map_at_100
338
+ value: 81.417
339
+ - type: map_at_1000
340
+ value: 81.46600000000001
341
+ - type: map_at_3
342
+ value: 55.077
343
+ - type: map_at_5
344
+ value: 69.18900000000001
345
+ - type: mrr_at_1
346
+ value: 90.55
347
+ - type: mrr_at_10
348
+ value: 93.42200000000001
349
+ - type: mrr_at_100
350
+ value: 93.46900000000001
351
+ - type: mrr_at_1000
352
+ value: 93.472
353
+ - type: mrr_at_3
354
+ value: 93.108
355
+ - type: mrr_at_5
356
+ value: 93.318
357
+ - type: ndcg_at_1
358
+ value: 90.55
359
+ - type: ndcg_at_10
360
+ value: 86.227
361
+ - type: ndcg_at_100
362
+ value: 89.201
363
+ - type: ndcg_at_1000
364
+ value: 89.655
365
+ - type: ndcg_at_3
366
+ value: 85.89099999999999
367
+ - type: ndcg_at_5
368
+ value: 84.443
369
+ - type: precision_at_1
370
+ value: 90.55
371
+ - type: precision_at_10
372
+ value: 40.915
373
+ - type: precision_at_100
374
+ value: 4.749
375
+ - type: precision_at_1000
376
+ value: 0.486
377
+ - type: precision_at_3
378
+ value: 76.9
379
+ - type: precision_at_5
380
+ value: 64.56
381
+ - type: recall_at_1
382
+ value: 26.104
383
+ - type: recall_at_10
384
+ value: 86.924
385
+ - type: recall_at_100
386
+ value: 96.52
387
+ - type: recall_at_1000
388
+ value: 98.83800000000001
389
+ - type: recall_at_3
390
+ value: 57.196999999999996
391
+ - type: recall_at_5
392
+ value: 73.595
393
+ - task:
394
+ type: Retrieval
395
+ dataset:
396
+ type: C-MTEB/EcomRetrieval
397
+ name: MTEB EcomRetrieval
398
+ config: default
399
+ split: dev
400
+ revision: None
401
+ metrics:
402
+ - type: map_at_1
403
+ value: 51.9
404
+ - type: map_at_10
405
+ value: 62.446
406
+ - type: map_at_100
407
+ value: 62.922
408
+ - type: map_at_1000
409
+ value: 62.934999999999995
410
+ - type: map_at_3
411
+ value: 59.933
412
+ - type: map_at_5
413
+ value: 61.548
414
+ - type: mrr_at_1
415
+ value: 51.9
416
+ - type: mrr_at_10
417
+ value: 62.446
418
+ - type: mrr_at_100
419
+ value: 62.922
420
+ - type: mrr_at_1000
421
+ value: 62.934999999999995
422
+ - type: mrr_at_3
423
+ value: 59.933
424
+ - type: mrr_at_5
425
+ value: 61.548
426
+ - type: ndcg_at_1
427
+ value: 51.9
428
+ - type: ndcg_at_10
429
+ value: 67.561
430
+ - type: ndcg_at_100
431
+ value: 69.87400000000001
432
+ - type: ndcg_at_1000
433
+ value: 70.19800000000001
434
+ - type: ndcg_at_3
435
+ value: 62.474
436
+ - type: ndcg_at_5
437
+ value: 65.391
438
+ - type: precision_at_1
439
+ value: 51.9
440
+ - type: precision_at_10
441
+ value: 8.36
442
+ - type: precision_at_100
443
+ value: 0.9440000000000001
444
+ - type: precision_at_1000
445
+ value: 0.097
446
+ - type: precision_at_3
447
+ value: 23.267
448
+ - type: precision_at_5
449
+ value: 15.379999999999999
450
+ - type: recall_at_1
451
+ value: 51.9
452
+ - type: recall_at_10
453
+ value: 83.6
454
+ - type: recall_at_100
455
+ value: 94.39999999999999
456
+ - type: recall_at_1000
457
+ value: 96.89999999999999
458
+ - type: recall_at_3
459
+ value: 69.8
460
+ - type: recall_at_5
461
+ value: 76.9
462
+ - task:
463
+ type: Classification
464
+ dataset:
465
+ type: C-MTEB/IFlyTek-classification
466
+ name: MTEB IFlyTek
467
+ config: default
468
+ split: validation
469
+ revision: None
470
+ metrics:
471
+ - type: accuracy
472
+ value: 49.672951135051946
473
+ - type: f1
474
+ value: 38.246634605142084
475
+ - task:
476
+ type: Classification
477
+ dataset:
478
+ type: C-MTEB/JDReview-classification
479
+ name: MTEB JDReview
480
+ config: default
481
+ split: test
482
+ revision: None
483
+ metrics:
484
+ - type: accuracy
485
+ value: 86.52908067542214
486
+ - type: ap
487
+ value: 55.415146961759135
488
+ - type: f1
489
+ value: 81.38343036361825
490
+ - task:
491
+ type: STS
492
+ dataset:
493
+ type: C-MTEB/LCQMC
494
+ name: MTEB LCQMC
495
+ config: default
496
+ split: test
497
+ revision: None
498
+ metrics:
499
+ - type: cos_sim_pearson
500
+ value: 70.15572724302896
501
+ - type: cos_sim_spearman
502
+ value: 75.11630463239744
503
+ - type: euclidean_pearson
504
+ value: 74.2927184018677
505
+ - type: euclidean_spearman
506
+ value: 75.11630463089752
507
+ - type: manhattan_pearson
508
+ value: 74.27724224882166
509
+ - type: manhattan_spearman
510
+ value: 75.10012699894408
511
+ - task:
512
+ type: Reranking
513
+ dataset:
514
+ type: C-MTEB/Mmarco-reranking
515
+ name: MTEB MMarcoReranking
516
+ config: default
517
+ split: dev
518
+ revision: None
519
+ metrics:
520
+ - type: map
521
+ value: 30.62934327678744
522
+ - type: mrr
523
+ value: 29.48730158730159
524
+ - task:
525
+ type: Retrieval
526
+ dataset:
527
+ type: C-MTEB/MMarcoRetrieval
528
+ name: MTEB MMarcoRetrieval
529
+ config: default
530
+ split: dev
531
+ revision: None
532
+ metrics:
533
+ - type: map_at_1
534
+ value: 65.33
535
+ - type: map_at_10
536
+ value: 74.524
537
+ - type: map_at_100
538
+ value: 74.851
539
+ - type: map_at_1000
540
+ value: 74.86500000000001
541
+ - type: map_at_3
542
+ value: 72.748
543
+ - type: map_at_5
544
+ value: 73.896
545
+ - type: mrr_at_1
546
+ value: 67.593
547
+ - type: mrr_at_10
548
+ value: 75.19
549
+ - type: mrr_at_100
550
+ value: 75.472
551
+ - type: mrr_at_1000
552
+ value: 75.484
553
+ - type: mrr_at_3
554
+ value: 73.634
555
+ - type: mrr_at_5
556
+ value: 74.638
557
+ - type: ndcg_at_1
558
+ value: 67.593
559
+ - type: ndcg_at_10
560
+ value: 78.254
561
+ - type: ndcg_at_100
562
+ value: 79.727
563
+ - type: ndcg_at_1000
564
+ value: 80.09100000000001
565
+ - type: ndcg_at_3
566
+ value: 74.892
567
+ - type: ndcg_at_5
568
+ value: 76.835
569
+ - type: precision_at_1
570
+ value: 67.593
571
+ - type: precision_at_10
572
+ value: 9.46
573
+ - type: precision_at_100
574
+ value: 1.02
575
+ - type: precision_at_1000
576
+ value: 0.105
577
+ - type: precision_at_3
578
+ value: 28.227999999999998
579
+ - type: precision_at_5
580
+ value: 17.965999999999998
581
+ - type: recall_at_1
582
+ value: 65.33
583
+ - type: recall_at_10
584
+ value: 89.048
585
+ - type: recall_at_100
586
+ value: 95.732
587
+ - type: recall_at_1000
588
+ value: 98.598
589
+ - type: recall_at_3
590
+ value: 80.209
591
+ - type: recall_at_5
592
+ value: 84.824
593
+ - task:
594
+ type: Classification
595
+ dataset:
596
+ type: mteb/amazon_massive_intent
597
+ name: MTEB MassiveIntentClassification (zh-CN)
598
+ config: zh-CN
599
+ split: test
600
+ revision: 31efe3c427b0bae9c22cbb560b8f15491cc6bed7
601
+ metrics:
602
+ - type: accuracy
603
+ value: 73.38937457969065
604
+ - type: f1
605
+ value: 70.87692475465195
606
+ - task:
607
+ type: Classification
608
+ dataset:
609
+ type: mteb/amazon_massive_scenario
610
+ name: MTEB MassiveScenarioClassification (zh-CN)
611
+ config: zh-CN
612
+ split: test
613
+ revision: 7d571f92784cd94a019292a1f45445077d0ef634
614
+ metrics:
615
+ - type: accuracy
616
+ value: 76.04236718224612
617
+ - type: f1
618
+ value: 75.52425703483891
619
+ - task:
620
+ type: Retrieval
621
+ dataset:
622
+ type: C-MTEB/MedicalRetrieval
623
+ name: MTEB MedicalRetrieval
624
+ config: default
625
+ split: dev
626
+ revision: None
627
+ metrics:
628
+ - type: map_at_1
629
+ value: 53.1
630
+ - type: map_at_10
631
+ value: 60.24
632
+ - type: map_at_100
633
+ value: 60.781
634
+ - type: map_at_1000
635
+ value: 60.81999999999999
636
+ - type: map_at_3
637
+ value: 58.733000000000004
638
+ - type: map_at_5
639
+ value: 59.618
640
+ - type: mrr_at_1
641
+ value: 53.0
642
+ - type: mrr_at_10
643
+ value: 60.195
644
+ - type: mrr_at_100
645
+ value: 60.736000000000004
646
+ - type: mrr_at_1000
647
+ value: 60.775
648
+ - type: mrr_at_3
649
+ value: 58.68299999999999
650
+ - type: mrr_at_5
651
+ value: 59.573
652
+ - type: ndcg_at_1
653
+ value: 53.1
654
+ - type: ndcg_at_10
655
+ value: 63.568999999999996
656
+ - type: ndcg_at_100
657
+ value: 66.401
658
+ - type: ndcg_at_1000
659
+ value: 67.597
660
+ - type: ndcg_at_3
661
+ value: 60.455000000000005
662
+ - type: ndcg_at_5
663
+ value: 62.05500000000001
664
+ - type: precision_at_1
665
+ value: 53.1
666
+ - type: precision_at_10
667
+ value: 7.3999999999999995
668
+ - type: precision_at_100
669
+ value: 0.877
670
+ - type: precision_at_1000
671
+ value: 0.097
672
+ - type: precision_at_3
673
+ value: 21.8
674
+ - type: precision_at_5
675
+ value: 13.86
676
+ - type: recall_at_1
677
+ value: 53.1
678
+ - type: recall_at_10
679
+ value: 74.0
680
+ - type: recall_at_100
681
+ value: 87.7
682
+ - type: recall_at_1000
683
+ value: 97.39999999999999
684
+ - type: recall_at_3
685
+ value: 65.4
686
+ - type: recall_at_5
687
+ value: 69.3
688
+ - task:
689
+ type: Classification
690
+ dataset:
691
+ type: C-MTEB/MultilingualSentiment-classification
692
+ name: MTEB MultilingualSentiment
693
+ config: default
694
+ split: validation
695
+ revision: None
696
+ metrics:
697
+ - type: accuracy
698
+ value: 76.40333333333332
699
+ - type: f1
700
+ value: 76.40924131087777
701
+ - task:
702
+ type: PairClassification
703
+ dataset:
704
+ type: C-MTEB/OCNLI
705
+ name: MTEB Ocnli
706
+ config: default
707
+ split: validation
708
+ revision: None
709
+ metrics:
710
+ - type: cos_sim_accuracy
711
+ value: 77.15213860314023
712
+ - type: cos_sim_ap
713
+ value: 79.30594584166899
714
+ - type: cos_sim_f1
715
+ value: 80.25889967637539
716
+ - type: cos_sim_precision
717
+ value: 71.38157894736842
718
+ - type: cos_sim_recall
719
+ value: 91.65786694825766
720
+ - type: dot_accuracy
721
+ value: 77.15213860314023
722
+ - type: dot_ap
723
+ value: 79.30594584166899
724
+ - type: dot_f1
725
+ value: 80.25889967637539
726
+ - type: dot_precision
727
+ value: 71.38157894736842
728
+ - type: dot_recall
729
+ value: 91.65786694825766
730
+ - type: euclidean_accuracy
731
+ value: 77.15213860314023
732
+ - type: euclidean_ap
733
+ value: 79.30594584166899
734
+ - type: euclidean_f1
735
+ value: 80.25889967637539
736
+ - type: euclidean_precision
737
+ value: 71.38157894736842
738
+ - type: euclidean_recall
739
+ value: 91.65786694825766
740
+ - type: manhattan_accuracy
741
+ value: 77.36870600974554
742
+ - type: manhattan_ap
743
+ value: 79.23401219102254
744
+ - type: manhattan_f1
745
+ value: 80.44901777362021
746
+ - type: manhattan_precision
747
+ value: 72.20822837951302
748
+ - type: manhattan_recall
749
+ value: 90.8130939809926
750
+ - type: max_accuracy
751
+ value: 77.36870600974554
752
+ - type: max_ap
753
+ value: 79.30594584166899
754
+ - type: max_f1
755
+ value: 80.44901777362021
756
+ - task:
757
+ type: Classification
758
+ dataset:
759
+ type: C-MTEB/OnlineShopping-classification
760
+ name: MTEB OnlineShopping
761
+ config: default
762
+ split: test
763
+ revision: None
764
+ metrics:
765
+ - type: accuracy
766
+ value: 92.6
767
+ - type: ap
768
+ value: 90.78779333103819
769
+ - type: f1
770
+ value: 92.59253441654515
771
+ - task:
772
+ type: STS
773
+ dataset:
774
+ type: C-MTEB/PAWSX
775
+ name: MTEB PAWSX
776
+ config: default
777
+ split: test
778
+ revision: None
779
+ metrics:
780
+ - type: cos_sim_pearson
781
+ value: 34.4442917065113
782
+ - type: cos_sim_spearman
783
+ value: 37.93070836936766
784
+ - type: euclidean_pearson
785
+ value: 38.35141108502335
786
+ - type: euclidean_spearman
787
+ value: 37.936378767247106
788
+ - type: manhattan_pearson
789
+ value: 38.357078125497566
790
+ - type: manhattan_spearman
791
+ value: 37.94413026678537
792
+ - task:
793
+ type: STS
794
+ dataset:
795
+ type: C-MTEB/QBQTC
796
+ name: MTEB QBQTC
797
+ config: default
798
+ split: test
799
+ revision: None
800
+ metrics:
801
+ - type: cos_sim_pearson
802
+ value: 32.84777948741198
803
+ - type: cos_sim_spearman
804
+ value: 34.212129449696285
805
+ - type: euclidean_pearson
806
+ value: 32.69161407750465
807
+ - type: euclidean_spearman
808
+ value: 34.21178008084197
809
+ - type: manhattan_pearson
810
+ value: 32.675418316752506
811
+ - type: manhattan_spearman
812
+ value: 34.178590557249
813
+ - task:
814
+ type: STS
815
+ dataset:
816
+ type: mteb/sts22-crosslingual-sts
817
+ name: MTEB STS22 (zh)
818
+ config: zh
819
+ split: test
820
+ revision: 6d1ba47164174a496b7fa5d3569dae26a6813b80
821
+ metrics:
822
+ - type: cos_sim_pearson
823
+ value: 64.65903821549742
824
+ - type: cos_sim_spearman
825
+ value: 64.54376284777354
826
+ - type: euclidean_pearson
827
+ value: 63.70022677799055
828
+ - type: euclidean_spearman
829
+ value: 64.54376284777354
830
+ - type: manhattan_pearson
831
+ value: 64.46392290759724
832
+ - type: manhattan_spearman
833
+ value: 65.2496975447815
834
+ - task:
835
+ type: STS
836
+ dataset:
837
+ type: C-MTEB/STSB
838
+ name: MTEB STSB
839
+ config: default
840
+ split: test
841
+ revision: None
842
+ metrics:
843
+ - type: cos_sim_pearson
844
+ value: 80.05773088991484
845
+ - type: cos_sim_spearman
846
+ value: 80.71550237522008
847
+ - type: euclidean_pearson
848
+ value: 80.31115977415573
849
+ - type: euclidean_spearman
850
+ value: 80.71510951779365
851
+ - type: manhattan_pearson
852
+ value: 80.25235514937249
853
+ - type: manhattan_spearman
854
+ value: 80.65958309383224
855
+ - task:
856
+ type: Reranking
857
+ dataset:
858
+ type: C-MTEB/T2Reranking
859
+ name: MTEB T2Reranking
860
+ config: default
861
+ split: dev
862
+ revision: None
863
+ metrics:
864
+ - type: map
865
+ value: 66.18255262304848
866
+ - type: mrr
867
+ value: 75.95393252087565
868
+ - task:
869
+ type: Retrieval
870
+ dataset:
871
+ type: C-MTEB/T2Retrieval
872
+ name: MTEB T2Retrieval
873
+ config: default
874
+ split: dev
875
+ revision: None
876
+ metrics:
877
+ - type: map_at_1
878
+ value: 28.651
879
+ - type: map_at_10
880
+ value: 76.281
881
+ - type: map_at_100
882
+ value: 80.018
883
+ - type: map_at_1000
884
+ value: 80.098
885
+ - type: map_at_3
886
+ value: 54.783
887
+ - type: map_at_5
888
+ value: 66.508
889
+ - type: mrr_at_1
890
+ value: 90.99199999999999
891
+ - type: mrr_at_10
892
+ value: 93.812
893
+ - type: mrr_at_100
894
+ value: 93.87100000000001
895
+ - type: mrr_at_1000
896
+ value: 93.87299999999999
897
+ - type: mrr_at_3
898
+ value: 93.415
899
+ - type: mrr_at_5
900
+ value: 93.685
901
+ - type: ndcg_at_1
902
+ value: 90.99199999999999
903
+ - type: ndcg_at_10
904
+ value: 84.57900000000001
905
+ - type: ndcg_at_100
906
+ value: 88.474
907
+ - type: ndcg_at_1000
908
+ value: 89.172
909
+ - type: ndcg_at_3
910
+ value: 86.56099999999999
911
+ - type: ndcg_at_5
912
+ value: 84.811
913
+ - type: precision_at_1
914
+ value: 90.99199999999999
915
+ - type: precision_at_10
916
+ value: 40.969
917
+ - type: precision_at_100
918
+ value: 4.97
919
+ - type: precision_at_1000
920
+ value: 0.515
921
+ - type: precision_at_3
922
+ value: 74.734
923
+ - type: precision_at_5
924
+ value: 61.980999999999995
925
+ - type: recall_at_1
926
+ value: 28.651
927
+ - type: recall_at_10
928
+ value: 83.321
929
+ - type: recall_at_100
930
+ value: 95.498
931
+ - type: recall_at_1000
932
+ value: 98.759
933
+ - type: recall_at_3
934
+ value: 56.708000000000006
935
+ - type: recall_at_5
936
+ value: 70.25200000000001
937
+ - task:
938
+ type: Classification
939
+ dataset:
940
+ type: C-MTEB/TNews-classification
941
+ name: MTEB TNews
942
+ config: default
943
+ split: validation
944
+ revision: None
945
+ metrics:
946
+ - type: accuracy
947
+ value: 52.037
948
+ - type: f1
949
+ value: 50.3832093595745
950
+ - task:
951
+ type: Clustering
952
+ dataset:
953
+ type: C-MTEB/ThuNewsClusteringP2P
954
+ name: MTEB ThuNewsClusteringP2P
955
+ config: default
956
+ split: test
957
+ revision: None
958
+ metrics:
959
+ - type: v_measure
960
+ value: 70.09793315196697
961
+ - task:
962
+ type: Clustering
963
+ dataset:
964
+ type: C-MTEB/ThuNewsClusteringS2S
965
+ name: MTEB ThuNewsClusteringS2S
966
+ config: default
967
+ split: test
968
+ revision: None
969
+ metrics:
970
+ - type: v_measure
971
+ value: 63.66930246094367
972
+ - task:
973
+ type: Retrieval
974
+ dataset:
975
+ type: C-MTEB/VideoRetrieval
976
+ name: MTEB VideoRetrieval
977
+ config: default
978
+ split: dev
979
+ revision: None
980
+ metrics:
981
+ - type: map_at_1
982
+ value: 60.4
983
+ - type: map_at_10
984
+ value: 69.878
985
+ - type: map_at_100
986
+ value: 70.285
987
+ - type: map_at_1000
988
+ value: 70.295
989
+ - type: map_at_3
990
+ value: 68.033
991
+ - type: map_at_5
992
+ value: 69.233
993
+ - type: mrr_at_1
994
+ value: 60.3
995
+ - type: mrr_at_10
996
+ value: 69.828
997
+ - type: mrr_at_100
998
+ value: 70.235
999
+ - type: mrr_at_1000
1000
+ value: 70.245
1001
+ - type: mrr_at_3
1002
+ value: 67.983
1003
+ - type: mrr_at_5
1004
+ value: 69.18299999999999
1005
+ - type: ndcg_at_1
1006
+ value: 60.4
1007
+ - type: ndcg_at_10
1008
+ value: 74.155
1009
+ - type: ndcg_at_100
1010
+ value: 76.173
1011
+ - type: ndcg_at_1000
1012
+ value: 76.44800000000001
1013
+ - type: ndcg_at_3
1014
+ value: 70.44500000000001
1015
+ - type: ndcg_at_5
1016
+ value: 72.61800000000001
1017
+ - type: precision_at_1
1018
+ value: 60.4
1019
+ - type: precision_at_10
1020
+ value: 8.74
1021
+ - type: precision_at_100
1022
+ value: 0.9690000000000001
1023
+ - type: precision_at_1000
1024
+ value: 0.099
1025
+ - type: precision_at_3
1026
+ value: 25.8
1027
+ - type: precision_at_5
1028
+ value: 16.54
1029
+ - type: recall_at_1
1030
+ value: 60.4
1031
+ - type: recall_at_10
1032
+ value: 87.4
1033
+ - type: recall_at_100
1034
+ value: 96.89999999999999
1035
+ - type: recall_at_1000
1036
+ value: 99.1
1037
+ - type: recall_at_3
1038
+ value: 77.4
1039
+ - type: recall_at_5
1040
+ value: 82.69999999999999
1041
+ - task:
1042
+ type: Classification
1043
+ dataset:
1044
+ type: C-MTEB/waimai-classification
1045
+ name: MTEB Waimai
1046
+ config: default
1047
+ split: test
1048
+ revision: None
1049
+ metrics:
1050
+ - type: accuracy
1051
+ value: 88.49000000000001
1052
+ - type: ap
1053
+ value: 73.5441395538586
1054
+ - type: f1
1055
+ value: 86.88114969870975
1056
  ---
1057
+
1058
+
1059
+ # {alime-embedding-large-zh}
1060
+
1061
+ The alime embedding model.
1062
+
1063
+ <!--- Describe your model here -->
1064
+
1065
+ ## Usage (Sentence-Transformers)
1066
+
1067
+ Using this model becomes easy when you have [sentence-transformers](https://www.SBERT.net) installed:
1068
+
1069
+ ```
1070
+ pip install -U sentence-transformers
1071
+ ```
1072
+
1073
+ Then you can use the model like this:
1074
+
1075
+ ```python
1076
+ from sentence_transformers import SentenceTransformer
1077
+ sentences = ["西湖在哪?", "西湖风景名胜区位于浙江省杭州市"]
1078
+
1079
+ model = SentenceTransformer('Pristinenlp/alime-embedding-large-zh')
1080
+ embeddings = model.encode(sentences)
1081
+ print(embeddings)
1082
+ ```
config.json ADDED
@@ -0,0 +1,31 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "BertModel"
4
+ ],
5
+ "attention_probs_dropout_prob": 0.1,
6
+ "classifier_dropout": null,
7
+ "directionality": "bidi",
8
+ "gradient_checkpointing": false,
9
+ "hidden_act": "gelu",
10
+ "hidden_dropout_prob": 0.1,
11
+ "hidden_size": 1024,
12
+ "initializer_range": 0.02,
13
+ "intermediate_size": 4096,
14
+ "layer_norm_eps": 1e-12,
15
+ "max_position_embeddings": 512,
16
+ "model_type": "bert",
17
+ "num_attention_heads": 16,
18
+ "num_hidden_layers": 24,
19
+ "pad_token_id": 0,
20
+ "pooler_fc_size": 768,
21
+ "pooler_num_attention_heads": 12,
22
+ "pooler_num_fc_layers": 3,
23
+ "pooler_size_per_head": 128,
24
+ "pooler_type": "first_token_transform",
25
+ "position_embedding_type": "absolute",
26
+ "torch_dtype": "float32",
27
+ "transformers_version": "4.31.0",
28
+ "type_vocab_size": 2,
29
+ "use_cache": true,
30
+ "vocab_size": 21128
31
+ }
config_sentence_transformers.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "__version__": {
3
+ "sentence_transformers": "2.2.2",
4
+ "transformers": "4.31.0",
5
+ "pytorch": "2.0.1"
6
+ }
7
+ }
modules.json ADDED
@@ -0,0 +1,20 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ [
2
+ {
3
+ "idx": 0,
4
+ "name": "0",
5
+ "path": "",
6
+ "type": "sentence_transformers.models.Transformer"
7
+ },
8
+ {
9
+ "idx": 1,
10
+ "name": "1",
11
+ "path": "1_Pooling",
12
+ "type": "sentence_transformers.models.Pooling"
13
+ },
14
+ {
15
+ "idx": 2,
16
+ "name": "2",
17
+ "path": "2_Normalize",
18
+ "type": "sentence_transformers.models.Normalize"
19
+ }
20
+ ]
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bc38380a162eb4296219f38cba0df2f2bba407d9bd366a1f445c669f5162b78
3
+ size 1302216105
sentence_bert_config.json ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ {
2
+ "max_seq_length": 512,
3
+ "do_lower_case": false
4
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,7 @@
 
 
 
 
 
 
 
 
1
+ {
2
+ "cls_token": "[CLS]",
3
+ "mask_token": "[MASK]",
4
+ "pad_token": "[PAD]",
5
+ "sep_token": "[SEP]",
6
+ "unk_token": "[UNK]"
7
+ }
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,15 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "clean_up_tokenization_spaces": true,
3
+ "cls_token": "[CLS]",
4
+ "do_basic_tokenize": true,
5
+ "do_lower_case": true,
6
+ "mask_token": "[MASK]",
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "never_split": null,
9
+ "pad_token": "[PAD]",
10
+ "sep_token": "[SEP]",
11
+ "strip_accents": null,
12
+ "tokenize_chinese_chars": true,
13
+ "tokenizer_class": "BertTokenizer",
14
+ "unk_token": "[UNK]"
15
+ }
vocab.txt ADDED
The diff for this file is too large to render. See raw diff