eduagarcia commited on
Commit
21c6649
1 Parent(s): 911be8d

Code refactor: moved leaderboards configs from app.py to .yaml files

Browse files
Files changed (4) hide show
  1. app.py +0 -0
  2. config.yaml +351 -0
  3. envs.py +47 -0
  4. model_meta.yaml +1160 -0
app.py CHANGED
The diff for this file is too large to render. See raw diff
 
config.yaml ADDED
@@ -0,0 +1,351 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ version: 0.0.1
2
+ config:
3
+ REPO_ID: "mteb/leaderboard"
4
+ RESULTS_REPO: mteb/results
5
+ LEADERBOARD_NAME: "MTEB Leaderboard"
6
+ tasks:
7
+ BitextMining:
8
+ icon: "🎌"
9
+ metric: f1
10
+ metric_description: "[F1](https://huggingface.co/spaces/evaluate-metric/f1)"
11
+ Classification:
12
+ icon: "❤️"
13
+ metric: accuracy
14
+ metric_description: "[Accuracy](https://huggingface.co/spaces/evaluate-metric/accuracy)"
15
+ Clustering:
16
+ icon: "✨"
17
+ metric: v_measure
18
+ metric_description: "Validity Measure (v_measure)"
19
+ PairClassification:
20
+ icon: "🎭"
21
+ metric: cos_sim_ap
22
+ metric_description: "Average Precision based on Cosine Similarities (cos_sim_ap)"
23
+ Reranking:
24
+ icon: "🥈"
25
+ metric: map
26
+ metric_description: "Mean Average Precision (MAP)"
27
+ Retrieval:
28
+ icon: "🔎"
29
+ metric: ndcg_at_10
30
+ metric_description: "Normalized Discounted Cumulative Gain @ k (ndcg_at_10)"
31
+ STS:
32
+ icon: "🤖"
33
+ metric: cos_sim_spearman
34
+ metric_description: "Spearman correlation based on cosine similarity"
35
+ Summarization:
36
+ icon: "📜"
37
+ metric: cos_sim_spearman
38
+ metric_description: "Spearman correlation based on cosine similarity"
39
+ boards:
40
+ en:
41
+ title: English
42
+ language_long: "English"
43
+ has_overall: true
44
+ acronym: null
45
+ icon: null
46
+ special_icons: null
47
+ credits: null
48
+ tasks:
49
+ Classification:
50
+ - AmazonCounterfactualClassification (en)
51
+ - AmazonPolarityClassification
52
+ - AmazonReviewsClassification (en)
53
+ - Banking77Classification
54
+ - EmotionClassification
55
+ - ImdbClassification
56
+ - MassiveIntentClassification (en)
57
+ - MassiveScenarioClassification (en)
58
+ - MTOPDomainClassification (en)
59
+ - MTOPIntentClassification (en)
60
+ - ToxicConversationsClassification
61
+ - TweetSentimentExtractionClassification
62
+ Clustering:
63
+ - ArxivClusteringP2P
64
+ - ArxivClusteringS2S
65
+ - BiorxivClusteringP2P
66
+ - BiorxivClusteringS2S
67
+ - MedrxivClusteringP2P
68
+ - MedrxivClusteringS2S
69
+ - RedditClustering
70
+ - RedditClusteringP2P
71
+ - StackExchangeClustering
72
+ - StackExchangeClusteringP2P
73
+ - TwentyNewsgroupsClustering
74
+ PairClassification:
75
+ - SprintDuplicateQuestions
76
+ - TwitterSemEval2015
77
+ - TwitterURLCorpus
78
+ Reranking:
79
+ - AskUbuntuDupQuestions
80
+ - MindSmallReranking
81
+ - SciDocsRR
82
+ - StackOverflowDupQuestions
83
+ Retrieval:
84
+ - ArguAna
85
+ - ClimateFEVER
86
+ - CQADupstackRetrieval
87
+ - DBPedia
88
+ - FEVER
89
+ - FiQA2018
90
+ - HotpotQA
91
+ - MSMARCO
92
+ - NFCorpus
93
+ - NQ
94
+ - QuoraRetrieval
95
+ - SCIDOCS
96
+ - SciFact
97
+ - Touche2020
98
+ - TRECCOVID
99
+ STS:
100
+ - BIOSSES
101
+ - SICK-R
102
+ - STS12
103
+ - STS13
104
+ - STS14
105
+ - STS15
106
+ - STS16
107
+ - STS17 (en-en)
108
+ - STS22 (en)
109
+ - STSBenchmark
110
+ Summarization:
111
+ - SummEval
112
+ en-x:
113
+ title: "English-X"
114
+ language_long: " 117 (Pairs of: English & other language)"
115
+ has_overall: false
116
+ acronym: null
117
+ icon: null
118
+ special_icons: null
119
+ credits: null
120
+ tasks:
121
+ BitextMining: ['BUCC (de-en)', 'BUCC (fr-en)', 'BUCC (ru-en)', 'BUCC (zh-en)', 'Tatoeba (afr-eng)', 'Tatoeba (amh-eng)', 'Tatoeba (ang-eng)', 'Tatoeba (ara-eng)', 'Tatoeba (arq-eng)', 'Tatoeba (arz-eng)', 'Tatoeba (ast-eng)', 'Tatoeba (awa-eng)', 'Tatoeba (aze-eng)', 'Tatoeba (bel-eng)', 'Tatoeba (ben-eng)', 'Tatoeba (ber-eng)', 'Tatoeba (bos-eng)', 'Tatoeba (bre-eng)', 'Tatoeba (bul-eng)', 'Tatoeba (cat-eng)', 'Tatoeba (cbk-eng)', 'Tatoeba (ceb-eng)', 'Tatoeba (ces-eng)', 'Tatoeba (cha-eng)', 'Tatoeba (cmn-eng)', 'Tatoeba (cor-eng)', 'Tatoeba (csb-eng)', 'Tatoeba (cym-eng)', 'Tatoeba (dan-eng)', 'Tatoeba (deu-eng)', 'Tatoeba (dsb-eng)', 'Tatoeba (dtp-eng)', 'Tatoeba (ell-eng)', 'Tatoeba (epo-eng)', 'Tatoeba (est-eng)', 'Tatoeba (eus-eng)', 'Tatoeba (fao-eng)', 'Tatoeba (fin-eng)', 'Tatoeba (fra-eng)', 'Tatoeba (fry-eng)', 'Tatoeba (gla-eng)', 'Tatoeba (gle-eng)', 'Tatoeba (glg-eng)', 'Tatoeba (gsw-eng)', 'Tatoeba (heb-eng)', 'Tatoeba (hin-eng)', 'Tatoeba (hrv-eng)', 'Tatoeba (hsb-eng)', 'Tatoeba (hun-eng)', 'Tatoeba (hye-eng)', 'Tatoeba (ido-eng)', 'Tatoeba (ile-eng)', 'Tatoeba (ina-eng)', 'Tatoeba (ind-eng)', 'Tatoeba (isl-eng)', 'Tatoeba (ita-eng)', 'Tatoeba (jav-eng)', 'Tatoeba (jpn-eng)', 'Tatoeba (kab-eng)', 'Tatoeba (kat-eng)', 'Tatoeba (kaz-eng)', 'Tatoeba (khm-eng)', 'Tatoeba (kor-eng)', 'Tatoeba (kur-eng)', 'Tatoeba (kzj-eng)', 'Tatoeba (lat-eng)', 'Tatoeba (lfn-eng)', 'Tatoeba (lit-eng)', 'Tatoeba (lvs-eng)', 'Tatoeba (mal-eng)', 'Tatoeba (mar-eng)', 'Tatoeba (max-eng)', 'Tatoeba (mhr-eng)', 'Tatoeba (mkd-eng)', 'Tatoeba (mon-eng)', 'Tatoeba (nds-eng)', 'Tatoeba (nld-eng)', 'Tatoeba (nno-eng)', 'Tatoeba (nob-eng)', 'Tatoeba (nov-eng)', 'Tatoeba (oci-eng)', 'Tatoeba (orv-eng)', 'Tatoeba (pam-eng)', 'Tatoeba (pes-eng)', 'Tatoeba (pms-eng)', 'Tatoeba (pol-eng)', 'Tatoeba (por-eng)', 'Tatoeba (ron-eng)', 'Tatoeba (rus-eng)', 'Tatoeba (slk-eng)', 'Tatoeba (slv-eng)', 'Tatoeba (spa-eng)', 'Tatoeba (sqi-eng)', 'Tatoeba (srp-eng)', 'Tatoeba (swe-eng)', 'Tatoeba (swg-eng)', 'Tatoeba (swh-eng)', 'Tatoeba (tam-eng)', 'Tatoeba (tat-eng)', 'Tatoeba (tel-eng)', 'Tatoeba (tgl-eng)', 'Tatoeba (tha-eng)', 'Tatoeba (tuk-eng)', 'Tatoeba (tur-eng)', 'Tatoeba (tzl-eng)', 'Tatoeba (uig-eng)', 'Tatoeba (ukr-eng)', 'Tatoeba (urd-eng)', 'Tatoeba (uzb-eng)', 'Tatoeba (vie-eng)', 'Tatoeba (war-eng)', 'Tatoeba (wuu-eng)', 'Tatoeba (xho-eng)', 'Tatoeba (yid-eng)', 'Tatoeba (yue-eng)', 'Tatoeba (zsm-eng)']
122
+ zh:
123
+ title: Chinese
124
+ language_long: Chinese
125
+ has_overall: true
126
+ acronym: C-MTEB
127
+ icon: "🇨🇳"
128
+ special_icons:
129
+ Classification: "🧡"
130
+ credits: "[FlagEmbedding](https://github.com/FlagOpen/FlagEmbedding)"
131
+ tasks:
132
+ Classification:
133
+ - AmazonReviewsClassification (zh)
134
+ - IFlyTek
135
+ - JDReview
136
+ - MassiveIntentClassification (zh-CN)
137
+ - MassiveScenarioClassification (zh-CN)
138
+ - MultilingualSentiment
139
+ - OnlineShopping
140
+ - TNews
141
+ - Waimai
142
+ Clustering:
143
+ - CLSClusteringP2P
144
+ - CLSClusteringS2S
145
+ - ThuNewsClusteringP2P
146
+ - ThuNewsClusteringS2S
147
+ PairClassification:
148
+ - Cmnli
149
+ - Ocnli
150
+ Reranking:
151
+ - CMedQAv1
152
+ - CMedQAv2
153
+ - MMarcoReranking
154
+ - T2Reranking
155
+ Retrieval:
156
+ - CmedqaRetrieval
157
+ - CovidRetrieval
158
+ - DuRetrieval
159
+ - EcomRetrieval
160
+ - MedicalRetrieval
161
+ - MMarcoRetrieval
162
+ - T2Retrieval
163
+ - VideoRetrieval
164
+ STS:
165
+ - AFQMC
166
+ - ATEC
167
+ - BQ
168
+ - LCQMC
169
+ - PAWSX
170
+ - QBQTC
171
+ - STS22 (zh)
172
+ - STSB
173
+ da:
174
+ title: Danish
175
+ language_long: Danish
176
+ has_overall: false
177
+ acronym: null
178
+ icon: "🇩🇰"
179
+ special_icons:
180
+ Classification: "🤍"
181
+ credits: "[Kenneth Enevoldsen](https://github.com/KennethEnevoldsen), [scandinavian-embedding-benchmark](https://kennethenevoldsen.github.io/scandinavian-embedding-benchmark/)"
182
+ tasks:
183
+ BitextMining:
184
+ - BornholmBitextMining
185
+ Classification:
186
+ - AngryTweetsClassification
187
+ - DanishPoliticalCommentsClassification
188
+ - DKHateClassification
189
+ - LccSentimentClassification
190
+ - MassiveIntentClassification (da)
191
+ - MassiveScenarioClassification (da)
192
+ - NordicLangClassification
193
+ - ScalaDaClassification
194
+ fr:
195
+ title: French
196
+ language_long: "French"
197
+ has_overall: true
198
+ acronym: "F-MTEB"
199
+ icon: "🇫🇷"
200
+ special_icons:
201
+ Classification: "💙"
202
+ credits: "[Lyon-NLP](https://github.com/Lyon-NLP): [Gabriel Sequeira](https://github.com/GabrielSequeira), [Imene Kerboua](https://github.com/imenelydiaker), [Wissam Siblini](https://github.com/wissam-sib), [Mathieu Ciancone](https://github.com/MathieuCiancone), [Marion Schaeffer](https://github.com/schmarion)"
203
+ tasks:
204
+ Classification:
205
+ - AmazonReviewsClassification (fr)
206
+ - MasakhaNEWSClassification (fra)
207
+ - MassiveIntentClassification (fr)
208
+ - MassiveScenarioClassification (fr)
209
+ - MTOPDomainClassification (fr)
210
+ - MTOPIntentClassification (fr)
211
+ Clustering:
212
+ - AlloProfClusteringP2P
213
+ - AlloProfClusteringS2S
214
+ - HALClusteringS2S
215
+ - MLSUMClusteringP2P
216
+ - MLSUMClusteringS2S
217
+ - MasakhaNEWSClusteringP2P (fra)
218
+ - MasakhaNEWSClusteringS2S (fra)
219
+ PairClassification:
220
+ - OpusparcusPC (fr)
221
+ - PawsX (fr)
222
+ Reranking:
223
+ - AlloprofReranking
224
+ - SyntecReranking
225
+ Retrieval:
226
+ - AlloprofRetrieval
227
+ - BSARDRetrieval
228
+ - MintakaRetrieval (fr)
229
+ - SyntecRetrieval
230
+ - XPQARetrieval (fr)
231
+ STS:
232
+ - STS22 (fr)
233
+ - STSBenchmarkMultilingualSTS (fr)
234
+ - SICKFr
235
+ Summarization:
236
+ - SummEvalFr
237
+ 'no':
238
+ title: Norwegian
239
+ language_long: "Norwegian Bokmål"
240
+ has_overall: false
241
+ acronym: null
242
+ icon: "🇳🇴"
243
+ special_icons:
244
+ Classification: "💙"
245
+ credits: "[Kenneth Enevoldsen](https://github.com/KennethEnevoldsen), [scandinavian-embedding-benchmark](https://kennethenevoldsen.github.io/scandinavian-embedding-benchmark/)"
246
+ tasks:
247
+ Classification: &id001
248
+ - NoRecClassification
249
+ - NordicLangClassification
250
+ - NorwegianParliament
251
+ - MassiveIntentClassification (nb)
252
+ - MassiveScenarioClassification (nb)
253
+ - ScalaNbClassification
254
+ law:
255
+ title: Law
256
+ language_long: "English, German, Chinese"
257
+ has_overall: false
258
+ acronym: null
259
+ icon: "⚖️"
260
+ special_icons: null
261
+ credits: "[Voyage AI](https://www.voyageai.com/)"
262
+ tasks:
263
+ Retrieval:
264
+ - AILACasedocs
265
+ - AILAStatutes
266
+ - GerDaLIRSmall
267
+ - LeCaRDv2
268
+ - LegalBenchConsumerContractsQA
269
+ - LegalBenchCorporateLobbying
270
+ - LegalQuAD
271
+ - LegalSummarization
272
+ pl:
273
+ title: Polish
274
+ language_long: Polish
275
+ has_overall: true
276
+ acronym: null
277
+ icon: "🇵🇱"
278
+ special_icons:
279
+ Classification: "🤍"
280
+ credits: "[Rafał Poświata](https://github.com/rafalposwiata)"
281
+ tasks:
282
+ Classification:
283
+ - AllegroReviews
284
+ - CBD
285
+ - MassiveIntentClassification (pl)
286
+ - MassiveScenarioClassification (pl)
287
+ - PAC
288
+ - PolEmo2.0-IN
289
+ - PolEmo2.0-OUT
290
+ Clustering:
291
+ - 8TagsClustering
292
+ PairClassification:
293
+ - CDSC-E
294
+ - PPC
295
+ - PSC
296
+ - SICK-E-PL
297
+ Retrieval:
298
+ - ArguAna-PL
299
+ - DBPedia-PL
300
+ - FiQA-PL
301
+ - HotpotQA-PL
302
+ - MSMARCO-PL
303
+ - NFCorpus-PL
304
+ - NQ-PL
305
+ - Quora-PL
306
+ - SCIDOCS-PL
307
+ - SciFact-PL
308
+ - TRECCOVID-PL
309
+ STS:
310
+ - CDSC-R
311
+ - SICK-R-PL
312
+ - STS22 (pl)
313
+ se:
314
+ title: Swedish
315
+ language_long: Swedish
316
+ has_overall: false
317
+ acronym: null
318
+ icon: "🇸🇪"
319
+ special_icons:
320
+ Classification: "💛"
321
+ credits: "[Kenneth Enevoldsen](https://github.com/KennethEnevoldsen), [scandinavian-embedding-benchmark](https://kennethenevoldsen.github.io/scandinavian-embedding-benchmark/)"
322
+ tasks:
323
+ Classification:
324
+ - NoRecClassification
325
+ - NordicLangClassification
326
+ - NorwegianParliament
327
+ - MassiveIntentClassification (nb)
328
+ - MassiveScenarioClassification (nb)
329
+ - ScalaNbClassification
330
+ other-cls:
331
+ title: "Other Languages"
332
+ language_long: "47 (Only languages not included in the other tabs)"
333
+ has_overall: false
334
+ acronym: null
335
+ icon: null
336
+ special_icons:
337
+ Classification: "💜💚💙"
338
+ credits: null
339
+ tasks:
340
+ Classification: ['AmazonCounterfactualClassification (de)', 'AmazonCounterfactualClassification (ja)', 'AmazonReviewsClassification (de)', 'AmazonReviewsClassification (es)', 'AmazonReviewsClassification (fr)', 'AmazonReviewsClassification (ja)', 'AmazonReviewsClassification (zh)', 'MTOPDomainClassification (de)', 'MTOPDomainClassification (es)', 'MTOPDomainClassification (fr)', 'MTOPDomainClassification (hi)', 'MTOPDomainClassification (th)', 'MTOPIntentClassification (de)', 'MTOPIntentClassification (es)', 'MTOPIntentClassification (fr)', 'MTOPIntentClassification (hi)', 'MTOPIntentClassification (th)', 'MassiveIntentClassification (af)', 'MassiveIntentClassification (am)', 'MassiveIntentClassification (ar)', 'MassiveIntentClassification (az)', 'MassiveIntentClassification (bn)', 'MassiveIntentClassification (cy)', 'MassiveIntentClassification (de)', 'MassiveIntentClassification (el)', 'MassiveIntentClassification (es)', 'MassiveIntentClassification (fa)', 'MassiveIntentClassification (fi)', 'MassiveIntentClassification (fr)', 'MassiveIntentClassification (he)', 'MassiveIntentClassification (hi)', 'MassiveIntentClassification (hu)', 'MassiveIntentClassification (hy)', 'MassiveIntentClassification (id)', 'MassiveIntentClassification (is)', 'MassiveIntentClassification (it)', 'MassiveIntentClassification (ja)', 'MassiveIntentClassification (jv)', 'MassiveIntentClassification (ka)', 'MassiveIntentClassification (km)', 'MassiveIntentClassification (kn)', 'MassiveIntentClassification (ko)', 'MassiveIntentClassification (lv)', 'MassiveIntentClassification (ml)', 'MassiveIntentClassification (mn)', 'MassiveIntentClassification (ms)', 'MassiveIntentClassification (my)', 'MassiveIntentClassification (nl)', 'MassiveIntentClassification (pt)', 'MassiveIntentClassification (ro)', 'MassiveIntentClassification (ru)', 'MassiveIntentClassification (sl)', 'MassiveIntentClassification (sq)', 'MassiveIntentClassification (sw)', 'MassiveIntentClassification (ta)', 'MassiveIntentClassification (te)', 'MassiveIntentClassification (th)', 'MassiveIntentClassification (tl)', 'MassiveIntentClassification (tr)', 'MassiveIntentClassification (ur)', 'MassiveIntentClassification (vi)', 'MassiveIntentClassification (zh-TW)', 'MassiveScenarioClassification (af)', 'MassiveScenarioClassification (am)', 'MassiveScenarioClassification (ar)', 'MassiveScenarioClassification (az)', 'MassiveScenarioClassification (bn)', 'MassiveScenarioClassification (cy)', 'MassiveScenarioClassification (de)', 'MassiveScenarioClassification (el)', 'MassiveScenarioClassification (es)', 'MassiveScenarioClassification (fa)', 'MassiveScenarioClassification (fi)', 'MassiveScenarioClassification (fr)', 'MassiveScenarioClassification (he)', 'MassiveScenarioClassification (hi)', 'MassiveScenarioClassification (hu)', 'MassiveScenarioClassification (hy)', 'MassiveScenarioClassification (id)', 'MassiveScenarioClassification (is)', 'MassiveScenarioClassification (it)', 'MassiveScenarioClassification (ja)', 'MassiveScenarioClassification (jv)', 'MassiveScenarioClassification (ka)', 'MassiveScenarioClassification (km)', 'MassiveScenarioClassification (kn)', 'MassiveScenarioClassification (ko)', 'MassiveScenarioClassification (lv)', 'MassiveScenarioClassification (ml)', 'MassiveScenarioClassification (mn)', 'MassiveScenarioClassification (ms)', 'MassiveScenarioClassification (my)', 'MassiveScenarioClassification (nl)', 'MassiveScenarioClassification (pt)', 'MassiveScenarioClassification (ro)', 'MassiveScenarioClassification (ru)', 'MassiveScenarioClassification (sl)', 'MassiveScenarioClassification (sq)', 'MassiveScenarioClassification (sw)', 'MassiveScenarioClassification (ta)', 'MassiveScenarioClassification (te)', 'MassiveScenarioClassification (th)', 'MassiveScenarioClassification (tl)', 'MassiveScenarioClassification (tr)', 'MassiveScenarioClassification (ur)', 'MassiveScenarioClassification (vi)', 'MassiveScenarioClassification (zh-TW)']
341
+ other-sts:
342
+ title: Other
343
+ language_long: "Arabic, Chinese, Dutch, English, French, German, Italian, Korean, Polish, Russian, Spanish (Only language combos not included in the other tabs)"
344
+ has_overall: false
345
+ acronym: null
346
+ icon: null
347
+ special_icons:
348
+ STS: "👽"
349
+ credits: null
350
+ tasks:
351
+ STS: ["STS17 (ar-ar)", "STS17 (en-ar)", "STS17 (en-de)", "STS17 (en-tr)", "STS17 (es-en)", "STS17 (es-es)", "STS17 (fr-en)", "STS17 (it-en)", "STS17 (ko-ko)", "STS17 (nl-en)", "STS22 (ar)", "STS22 (de)", "STS22 (de-en)", "STS22 (de-fr)", "STS22 (de-pl)", "STS22 (es)", "STS22 (es-en)", "STS22 (es-it)", "STS22 (fr)", "STS22 (fr-pl)", "STS22 (it)", "STS22 (pl)", "STS22 (pl-en)", "STS22 (ru)", "STS22 (tr)", "STS22 (zh-en)", "STSBenchmark"]
envs.py ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ from yaml import safe_load
3
+
4
+ from huggingface_hub import HfApi
5
+
6
+ LEADERBOARD_CONFIG_PATH = "config.yaml"
7
+ with open(LEADERBOARD_CONFIG_PATH, 'r', encoding='utf-8') as f:
8
+ LEADERBOARD_CONFIG = safe_load(f)
9
+ MODEL_META_PATH = "model_meta.yaml"
10
+ with open(MODEL_META_PATH, 'r', encoding='utf-8') as f:
11
+ MODEL_META = safe_load(f)
12
+
13
+ # Try first to get the config from the environment variables, then from the config.yaml file
14
+ def get_config(name, default):
15
+ res = None
16
+
17
+ if name in os.environ:
18
+ res = os.environ[name]
19
+ elif 'config' in LEADERBOARD_CONFIG:
20
+ res = LEADERBOARD_CONFIG['config'].get(name, None)
21
+
22
+ if res is None:
23
+ return default
24
+ return res
25
+
26
+ def str2bool(v):
27
+ return str(v).lower() in ("yes", "true", "t", "1")
28
+
29
+ # clone / pull the lmeh eval data
30
+ HF_TOKEN = get_config("HF_TOKEN", None)
31
+
32
+ LEADERBOARD_NAME = get_config("LEADERBOARD_NAME", "MTEB Leaderboard")
33
+
34
+ REPO_ID = get_config("REPO_ID", "mteb/leaderboard")
35
+ RESULTS_REPO = get_config("RESULTS_REPO", "mteb/results")
36
+
37
+ CACHE_PATH=get_config("HF_HOME", ".")
38
+ os.environ["HF_HOME"] = CACHE_PATH
39
+
40
+ if not os.access(CACHE_PATH, os.W_OK):
41
+ print(f"No write access to HF_HOME: {CACHE_PATH}. Resetting to current directory.")
42
+ CACHE_PATH = "."
43
+ os.environ["HF_HOME"] = CACHE_PATH
44
+ else:
45
+ print(f"Write access confirmed for HF_HOME")
46
+
47
+ API = HfApi(token=HF_TOKEN)
model_meta.yaml ADDED
@@ -0,0 +1,1160 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ models_meta:
2
+ Baichuan-text-embedding:
3
+ dim: 1024
4
+ is_external: true
5
+ is_proprietary: true
6
+ is_sentence_transformers_compatible: false
7
+ link: https://platform.baichuan-ai.com/docs/text-Embedding
8
+ seq_len: 512
9
+ size: null
10
+ Cohere-embed-english-v3.0:
11
+ dim: 1024
12
+ is_external: true
13
+ is_proprietary: true
14
+ is_sentence_transformers_compatible: false
15
+ link: https://huggingface.co/Cohere/Cohere-embed-english-v3.0
16
+ seq_len: 512
17
+ size: null
18
+ Cohere-embed-multilingual-light-v3.0:
19
+ dim: 384
20
+ is_external: true
21
+ is_proprietary: true
22
+ is_sentence_transformers_compatible: false
23
+ link: https://huggingface.co/Cohere/Cohere-embed-multilingual-light-v3.0
24
+ seq_len: 512
25
+ size: null
26
+ Cohere-embed-multilingual-v3.0:
27
+ dim: 1024
28
+ is_external: true
29
+ is_proprietary: true
30
+ is_sentence_transformers_compatible: false
31
+ link: https://huggingface.co/Cohere/Cohere-embed-multilingual-v3.0
32
+ seq_len: 512
33
+ size: null
34
+ DanskBERT:
35
+ dim: 768
36
+ is_external: true
37
+ is_proprietary: false
38
+ is_sentence_transformers_compatible: true
39
+ link: https://huggingface.co/vesteinn/DanskBERT
40
+ seq_len: 514
41
+ size: 125
42
+ LASER2:
43
+ dim: 1024
44
+ is_external: true
45
+ is_proprietary: false
46
+ is_sentence_transformers_compatible: false
47
+ link: https://github.com/facebookresearch/LASER
48
+ seq_len: N/A
49
+ size: 43
50
+ LLM2Vec-Llama-supervised:
51
+ dim: 4096
52
+ is_external: true
53
+ is_proprietary: false
54
+ is_sentence_transformers_compatible: false
55
+ link: https://huggingface.co/McGill-NLP/LLM2Vec-Llama-2-7b-chat-hf-mntp-supervised
56
+ seq_len: 4096
57
+ size: 6607
58
+ LLM2Vec-Llama-unsupervised:
59
+ dim: 4096
60
+ is_external: true
61
+ is_proprietary: false
62
+ is_sentence_transformers_compatible: false
63
+ link: https://huggingface.co/McGill-NLP/LLM2Vec-Llama-2-7b-chat-hf-mntp
64
+ seq_len: 4096
65
+ size: 6607
66
+ LLM2Vec-Mistral-supervised:
67
+ dim: 4096
68
+ is_external: true
69
+ is_proprietary: false
70
+ is_sentence_transformers_compatible: false
71
+ link: https://huggingface.co/McGill-NLP/LLM2Vec-Mistral-7B-Instruct-v2-mntp-supervised
72
+ seq_len: 32768
73
+ size: 7111
74
+ LLM2Vec-Mistral-unsupervised:
75
+ dim: 4096
76
+ is_external: true
77
+ is_proprietary: false
78
+ is_sentence_transformers_compatible: false
79
+ link: https://huggingface.co/McGill-NLP/LLM2Vec-Mistral-7B-Instruct-v2-mntp
80
+ seq_len: 32768
81
+ size: 7111
82
+ LLM2Vec-Sheared-Llama-supervised:
83
+ dim: 2048
84
+ is_external: true
85
+ is_proprietary: false
86
+ is_sentence_transformers_compatible: false
87
+ link: https://huggingface.co/McGill-NLP/LLM2Vec-Sheared-LLaMA-mntp-supervised
88
+ seq_len: 4096
89
+ size: 1280
90
+ LLM2Vec-Sheared-Llama-unsupervised:
91
+ dim: 2048
92
+ is_external: true
93
+ is_proprietary: false
94
+ is_sentence_transformers_compatible: false
95
+ link: https://huggingface.co/McGill-NLP/LLM2Vec-Sheared-LLaMA-mntp
96
+ seq_len: 4096
97
+ size: 1280
98
+ LaBSE:
99
+ dim: 768
100
+ is_external: true
101
+ is_proprietary: false
102
+ is_sentence_transformers_compatible: true
103
+ link: https://huggingface.co/sentence-transformers/LaBSE
104
+ seq_len: 512
105
+ size: 471
106
+ OpenSearch-text-hybrid:
107
+ dim: 1792
108
+ is_external: true
109
+ is_proprietary: true
110
+ is_sentence_transformers_compatible: false
111
+ link: https://help.aliyun.com/zh/open-search/vector-search-edition/hybrid-retrieval
112
+ seq_len: 512
113
+ size: null
114
+ all-MiniLM-L12-v2:
115
+ dim: 384
116
+ is_external: true
117
+ is_proprietary: false
118
+ is_sentence_transformers_compatible: true
119
+ link: https://huggingface.co/sentence-transformers/all-MiniLM-L12-v2
120
+ seq_len: 512
121
+ size: 33
122
+ all-MiniLM-L6-v2:
123
+ dim: 384
124
+ is_external: true
125
+ is_proprietary: false
126
+ is_sentence_transformers_compatible: true
127
+ link: https://huggingface.co/sentence-transformers/all-MiniLM-L6-v2
128
+ seq_len: 512
129
+ size: 23
130
+ all-mpnet-base-v2:
131
+ dim: 768
132
+ is_external: true
133
+ is_proprietary: false
134
+ is_sentence_transformers_compatible: true
135
+ link: https://huggingface.co/sentence-transformers/all-mpnet-base-v2
136
+ seq_len: 514
137
+ size: 110
138
+ allenai-specter:
139
+ dim: 768
140
+ is_external: true
141
+ is_proprietary: false
142
+ is_sentence_transformers_compatible: true
143
+ link: https://huggingface.co/sentence-transformers/allenai-specter
144
+ seq_len: 512
145
+ size: 110
146
+ bert-base-10lang-cased:
147
+ dim: 768
148
+ is_external: true
149
+ is_proprietary: false
150
+ is_sentence_transformers_compatible: true
151
+ link: https://huggingface.co/Geotrend/bert-base-10lang-cased
152
+ seq_len: 512
153
+ size: 138
154
+ bert-base-15lang-cased:
155
+ dim: 768
156
+ is_external: true
157
+ is_proprietary: false
158
+ is_sentence_transformers_compatible: true
159
+ link: https://huggingface.co/Geotrend/bert-base-15lang-cased
160
+ seq_len: 512
161
+ size: 138
162
+ bert-base-25lang-cased:
163
+ dim: 768
164
+ is_external: true
165
+ is_proprietary: false
166
+ is_sentence_transformers_compatible: true
167
+ link: https://huggingface.co/Geotrend/bert-base-25lang-cased
168
+ seq_len: 512
169
+ size: 138
170
+ bert-base-multilingual-cased:
171
+ dim: 768
172
+ is_external: true
173
+ is_proprietary: false
174
+ is_sentence_transformers_compatible: true
175
+ link: https://huggingface.co/google-bert/bert-base-multilingual-cased
176
+ seq_len: 512
177
+ size: 179
178
+ bert-base-multilingual-uncased:
179
+ dim: 768
180
+ is_external: true
181
+ is_proprietary: false
182
+ is_sentence_transformers_compatible: true
183
+ link: https://huggingface.co/google-bert/bert-base-multilingual-uncased
184
+ seq_len: 512
185
+ size: 168
186
+ bert-base-swedish-cased:
187
+ dim: 768
188
+ is_external: true
189
+ is_proprietary: false
190
+ is_sentence_transformers_compatible: true
191
+ link: https://huggingface.co/KB/bert-base-swedish-cased
192
+ seq_len: 512
193
+ size: 125
194
+ bert-base-uncased:
195
+ dim: 768
196
+ is_external: true
197
+ is_proprietary: false
198
+ is_sentence_transformers_compatible: true
199
+ link: https://huggingface.co/bert-base-uncased
200
+ seq_len: 512
201
+ size: 110
202
+ bge-base-zh-v1.5:
203
+ dim: 768
204
+ is_external: true
205
+ is_proprietary: false
206
+ is_sentence_transformers_compatible: true
207
+ link: https://huggingface.co/BAAI/bge-base-zh-v1.5
208
+ seq_len: 512
209
+ size: 102
210
+ bge-large-en-v1.5:
211
+ dim: 1024
212
+ is_external: true
213
+ is_proprietary: false
214
+ is_sentence_transformers_compatible: false
215
+ link: https://huggingface.co/BAAI/bge-large-en-v1.5
216
+ seq_len: 512
217
+ size: null
218
+ bge-large-zh-noinstruct:
219
+ dim: 1024
220
+ is_external: true
221
+ is_proprietary: false
222
+ is_sentence_transformers_compatible: true
223
+ link: https://huggingface.co/BAAI/bge-large-zh-noinstruct
224
+ seq_len: 512
225
+ size: 326
226
+ bge-large-zh-v1.5:
227
+ dim: 1024
228
+ is_external: true
229
+ is_proprietary: false
230
+ is_sentence_transformers_compatible: true
231
+ link: https://huggingface.co/BAAI/bge-large-zh-v1.5
232
+ seq_len: 512
233
+ size: 326
234
+ bge-small-zh-v1.5:
235
+ dim: 512
236
+ is_external: true
237
+ is_proprietary: false
238
+ is_sentence_transformers_compatible: true
239
+ link: https://huggingface.co/BAAI/bge-small-zh-v1.5
240
+ seq_len: 512
241
+ size: 24
242
+ camembert-base:
243
+ dim: 512
244
+ is_external: false
245
+ is_proprietary: false
246
+ is_sentence_transformers_compatible: true
247
+ link: https://huggingface.co/almanach/camembert-base
248
+ seq_len: 512
249
+ size: 111
250
+ camembert-large:
251
+ dim: 768
252
+ is_external: false
253
+ is_proprietary: false
254
+ is_sentence_transformers_compatible: true
255
+ link: https://huggingface.co/almanach/camembert-large
256
+ seq_len: 512
257
+ size: 338
258
+ contriever-base-msmarco:
259
+ dim: 768
260
+ is_external: true
261
+ is_proprietary: false
262
+ is_sentence_transformers_compatible: true
263
+ link: https://huggingface.co/nthakur/contriever-base-msmarco
264
+ seq_len: 512
265
+ size: 110
266
+ cross-en-de-roberta-sentence-transformer:
267
+ dim: 768
268
+ is_external: true
269
+ is_proprietary: false
270
+ is_sentence_transformers_compatible: true
271
+ link: https://huggingface.co/T-Systems-onsite/cross-en-de-roberta-sentence-transformer
272
+ seq_len: 514
273
+ size: 278
274
+ dfm-encoder-large-v1:
275
+ dim: 1024
276
+ is_external: true
277
+ is_proprietary: false
278
+ is_sentence_transformers_compatible: true
279
+ link: https://huggingface.co/chcaa/dfm-encoder-large-v1
280
+ seq_len: 512
281
+ size: 355
282
+ dfm-sentence-encoder-large-1:
283
+ dim: 1024
284
+ is_external: true
285
+ is_proprietary: false
286
+ is_sentence_transformers_compatible: true
287
+ link: https://huggingface.co/chcaa/dfm-encoder-large-v1
288
+ seq_len: 512
289
+ size: 355
290
+ distilbert-base-25lang-cased:
291
+ dim: 768
292
+ is_external: false
293
+ is_proprietary: false
294
+ is_sentence_transformers_compatible: true
295
+ link: https://huggingface.co/Geotrend/distilbert-base-25lang-cased
296
+ seq_len: 512
297
+ size: 110
298
+ distilbert-base-en-fr-cased:
299
+ dim: 768
300
+ is_external: false
301
+ is_proprietary: false
302
+ is_sentence_transformers_compatible: true
303
+ link: https://huggingface.co/Geotrend/distilbert-base-en-fr-cased
304
+ seq_len: 512
305
+ size: 110
306
+ distilbert-base-en-fr-es-pt-it-cased:
307
+ dim: 768
308
+ is_external: false
309
+ is_proprietary: false
310
+ is_sentence_transformers_compatible: true
311
+ link: https://huggingface.co/Geotrend/distilbert-base-en-fr-es-pt-it-cased
312
+ seq_len: 512
313
+ size: 110
314
+ distilbert-base-fr-cased:
315
+ dim: 768
316
+ is_external: false
317
+ is_proprietary: false
318
+ is_sentence_transformers_compatible: true
319
+ link: https://huggingface.co/Geotrend/distilbert-base-fr-cased
320
+ seq_len: 512
321
+ size: 110
322
+ distilbert-base-uncased:
323
+ dim: 768
324
+ is_external: false
325
+ is_proprietary: false
326
+ is_sentence_transformers_compatible: true
327
+ link: https://huggingface.co/distilbert-base-uncased
328
+ seq_len: 512
329
+ size: 110
330
+ distiluse-base-multilingual-cased-v2:
331
+ dim: 512
332
+ is_external: true
333
+ is_proprietary: false
334
+ is_sentence_transformers_compatible: true
335
+ link: https://huggingface.co/sentence-transformers/distiluse-base-multilingual-cased-v2
336
+ seq_len: 512
337
+ size: 135
338
+ e5-base:
339
+ dim: 768
340
+ is_external: true
341
+ is_proprietary: false
342
+ is_sentence_transformers_compatible: true
343
+ link: https://huggingface.co/intfloat/e5-base
344
+ seq_len: 512
345
+ size: 110
346
+ e5-large:
347
+ dim: 1024
348
+ is_external: true
349
+ is_proprietary: false
350
+ is_sentence_transformers_compatible: true
351
+ link: https://huggingface.co/intfloat/e5-large
352
+ seq_len: 512
353
+ size: 335
354
+ e5-mistral-7b-instruct:
355
+ dim: 4096
356
+ is_external: true
357
+ is_proprietary: false
358
+ is_sentence_transformers_compatible: true
359
+ link: https://huggingface.co/intfloat/e5-mistral-7b-instruct
360
+ seq_len: 32768
361
+ size: 7111
362
+ e5-small:
363
+ dim: 384
364
+ is_external: true
365
+ is_proprietary: false
366
+ is_sentence_transformers_compatible: true
367
+ link: https://huggingface.co/intfloat/e5-small
368
+ seq_len: 512
369
+ size: 33
370
+ electra-small-nordic:
371
+ dim: 256
372
+ is_external: true
373
+ is_proprietary: false
374
+ is_sentence_transformers_compatible: true
375
+ link: https://huggingface.co/jonfd/electra-small-nordic
376
+ seq_len: 512
377
+ size: 23
378
+ electra-small-swedish-cased-discriminator:
379
+ dim: 256
380
+ is_external: true
381
+ is_proprietary: false
382
+ is_sentence_transformers_compatible: true
383
+ link: https://huggingface.co/KBLab/electra-small-swedish-cased-discriminator
384
+ seq_len: 512
385
+ size: 16
386
+ flaubert_base_cased:
387
+ dim: 768
388
+ is_external: true
389
+ is_proprietary: false
390
+ is_sentence_transformers_compatible: true
391
+ link: https://huggingface.co/flaubert/flaubert_base_cased
392
+ seq_len: 512
393
+ size: 138
394
+ flaubert_base_uncased:
395
+ dim: 768
396
+ is_external: true
397
+ is_proprietary: false
398
+ is_sentence_transformers_compatible: true
399
+ link: https://huggingface.co/flaubert/flaubert_base_uncased
400
+ seq_len: 512
401
+ size: 138
402
+ flaubert_large_cased:
403
+ dim: 1024
404
+ is_external: true
405
+ is_proprietary: false
406
+ is_sentence_transformers_compatible: true
407
+ link: https://huggingface.co/flaubert/flaubert_large_cased
408
+ seq_len: 512
409
+ size: 372
410
+ gbert-base:
411
+ dim: 768
412
+ is_external: true
413
+ is_proprietary: false
414
+ is_sentence_transformers_compatible: true
415
+ link: https://huggingface.co/deepset/gbert-base
416
+ seq_len: 512
417
+ size: 110
418
+ gbert-large:
419
+ dim: 1024
420
+ is_external: true
421
+ is_proprietary: false
422
+ is_sentence_transformers_compatible: true
423
+ link: https://huggingface.co/deepset/gbert-large
424
+ seq_len: 512
425
+ size: 337
426
+ gelectra-base:
427
+ dim: 768
428
+ is_external: true
429
+ is_proprietary: false
430
+ is_sentence_transformers_compatible: true
431
+ link: https://huggingface.co/deepset/gelectra-base
432
+ seq_len: 512
433
+ size: 110
434
+ gelectra-large:
435
+ dim: 1024
436
+ is_external: true
437
+ is_proprietary: false
438
+ is_sentence_transformers_compatible: true
439
+ link: https://huggingface.co/deepset/gelectra-large
440
+ seq_len: 512
441
+ size: 335
442
+ glove.6B.300d:
443
+ dim: 300
444
+ is_external: true
445
+ is_proprietary: false
446
+ is_sentence_transformers_compatible: true
447
+ link: https://huggingface.co/sentence-transformers/average_word_embeddings_glove.6B.300d
448
+ seq_len: N/A
449
+ size: 120
450
+ google-gecko-256.text-embedding-preview-0409:
451
+ dim: 256
452
+ is_external: true
453
+ is_proprietary: true
454
+ is_sentence_transformers_compatible: false
455
+ link: https://cloud.google.com/vertex-ai/generative-ai/docs/embeddings/get-text-embeddings#latest_models
456
+ seq_len: 2048
457
+ size: 1200
458
+ google-gecko.text-embedding-preview-0409:
459
+ dim: 768
460
+ is_external: true
461
+ is_proprietary: true
462
+ is_sentence_transformers_compatible: false
463
+ link: https://cloud.google.com/vertex-ai/generative-ai/docs/embeddings/get-text-embeddings#latest_models
464
+ seq_len: 2048
465
+ size: 1200
466
+ gottbert-base:
467
+ dim: 768
468
+ is_external: true
469
+ is_proprietary: false
470
+ is_sentence_transformers_compatible: true
471
+ link: https://huggingface.co/uklfr/gottbert-base
472
+ seq_len: 512
473
+ size: 127
474
+ gtr-t5-base:
475
+ dim: 768
476
+ is_external: true
477
+ is_proprietary: false
478
+ is_sentence_transformers_compatible: true
479
+ link: https://huggingface.co/sentence-transformers/gtr-t5-base
480
+ seq_len: 512
481
+ size: 110
482
+ gtr-t5-large:
483
+ dim: 768
484
+ is_external: true
485
+ is_proprietary: false
486
+ is_sentence_transformers_compatible: true
487
+ link: https://huggingface.co/sentence-transformers/gtr-t5-large
488
+ seq_len: 512
489
+ size: 168
490
+ gtr-t5-xl:
491
+ dim: 768
492
+ is_external: true
493
+ is_proprietary: false
494
+ is_sentence_transformers_compatible: true
495
+ link: https://huggingface.co/sentence-transformers/gtr-t5-xl
496
+ seq_len: 512
497
+ size: 1240
498
+ gtr-t5-xxl:
499
+ dim: 768
500
+ is_external: true
501
+ is_proprietary: false
502
+ is_sentence_transformers_compatible: true
503
+ link: https://huggingface.co/sentence-transformers/gtr-t5-xxl
504
+ seq_len: 512
505
+ size: 4865
506
+ herbert-base-retrieval-v2:
507
+ dim: 768
508
+ is_external: true
509
+ is_proprietary: false
510
+ is_sentence_transformers_compatible: true
511
+ link: https://huggingface.co/ipipan/herbert-base-retrieval-v2
512
+ seq_len: 514
513
+ size: 125
514
+ komninos:
515
+ dim: 300
516
+ is_external: true
517
+ is_proprietary: false
518
+ is_sentence_transformers_compatible: true
519
+ link: https://huggingface.co/sentence-transformers/average_word_embeddings_komninos
520
+ seq_len: N/A
521
+ size: 134
522
+ luotuo-bert-medium:
523
+ dim: 768
524
+ is_external: true
525
+ is_proprietary: false
526
+ is_sentence_transformers_compatible: true
527
+ link: https://huggingface.co/silk-road/luotuo-bert-medium
528
+ seq_len: 512
529
+ size: 328
530
+ m3e-base:
531
+ dim: 768
532
+ is_external: true
533
+ is_proprietary: false
534
+ is_sentence_transformers_compatible: true
535
+ link: https://huggingface.co/moka-ai/m3e-base
536
+ seq_len: 512
537
+ size: 102
538
+ m3e-large:
539
+ dim: 768
540
+ is_external: true
541
+ is_proprietary: false
542
+ is_sentence_transformers_compatible: true
543
+ link: https://huggingface.co/moka-ai/m3e-large
544
+ seq_len: 512
545
+ size: 102
546
+ mistral-embed:
547
+ dim: 1024
548
+ is_external: true
549
+ is_proprietary: true
550
+ is_sentence_transformers_compatible: false
551
+ link: https://docs.mistral.ai/guides/embeddings
552
+ seq_len: null
553
+ size: null
554
+ msmarco-bert-co-condensor:
555
+ dim: 768
556
+ is_external: true
557
+ is_proprietary: false
558
+ is_sentence_transformers_compatible: true
559
+ link: https://huggingface.co/sentence-transformers/msmarco-bert-co-condensor
560
+ seq_len: 512
561
+ size: 110
562
+ multi-qa-MiniLM-L6-cos-v1:
563
+ dim: 384
564
+ is_external: true
565
+ is_proprietary: false
566
+ is_sentence_transformers_compatible: true
567
+ link: https://huggingface.co/sentence-transformers/multi-qa-MiniLM-L6-cos-v1
568
+ seq_len: 512
569
+ size: 23
570
+ multilingual-e5-base:
571
+ dim: 768
572
+ is_external: true
573
+ is_proprietary: false
574
+ is_sentence_transformers_compatible: true
575
+ link: https://huggingface.co/intfloat/multilingual-e5-base
576
+ seq_len: 514
577
+ size: 278
578
+ multilingual-e5-large:
579
+ dim: 1024
580
+ is_external: true
581
+ is_proprietary: false
582
+ is_sentence_transformers_compatible: true
583
+ link: https://huggingface.co/intfloat/multilingual-e5-large
584
+ seq_len: 514
585
+ size: 560
586
+ multilingual-e5-small:
587
+ dim: 384
588
+ is_external: true
589
+ is_proprietary: false
590
+ is_sentence_transformers_compatible: true
591
+ link: https://huggingface.co/intfloat/multilingual-e5-small
592
+ seq_len: 512
593
+ size: 118
594
+ nb-bert-base:
595
+ dim: 768
596
+ is_external: true
597
+ is_proprietary: false
598
+ is_sentence_transformers_compatible: true
599
+ link: https://huggingface.co/NbAiLab/nb-bert-base
600
+ seq_len: 512
601
+ size: 179
602
+ nb-bert-large:
603
+ dim: 1024
604
+ is_external: true
605
+ is_proprietary: false
606
+ is_sentence_transformers_compatible: true
607
+ link: https://huggingface.co/NbAiLab/nb-bert-large
608
+ seq_len: 512
609
+ size: 355
610
+ nomic-embed-text-v1.5-128:
611
+ dim: 128
612
+ is_external: true
613
+ is_proprietary: false
614
+ is_sentence_transformers_compatible: true
615
+ link: https://huggingface.co/nomic-ai/nomic-embed-text-v1.5
616
+ seq_len: 8192
617
+ size: 138
618
+ nomic-embed-text-v1.5-256:
619
+ dim: 256
620
+ is_external: true
621
+ is_proprietary: false
622
+ is_sentence_transformers_compatible: true
623
+ link: https://huggingface.co/nomic-ai/nomic-embed-text-v1.5
624
+ seq_len: 8192
625
+ size: 138
626
+ nomic-embed-text-v1.5-512:
627
+ dim: 512
628
+ is_external: true
629
+ is_proprietary: false
630
+ is_sentence_transformers_compatible: true
631
+ link: https://huggingface.co/nomic-ai/nomic-embed-text-v1.5
632
+ seq_len: 8192
633
+ size: 138
634
+ nomic-embed-text-v1.5-64:
635
+ dim: 64
636
+ is_external: true
637
+ is_proprietary: false
638
+ is_sentence_transformers_compatible: true
639
+ link: https://huggingface.co/nomic-ai/nomic-embed-text-v1.5
640
+ seq_len: 8192
641
+ size: 138
642
+ norbert3-base:
643
+ dim: 768
644
+ is_external: true
645
+ is_proprietary: false
646
+ is_sentence_transformers_compatible: true
647
+ link: https://huggingface.co/ltg/norbert3-base
648
+ seq_len: 512
649
+ size: 131
650
+ norbert3-large:
651
+ dim: 1024
652
+ is_external: true
653
+ is_proprietary: false
654
+ is_sentence_transformers_compatible: true
655
+ link: https://huggingface.co/ltg/norbert3-large
656
+ seq_len: 512
657
+ size: 368
658
+ paraphrase-multilingual-MiniLM-L12-v2:
659
+ dim: 384
660
+ is_external: true
661
+ is_proprietary: false
662
+ is_sentence_transformers_compatible: true
663
+ link: https://huggingface.co/sentence-transformers/paraphrase-multilingual-MiniLM-L12-v2
664
+ seq_len: 512
665
+ size: 118
666
+ paraphrase-multilingual-mpnet-base-v2:
667
+ dim: 768
668
+ is_external: true
669
+ is_proprietary: false
670
+ is_sentence_transformers_compatible: true
671
+ link: https://huggingface.co/sentence-transformers/paraphrase-multilingual-mpnet-base-v2
672
+ seq_len: 514
673
+ size: 278
674
+ sentence-bert-swedish-cased:
675
+ dim: 768
676
+ is_external: true
677
+ is_proprietary: false
678
+ is_sentence_transformers_compatible: true
679
+ link: https://huggingface.co/KBLab/sentence-bert-swedish-cased
680
+ seq_len: 512
681
+ size: 125
682
+ sentence-camembert-base:
683
+ dim: 768
684
+ is_external: true
685
+ is_proprietary: false
686
+ is_sentence_transformers_compatible: true
687
+ link: https://huggingface.co/dangvantuan/sentence-camembert-base
688
+ seq_len: 512
689
+ size: 110
690
+ sentence-camembert-large:
691
+ dim: 1024
692
+ is_external: true
693
+ is_proprietary: false
694
+ is_sentence_transformers_compatible: true
695
+ link: https://huggingface.co/dangvantuan/sentence-camembert-large
696
+ seq_len: 512
697
+ size: 337
698
+ sentence-croissant-llm-base:
699
+ dim: 2048
700
+ is_external: true
701
+ is_proprietary: false
702
+ is_sentence_transformers_compatible: true
703
+ link: https://huggingface.co/Wissam42/sentence-croissant-llm-base
704
+ seq_len: 2048
705
+ size: 1280
706
+ sentence-t5-base:
707
+ dim: 768
708
+ is_external: true
709
+ is_proprietary: false
710
+ is_sentence_transformers_compatible: true
711
+ link: https://huggingface.co/sentence-transformers/sentence-t5-base
712
+ seq_len: 512
713
+ size: 110
714
+ sentence-t5-large:
715
+ dim: 768
716
+ is_external: true
717
+ is_proprietary: false
718
+ is_sentence_transformers_compatible: true
719
+ link: https://huggingface.co/sentence-transformers/sentence-t5-large
720
+ seq_len: 512
721
+ size: 168
722
+ sentence-t5-xl:
723
+ dim: 768
724
+ is_external: true
725
+ is_proprietary: false
726
+ is_sentence_transformers_compatible: true
727
+ link: https://huggingface.co/sentence-transformers/sentence-t5-xl
728
+ seq_len: 512
729
+ size: 1240
730
+ sentence-t5-xxl:
731
+ dim: 768
732
+ is_external: true
733
+ is_proprietary: false
734
+ is_sentence_transformers_compatible: true
735
+ link: https://huggingface.co/sentence-transformers/sentence-t5-xxl
736
+ seq_len: 512
737
+ size: 4865
738
+ silver-retriever-base-v1:
739
+ dim: 768
740
+ is_external: true
741
+ is_proprietary: false
742
+ is_sentence_transformers_compatible: true
743
+ link: https://huggingface.co/ipipan/silver-retriever-base-v1
744
+ seq_len: 514
745
+ size: 125
746
+ st-polish-paraphrase-from-distilroberta:
747
+ dim: 768
748
+ is_external: true
749
+ is_proprietary: false
750
+ is_sentence_transformers_compatible: true
751
+ link: https://huggingface.co/sdadas/st-polish-paraphrase-from-distilroberta
752
+ seq_len: 514
753
+ size: 125
754
+ st-polish-paraphrase-from-mpnet:
755
+ dim: 768
756
+ is_external: true
757
+ is_proprietary: false
758
+ is_sentence_transformers_compatible: true
759
+ link: https://huggingface.co/sdadas/st-polish-paraphrase-from-mpnet
760
+ seq_len: 514
761
+ size: 125
762
+ sup-simcse-bert-base-uncased:
763
+ dim: 768
764
+ is_external: true
765
+ is_proprietary: false
766
+ is_sentence_transformers_compatible: true
767
+ link: https://huggingface.co/princeton-nlp/sup-simcse-bert-base-uncased
768
+ seq_len: 512
769
+ size: 110
770
+ text-embedding-3-large:
771
+ dim: 3072
772
+ is_external: true
773
+ is_proprietary: true
774
+ is_sentence_transformers_compatible: false
775
+ link: https://openai.com/blog/new-embedding-models-and-api-updates
776
+ seq_len: 8191
777
+ size: null
778
+ text-embedding-3-large-256:
779
+ dim: 256
780
+ is_external: true
781
+ is_proprietary: true
782
+ is_sentence_transformers_compatible: false
783
+ link: https://openai.com/blog/new-embedding-models-and-api-updates
784
+ seq_len: 8191
785
+ size: null
786
+ text-embedding-3-small:
787
+ dim: 1536
788
+ is_external: true
789
+ is_proprietary: true
790
+ is_sentence_transformers_compatible: false
791
+ link: https://openai.com/blog/new-embedding-models-and-api-updates
792
+ seq_len: 8191
793
+ size: null
794
+ text-embedding-ada-002:
795
+ dim: 1536
796
+ is_external: true
797
+ is_proprietary: true
798
+ is_sentence_transformers_compatible: false
799
+ link: https://openai.com/blog/new-and-improved-embedding-model
800
+ seq_len: 8191
801
+ size: null
802
+ text-search-ada-001:
803
+ dim: 1024
804
+ is_external: true
805
+ is_proprietary: true
806
+ is_sentence_transformers_compatible: false
807
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
808
+ seq_len: 2046
809
+ size: null
810
+ text-search-ada-doc-001:
811
+ dim: 1024
812
+ is_external: true
813
+ is_proprietary: true
814
+ is_sentence_transformers_compatible: false
815
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
816
+ seq_len: 2046
817
+ size: null
818
+ text-search-ada-query-001:
819
+ dim: 1024
820
+ is_external: false
821
+ is_proprietary: true
822
+ is_sentence_transformers_compatible: false
823
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
824
+ seq_len: 2046
825
+ size: null
826
+ text-search-babbage-001:
827
+ dim: 2048
828
+ is_external: true
829
+ is_proprietary: true
830
+ is_sentence_transformers_compatible: false
831
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
832
+ seq_len: 2046
833
+ size: null
834
+ text-search-curie-001:
835
+ dim: 4096
836
+ is_external: true
837
+ is_proprietary: true
838
+ is_sentence_transformers_compatible: false
839
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
840
+ seq_len: 2046
841
+ size: null
842
+ text-search-davinci-001:
843
+ dim: 12288
844
+ is_external: true
845
+ is_proprietary: true
846
+ is_sentence_transformers_compatible: false
847
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
848
+ seq_len: 2046
849
+ size: null
850
+ text-similarity-ada-001:
851
+ dim: 1024
852
+ is_external: true
853
+ is_proprietary: true
854
+ is_sentence_transformers_compatible: false
855
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
856
+ seq_len: 2046
857
+ size: null
858
+ text-similarity-babbage-001:
859
+ dim: 2048
860
+ is_external: true
861
+ is_proprietary: true
862
+ is_sentence_transformers_compatible: false
863
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
864
+ seq_len: 2046
865
+ size: null
866
+ text-similarity-curie-001:
867
+ dim: 4096
868
+ is_external: true
869
+ is_proprietary: true
870
+ is_sentence_transformers_compatible: false
871
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
872
+ seq_len: 2046
873
+ size: null
874
+ text-similarity-davinci-001:
875
+ dim: 12288
876
+ is_external: true
877
+ is_proprietary: true
878
+ is_sentence_transformers_compatible: false
879
+ link: https://openai.com/blog/introducing-text-and-code-embeddings
880
+ seq_len: 2046
881
+ size: null
882
+ text2vec-base-chinese:
883
+ dim: 768
884
+ is_external: true
885
+ is_proprietary: false
886
+ is_sentence_transformers_compatible: true
887
+ link: https://huggingface.co/shibing624/text2vec-base-chinese
888
+ seq_len: 512
889
+ size: 102
890
+ text2vec-base-multilingual:
891
+ dim: null
892
+ is_external: true
893
+ is_proprietary: false
894
+ is_sentence_transformers_compatible: false
895
+ link: null
896
+ seq_len: null
897
+ size: null
898
+ text2vec-large-chinese:
899
+ dim: 1024
900
+ is_external: true
901
+ is_proprietary: false
902
+ is_sentence_transformers_compatible: true
903
+ link: https://huggingface.co/GanymedeNil/text2vec-large-chinese
904
+ seq_len: 512
905
+ size: 326
906
+ titan-embed-text-v1:
907
+ dim: 1536
908
+ is_external: true
909
+ is_proprietary: true
910
+ is_sentence_transformers_compatible: false
911
+ link: https://docs.aws.amazon.com/bedrock/latest/userguide/embeddings.html
912
+ seq_len: 8000
913
+ size: null
914
+ udever-bloom-1b1:
915
+ dim: 1536
916
+ is_external: true
917
+ is_proprietary: false
918
+ is_sentence_transformers_compatible: true
919
+ link: https://huggingface.co/izhx/udever-bloom-1b1
920
+ seq_len: 2048
921
+ size: null
922
+ udever-bloom-560m:
923
+ dim: 1024
924
+ is_external: true
925
+ is_proprietary: false
926
+ is_sentence_transformers_compatible: true
927
+ link: https://huggingface.co/izhx/udever-bloom-560m
928
+ seq_len: 2048
929
+ size: null
930
+ universal-sentence-encoder-multilingual-3:
931
+ dim: 512
932
+ is_external: true
933
+ is_proprietary: false
934
+ is_sentence_transformers_compatible: true
935
+ link: https://huggingface.co/vprelovac/universal-sentence-encoder-multilingual-3
936
+ seq_len: 512
937
+ size: null
938
+ universal-sentence-encoder-multilingual-large-3:
939
+ dim: 512
940
+ is_external: true
941
+ is_proprietary: false
942
+ is_sentence_transformers_compatible: true
943
+ link: https://huggingface.co/vprelovac/universal-sentence-encoder-multilingual-large-3
944
+ seq_len: 512
945
+ size: null
946
+ unsup-simcse-bert-base-uncased:
947
+ dim: 768
948
+ is_external: true
949
+ is_proprietary: false
950
+ is_sentence_transformers_compatible: true
951
+ link: https://huggingface.co/princeton-nlp/unsup-simcse-bert-base-uncased
952
+ seq_len: 512
953
+ size: 110
954
+ use-cmlm-multilingual:
955
+ dim: 768
956
+ is_external: true
957
+ is_proprietary: false
958
+ is_sentence_transformers_compatible: true
959
+ link: https://huggingface.co/sentence-transformers/use-cmlm-multilingual
960
+ seq_len: 512
961
+ size: 472
962
+ voyage-2:
963
+ dim: 1024
964
+ is_external: true
965
+ is_proprietary: true
966
+ is_sentence_transformers_compatible: false
967
+ link: https://docs.voyageai.com/embeddings/
968
+ seq_len: 1024
969
+ size: null
970
+ voyage-code-2:
971
+ dim: 1536
972
+ is_external: true
973
+ is_proprietary: true
974
+ is_sentence_transformers_compatible: false
975
+ link: https://docs.voyageai.com/embeddings/
976
+ seq_len: 16000
977
+ size: null
978
+ voyage-large-2-instruct:
979
+ dim: 1024
980
+ is_external: true
981
+ is_proprietary: false
982
+ is_sentence_transformers_compatible: false
983
+ link: https://docs.voyageai.com/embeddings/
984
+ seq_len: 16000
985
+ size: null
986
+ voyage-law-2:
987
+ dim: 1024
988
+ is_external: true
989
+ is_proprietary: true
990
+ is_sentence_transformers_compatible: false
991
+ link: https://docs.voyageai.com/embeddings/
992
+ seq_len: 4000
993
+ size: null
994
+ voyage-lite-01-instruct:
995
+ dim: 1024
996
+ is_external: true
997
+ is_proprietary: true
998
+ is_sentence_transformers_compatible: false
999
+ link: https://docs.voyageai.com/embeddings/
1000
+ seq_len: 4000
1001
+ size: null
1002
+ voyage-lite-02-instruct:
1003
+ dim: 1024
1004
+ is_external: true
1005
+ is_proprietary: true
1006
+ is_sentence_transformers_compatible: false
1007
+ link: https://docs.voyageai.com/embeddings/
1008
+ seq_len: 4000
1009
+ size: 1220
1010
+ xlm-roberta-base:
1011
+ dim: 768
1012
+ is_external: true
1013
+ is_proprietary: false
1014
+ is_sentence_transformers_compatible: true
1015
+ link: https://huggingface.co/xlm-roberta-base
1016
+ seq_len: 514
1017
+ size: 279
1018
+ xlm-roberta-large:
1019
+ dim: 1024
1020
+ is_external: true
1021
+ is_proprietary: false
1022
+ is_sentence_transformers_compatible: true
1023
+ link: https://huggingface.co/xlm-roberta-large
1024
+ seq_len: 514
1025
+ size: 560
1026
+ models_to_skip:
1027
+ - michaelfeil/ct2fast-e5-large-v2
1028
+ - McGill-NLP/LLM2Vec-Sheared-LLaMA-mntp-unsup-simcse
1029
+ - newsrx/instructor-xl
1030
+ - sionic-ai/sionic-ai-v1
1031
+ - lsf1000/bge-evaluation
1032
+ - Intel/bge-small-en-v1.5-sst2
1033
+ - newsrx/instructor-xl-newsrx
1034
+ - McGill-NLP/LLM2Vec-Mistral-7B-Instruct-v2-mntp-unsup-simcse
1035
+ - davidpeer/gte-small
1036
+ - goldenrooster/multilingual-e5-large
1037
+ - kozistr/fused-large-en
1038
+ - mixamrepijey/instructor-small
1039
+ - McGill-NLP/LLM2Vec-Llama-2-7b-chat-hf-mntp-supervised
1040
+ - DecisionOptimizationSystem/DeepFeatEmbeddingLargeContext
1041
+ - Intel/bge-base-en-v1.5-sst2-int8-dynamic
1042
+ - morgendigital/multilingual-e5-large-quantized
1043
+ - BAAI/bge-small-en
1044
+ - ggrn/e5-small-v2
1045
+ - vectoriseai/gte-small
1046
+ - giulio98/placeholder
1047
+ - odunola/UAE-Large-VI
1048
+ - vectoriseai/e5-large-v2
1049
+ - gruber/e5-small-v2-ggml
1050
+ - Severian/nomic
1051
+ - arcdev/e5-mistral-7b-instruct
1052
+ - mlx-community/multilingual-e5-base-mlx
1053
+ - michaelfeil/ct2fast-bge-base-en-v1.5
1054
+ - Intel/bge-small-en-v1.5-sst2-int8-static
1055
+ - jncraton/stella-base-en-v2-ct2-int8
1056
+ - vectoriseai/multilingual-e5-large
1057
+ - rlsChapters/Chapters-SFR-Embedding-Mistral
1058
+ - arcdev/SFR-Embedding-Mistral
1059
+ - McGill-NLP/LLM2Vec-Mistral-7B-Instruct-v2-mntp-supervised
1060
+ - vectoriseai/gte-base
1061
+ - mixamrepijey/instructor-models
1062
+ - GovCompete/e5-large-v2
1063
+ - ef-zulla/e5-multi-sml-torch
1064
+ - khoa-klaytn/bge-small-en-v1.5-angle
1065
+ - krilecy/e5-mistral-7b-instruct
1066
+ - vectoriseai/bge-base-en-v1.5
1067
+ - vectoriseai/instructor-base
1068
+ - jingyeom/korean_embedding_model
1069
+ - rizki/bgr-tf
1070
+ - barisaydin/bge-base-en
1071
+ - jamesgpt1/zzz
1072
+ - Malmuk1/e5-large-v2_Sharded
1073
+ - vectoriseai/ember-v1
1074
+ - Consensus/instructor-base
1075
+ - barisaydin/bge-small-en
1076
+ - barisaydin/gte-base
1077
+ - woody72/multilingual-e5-base
1078
+ - Einas/einas_ashkar
1079
+ - michaelfeil/ct2fast-bge-large-en-v1.5
1080
+ - vectoriseai/bge-small-en-v1.5
1081
+ - iampanda/Test
1082
+ - cherubhao/yogamodel
1083
+ - ieasybooks/multilingual-e5-large-onnx
1084
+ - jncraton/e5-small-v2-ct2-int8
1085
+ - radames/e5-large
1086
+ - khoa-klaytn/bge-base-en-v1.5-angle
1087
+ - Intel/bge-base-en-v1.5-sst2-int8-static
1088
+ - vectoriseai/e5-large
1089
+ - TitanML/jina-v2-base-en-embed
1090
+ - Koat/gte-tiny
1091
+ - binqiangliu/EmbeddingModlebgelargeENv1.5
1092
+ - beademiguelperez/sentence-transformers-multilingual-e5-small
1093
+ - sionic-ai/sionic-ai-v2
1094
+ - jamesdborin/jina-v2-base-en-embed
1095
+ - maiyad/multilingual-e5-small
1096
+ - dmlls/all-mpnet-base-v2
1097
+ - odunola/e5-base-v2
1098
+ - vectoriseai/bge-large-en-v1.5
1099
+ - vectoriseai/bge-small-en
1100
+ - karrar-alwaili/UAE-Large-V1
1101
+ - t12e/instructor-base
1102
+ - Frazic/udever-bloom-3b-sentence
1103
+ - Geolumina/instructor-xl
1104
+ - hsikchi/dump
1105
+ - recipe/embeddings
1106
+ - michaelfeil/ct2fast-bge-small-en-v1.5
1107
+ - ildodeltaRule/multilingual-e5-large
1108
+ - shubham-bgi/UAE-Large
1109
+ - BAAI/bge-large-en
1110
+ - michaelfeil/ct2fast-e5-small-v2
1111
+ - cgldo/semanticClone
1112
+ - barisaydin/gte-small
1113
+ - aident-ai/bge-base-en-onnx
1114
+ - jamesgpt1/english-large-v1
1115
+ - michaelfeil/ct2fast-e5-small
1116
+ - baseplate/instructor-large-1
1117
+ - newsrx/instructor-large
1118
+ - Narsil/bge-base-en
1119
+ - michaelfeil/ct2fast-e5-large
1120
+ - mlx-community/multilingual-e5-small-mlx
1121
+ - lightbird-ai/nomic
1122
+ - MaziyarPanahi/GritLM-8x7B-GGUF
1123
+ - newsrx/instructor-large-newsrx
1124
+ - dhairya0907/thenlper-get-large
1125
+ - barisaydin/bge-large-en
1126
+ - jncraton/bge-small-en-ct2-int8
1127
+ - retrainai/instructor-xl
1128
+ - BAAI/bge-base-en
1129
+ - gentlebowl/instructor-large-safetensors
1130
+ - d0rj/e5-large-en-ru
1131
+ - atian-chapters/Chapters-SFR-Embedding-Mistral
1132
+ - Intel/bge-base-en-v1.5-sts-int8-static
1133
+ - Intel/bge-base-en-v1.5-sts-int8-dynamic
1134
+ - jncraton/GIST-small-Embedding-v0-ct2-int8
1135
+ - jncraton/gte-tiny-ct2-int8
1136
+ - d0rj/e5-small-en-ru
1137
+ - vectoriseai/e5-small-v2
1138
+ - SmartComponents/bge-micro-v2
1139
+ - michaelfeil/ct2fast-gte-base
1140
+ - vectoriseai/e5-base-v2
1141
+ - Intel/bge-base-en-v1.5-sst2
1142
+ - McGill-NLP/LLM2Vec-Sheared-LLaMA-mntp-supervised
1143
+ - Research2NLP/electrical_stella
1144
+ - weakit-v/bge-base-en-v1.5-onnx
1145
+ - GovCompete/instructor-xl
1146
+ - barisaydin/text2vec-base-multilingual
1147
+ - Intel/bge-small-en-v1.5-sst2-int8-dynamic
1148
+ - jncraton/gte-small-ct2-int8
1149
+ - d0rj/e5-base-en-ru
1150
+ - barisaydin/gte-large
1151
+ - fresha/e5-large-v2-endpoint
1152
+ - vectoriseai/instructor-large
1153
+ - Severian/embed
1154
+ - vectoriseai/e5-base
1155
+ - mlx-community/multilingual-e5-large-mlx
1156
+ - vectoriseai/gte-large
1157
+ - anttip/ct2fast-e5-small-v2-hfie
1158
+ - michaelfeil/ct2fast-gte-large
1159
+ - gizmo-ai/Cohere-embed-multilingual-v3.0
1160
+ - McGill-NLP/LLM2Vec-Llama-2-7b-chat-hf-mntp-unsup-simcse