Upload folder using huggingface_hub
Browse filesThis view is limited to 50 files because it contains too many changes.
See raw diff
- .gitattributes +14 -0
- BAAI_bge-base-en-v1.5_C-MTEB_CMedQAv1-reranking.json +0 -0
- BAAI_bge-base-en-v1.5_C-MTEB_Mmarco-reranking.json +0 -0
- BAAI_bge-base-en-v1.5_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
- BAAI_bge-base-en-v1.5_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
- BAAI_bge-base-en-v1.5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
- BAAI_bge-base-en-v1.5_mteb_askubuntudupquestions-reranking.json +0 -0
- BAAI_bge-base-en-v1.5_mteb_scidocs-reranking.json +0 -0
- BAAI_bge-base-en-v1.5_mteb_stackoverflowdupquestions-reranking.json +0 -0
- BAAI_bge-base-en_C-MTEB_Mmarco-reranking.json +0 -0
- BAAI_bge-reranker-base_C-MTEB_Mmarco-reranking.json +0 -0
- BAAI_bge-reranker-large_C-MTEB_Mmarco-reranking.json +0 -0
- BAAI_llm-embedder_C-MTEB_CMedQAv1-reranking.json +0 -0
- BAAI_llm-embedder_C-MTEB_Mmarco-reranking.json +0 -0
- BAAI_llm-embedder_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
- BAAI_llm-embedder_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
- BAAI_llm-embedder_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
- BAAI_llm-embedder_mteb_askubuntudupquestions-reranking.json +0 -0
- BAAI_llm-embedder_mteb_scidocs-reranking.json +0 -0
- BAAI_llm-embedder_mteb_stackoverflowdupquestions-reranking.json +0 -0
- all-MiniLM-L6-v2_C-MTEB_CMedQAv1-reranking.json +0 -0
- all-MiniLM-L6-v2_C-MTEB_Mmarco-reranking.json +0 -0
- all-MiniLM-L6-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
- all-MiniLM-L6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
- all-MiniLM-L6-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
- all-MiniLM-L6-v2_mteb_askubuntudupquestions-reranking.json +0 -0
- all-MiniLM-L6-v2_mteb_scidocs-reranking.json +0 -0
- all-MiniLM-L6-v2_mteb_stackoverflowdupquestions-reranking.json +0 -0
- all-distilroberta-v1_C-MTEB_CMedQAv1-reranking.json +0 -0
- all-distilroberta-v1_C-MTEB_Mmarco-reranking.json +0 -0
- all-distilroberta-v1_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
- all-distilroberta-v1_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
- all-distilroberta-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
- all-distilroberta-v1_mteb_askubuntudupquestions-reranking.json +0 -0
- all-distilroberta-v1_mteb_scidocs-reranking.json +0 -0
- all-distilroberta-v1_mteb_stackoverflowdupquestions-reranking.json +0 -0
- all-mpnet-base-v2_C-MTEB_CMedQAv1-reranking.json +0 -0
- all-mpnet-base-v2_C-MTEB_Mmarco-reranking.json +0 -0
- all-mpnet-base-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
- all-mpnet-base-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
- all-mpnet-base-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
- all-mpnet-base-v2_mteb_askubuntudupquestions-reranking.json +0 -0
- all-mpnet-base-v2_mteb_scidocs-reranking.json +0 -0
- all-mpnet-base-v2_mteb_stackoverflowdupquestions-reranking.json +0 -0
- cross-encoder_ms-marco-MiniLM-L-6-v2_C-MTEB_Mmarco-reranking.json +0 -0
- cross-encoder_ms-marco-MiniLM-L-6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
- cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_askubuntudupquestions-reranking.json +0 -0
- cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_scidocs-reranking.json +0 -0
- cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_stackoverflowdupquestions-reranking.json +0 -0
- cross-encoder_ms-marco-TinyBERT-L-2-v2_C-MTEB_Mmarco-reranking.json +0 -0
.gitattributes
CHANGED
@@ -53,3 +53,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
53 |
*.jpg filter=lfs diff=lfs merge=lfs -text
|
54 |
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
55 |
*.webp filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
53 |
*.jpg filter=lfs diff=lfs merge=lfs -text
|
54 |
*.jpeg filter=lfs diff=lfs merge=lfs -text
|
55 |
*.webp filter=lfs diff=lfs merge=lfs -text
|
56 |
+
BAAI_bge-base-en-v1.5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
57 |
+
BAAI_llm-embedder_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
58 |
+
all-MiniLM-L6-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
59 |
+
all-distilroberta-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
60 |
+
all-mpnet-base-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
61 |
+
intfloat_e5-large-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
62 |
+
intfloat_e5-small-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
63 |
+
intfloat_multilingual-e5-large_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
64 |
+
intfloat_multilingual-e5-small_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
65 |
+
llmrails_ember-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
66 |
+
msmarco-MiniLM-L6-cos-v5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
67 |
+
msmarco-distilbert-dot-v5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
68 |
+
multi-qa-MiniLM-L6-cos-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
69 |
+
multi-qa-distilbert-cos-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
|
BAAI_bge-base-en-v1.5_C-MTEB_CMedQAv1-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en-v1.5_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en-v1.5_OrdalieTech_MIRACL-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en-v1.5_OrdalieTech_Ordalie-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en-v1.5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c0d110dab1ae25af6c2d5d71e52ed40d863fc2c8d2936d59e98eedd8b84cffb2
|
3 |
+
size 101681226
|
BAAI_bge-base-en-v1.5_mteb_askubuntudupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en-v1.5_mteb_scidocs-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en-v1.5_mteb_stackoverflowdupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-base-en_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-reranker-base_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_bge-reranker-large_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_C-MTEB_CMedQAv1-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_OrdalieTech_MIRACL-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_OrdalieTech_Ordalie-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:5ed68223fbbc099f43018c1e506f6abfb4126665e914bc739cd7b3c0f2a72757
|
3 |
+
size 101680340
|
BAAI_llm-embedder_mteb_askubuntudupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_mteb_scidocs-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
BAAI_llm-embedder_mteb_stackoverflowdupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_C-MTEB_CMedQAv1-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:46e0e829e2232d8fc09078d396dec133864f65ff848b6c4f8ad9dd51d7b35371
|
3 |
+
size 101656380
|
all-MiniLM-L6-v2_mteb_askubuntudupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_mteb_scidocs-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-MiniLM-L6-v2_mteb_stackoverflowdupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_C-MTEB_CMedQAv1-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_OrdalieTech_MIRACL-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_OrdalieTech_Ordalie-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8bc6b0668983bfd7384c9e30ad5a8f6eacf39b3cf2f346b4a34ddc625fc69afc
|
3 |
+
size 101994634
|
all-distilroberta-v1_mteb_askubuntudupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_mteb_scidocs-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-distilroberta-v1_mteb_stackoverflowdupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_C-MTEB_CMedQAv1-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:4ecb6464b2d92995c88759c26164eb5487501b2067a7befbe4f99c49bd7793e6
|
3 |
+
size 101689889
|
all-mpnet-base-v2_mteb_askubuntudupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_mteb_scidocs-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
all-mpnet-base-v2_mteb_stackoverflowdupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
cross-encoder_ms-marco-MiniLM-L-6-v2_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
cross-encoder_ms-marco-MiniLM-L-6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_askubuntudupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_scidocs-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_stackoverflowdupquestions-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
cross-encoder_ms-marco-TinyBERT-L-2-v2_C-MTEB_Mmarco-reranking.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|