manu commited on
Commit
d1ab24f
1 Parent(s): 4c0cbf3

Upload folder using huggingface_hub

Browse files
This view is limited to 50 files because it contains too many changes.   See raw diff
Files changed (50) hide show
  1. .gitattributes +14 -0
  2. BAAI_bge-base-en-v1.5_C-MTEB_CMedQAv1-reranking.json +0 -0
  3. BAAI_bge-base-en-v1.5_C-MTEB_Mmarco-reranking.json +0 -0
  4. BAAI_bge-base-en-v1.5_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
  5. BAAI_bge-base-en-v1.5_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
  6. BAAI_bge-base-en-v1.5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
  7. BAAI_bge-base-en-v1.5_mteb_askubuntudupquestions-reranking.json +0 -0
  8. BAAI_bge-base-en-v1.5_mteb_scidocs-reranking.json +0 -0
  9. BAAI_bge-base-en-v1.5_mteb_stackoverflowdupquestions-reranking.json +0 -0
  10. BAAI_bge-base-en_C-MTEB_Mmarco-reranking.json +0 -0
  11. BAAI_bge-reranker-base_C-MTEB_Mmarco-reranking.json +0 -0
  12. BAAI_bge-reranker-large_C-MTEB_Mmarco-reranking.json +0 -0
  13. BAAI_llm-embedder_C-MTEB_CMedQAv1-reranking.json +0 -0
  14. BAAI_llm-embedder_C-MTEB_Mmarco-reranking.json +0 -0
  15. BAAI_llm-embedder_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
  16. BAAI_llm-embedder_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
  17. BAAI_llm-embedder_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
  18. BAAI_llm-embedder_mteb_askubuntudupquestions-reranking.json +0 -0
  19. BAAI_llm-embedder_mteb_scidocs-reranking.json +0 -0
  20. BAAI_llm-embedder_mteb_stackoverflowdupquestions-reranking.json +0 -0
  21. all-MiniLM-L6-v2_C-MTEB_CMedQAv1-reranking.json +0 -0
  22. all-MiniLM-L6-v2_C-MTEB_Mmarco-reranking.json +0 -0
  23. all-MiniLM-L6-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
  24. all-MiniLM-L6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
  25. all-MiniLM-L6-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
  26. all-MiniLM-L6-v2_mteb_askubuntudupquestions-reranking.json +0 -0
  27. all-MiniLM-L6-v2_mteb_scidocs-reranking.json +0 -0
  28. all-MiniLM-L6-v2_mteb_stackoverflowdupquestions-reranking.json +0 -0
  29. all-distilroberta-v1_C-MTEB_CMedQAv1-reranking.json +0 -0
  30. all-distilroberta-v1_C-MTEB_Mmarco-reranking.json +0 -0
  31. all-distilroberta-v1_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
  32. all-distilroberta-v1_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
  33. all-distilroberta-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
  34. all-distilroberta-v1_mteb_askubuntudupquestions-reranking.json +0 -0
  35. all-distilroberta-v1_mteb_scidocs-reranking.json +0 -0
  36. all-distilroberta-v1_mteb_stackoverflowdupquestions-reranking.json +0 -0
  37. all-mpnet-base-v2_C-MTEB_CMedQAv1-reranking.json +0 -0
  38. all-mpnet-base-v2_C-MTEB_Mmarco-reranking.json +0 -0
  39. all-mpnet-base-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json +0 -0
  40. all-mpnet-base-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
  41. all-mpnet-base-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json +3 -0
  42. all-mpnet-base-v2_mteb_askubuntudupquestions-reranking.json +0 -0
  43. all-mpnet-base-v2_mteb_scidocs-reranking.json +0 -0
  44. all-mpnet-base-v2_mteb_stackoverflowdupquestions-reranking.json +0 -0
  45. cross-encoder_ms-marco-MiniLM-L-6-v2_C-MTEB_Mmarco-reranking.json +0 -0
  46. cross-encoder_ms-marco-MiniLM-L-6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json +0 -0
  47. cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_askubuntudupquestions-reranking.json +0 -0
  48. cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_scidocs-reranking.json +0 -0
  49. cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_stackoverflowdupquestions-reranking.json +0 -0
  50. cross-encoder_ms-marco-TinyBERT-L-2-v2_C-MTEB_Mmarco-reranking.json +0 -0
.gitattributes CHANGED
@@ -53,3 +53,17 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
53
  *.jpg filter=lfs diff=lfs merge=lfs -text
54
  *.jpeg filter=lfs diff=lfs merge=lfs -text
55
  *.webp filter=lfs diff=lfs merge=lfs -text
56
+ BAAI_bge-base-en-v1.5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
57
+ BAAI_llm-embedder_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
58
+ all-MiniLM-L6-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
59
+ all-distilroberta-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
60
+ all-mpnet-base-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
61
+ intfloat_e5-large-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
62
+ intfloat_e5-small-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
63
+ intfloat_multilingual-e5-large_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
64
+ intfloat_multilingual-e5-small_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
65
+ llmrails_ember-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
66
+ msmarco-MiniLM-L6-cos-v5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
67
+ msmarco-distilbert-dot-v5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
68
+ multi-qa-MiniLM-L6-cos-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
69
+ multi-qa-distilbert-cos-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json filter=lfs diff=lfs merge=lfs -text
BAAI_bge-base-en-v1.5_C-MTEB_CMedQAv1-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en-v1.5_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en-v1.5_OrdalieTech_MIRACL-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en-v1.5_OrdalieTech_Ordalie-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en-v1.5_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c0d110dab1ae25af6c2d5d71e52ed40d863fc2c8d2936d59e98eedd8b84cffb2
3
+ size 101681226
BAAI_bge-base-en-v1.5_mteb_askubuntudupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en-v1.5_mteb_scidocs-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en-v1.5_mteb_stackoverflowdupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-base-en_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-reranker-base_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_bge-reranker-large_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_C-MTEB_CMedQAv1-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_OrdalieTech_MIRACL-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_OrdalieTech_Ordalie-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5ed68223fbbc099f43018c1e506f6abfb4126665e914bc739cd7b3c0f2a72757
3
+ size 101680340
BAAI_llm-embedder_mteb_askubuntudupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_mteb_scidocs-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
BAAI_llm-embedder_mteb_stackoverflowdupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_C-MTEB_CMedQAv1-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:46e0e829e2232d8fc09078d396dec133864f65ff848b6c4f8ad9dd51d7b35371
3
+ size 101656380
all-MiniLM-L6-v2_mteb_askubuntudupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_mteb_scidocs-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-MiniLM-L6-v2_mteb_stackoverflowdupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_C-MTEB_CMedQAv1-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_OrdalieTech_MIRACL-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_OrdalieTech_Ordalie-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8bc6b0668983bfd7384c9e30ad5a8f6eacf39b3cf2f346b4a34ddc625fc69afc
3
+ size 101994634
all-distilroberta-v1_mteb_askubuntudupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_mteb_scidocs-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-distilroberta-v1_mteb_stackoverflowdupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_C-MTEB_CMedQAv1-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_OrdalieTech_MIRACL-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_lyon-nlp_mteb-fr-reranking-alloprof-s2p.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ecb6464b2d92995c88759c26164eb5487501b2067a7befbe4f99c49bd7793e6
3
+ size 101689889
all-mpnet-base-v2_mteb_askubuntudupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_mteb_scidocs-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
all-mpnet-base-v2_mteb_stackoverflowdupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
cross-encoder_ms-marco-MiniLM-L-6-v2_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
cross-encoder_ms-marco-MiniLM-L-6-v2_OrdalieTech_Ordalie-FR-Reranking-benchmark.json ADDED
The diff for this file is too large to render. See raw diff
 
cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_askubuntudupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_scidocs-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
cross-encoder_ms-marco-MiniLM-L-6-v2_mteb_stackoverflowdupquestions-reranking.json ADDED
The diff for this file is too large to render. See raw diff
 
cross-encoder_ms-marco-TinyBERT-L-2-v2_C-MTEB_Mmarco-reranking.json ADDED
The diff for this file is too large to render. See raw diff