File size: 987 Bytes
00c7e19 3a5f1db 00c7e19 da83f9d |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 |
This is a cross-encoder model with dot-product based scoring mechanism trained on MS-MARCO dataset.
The parameters of the cross-encoder are initialized using a 6-layer [minilm model](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased)
and is trained via distillation using scores from three different teacher models --
[model 1](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_teacher_1_albert),
[model 2](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_teacher_2_bert_base), and
[model 3](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_teacher_3_bert_large_wwm).
This model is used in experiments of our [EMNLP 2023](https://aclanthology.org/2023.findings-emnlp.544/) and [ICLR 2024](https://openreview.net/forum?id=1CPta0bfN2) papers.
See our EMNLP 2022 paper titled "Efficient Nearest Neighbor Search for Cross-Encoder Models using Matrix Factorization" for more details on the dot-product based scoring mechanism.
---
license: apache-2.0
---
|