File size: 987 Bytes
00c7e19
3a5f1db
 
 
 
 
00c7e19
 
 
 
 
da83f9d
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
This is a cross-encoder model with dot-product based scoring mechanism trained on MS-MARCO dataset. 
The parameters of the cross-encoder are initialized using a 6-layer [minilm model](https://huggingface.co/nreimers/MiniLM-L6-H384-uncased) 
and is trained via distillation using scores from three different teacher models -- 
[model 1](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_teacher_1_albert), 
[model 2](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_teacher_2_bert_base), and 
[model 3](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_teacher_3_bert_large_wwm). 
This model is used in experiments of our [EMNLP 2023](https://aclanthology.org/2023.findings-emnlp.544/) and [ICLR 2024](https://openreview.net/forum?id=1CPta0bfN2) papers.

See our EMNLP 2022 paper titled "Efficient Nearest Neighbor Search for Cross-Encoder Models using Matrix Factorization" for more details on the dot-product based scoring mechanism.


---
license: apache-2.0
---