This is a cross-encoder model with dot-product based scoring mechanism trained on MS-MARCO dataset. The parameters of the cross-encoder are initialized using [albert-large-v2](https://huggingface.co/albert/albert-base-v2). This model is used as a teacher model for training a [MiniLM-based cross-encoder model](https://huggingface.co/nishantyadav/emb_crossenc_msmarco_miniLM) which is used in experiments of our [EMNLP 2023](https://aclanthology.org/2023.findings-emnlp.544/) and [ICLR 2024](https://openreview.net/forum?id=1CPta0bfN2) papers. See our EMNLP 2022 paper titled "Efficient Nearest Neighbor Search for Cross-Encoder Models using Matrix Factorization" for more details on the dot-product based scoring mechanism. --- license: apache-2.0 ---