--- license: apache-2.0 datasets: - xquad language: - multilingual library_name: transformers tags: - cross-lingual - exctractive-question-answering metrics: - f1 - exact_match --- # Description Best-performing "mBERT-qa-en, skd" model from the paper [Promoting Generalized Cross-lingual Question Answering in Few-resource Scenarios via Self-knowledge Distillation](https://arxiv.org/abs/2309.17134). Check the official [GitHub repository](https://github.com/ccasimiro88/self-distillation-gxlt-qa) to access the code used to implement the methods in the paper. **More info coming soon!** # How to Cite To cite our work use the following BibTex: ``` @misc{carrino2023promoting, title={Promoting Generalized Cross-lingual Question Answering in Few-resource Scenarios via Self-knowledge Distillation}, author={Casimiro Pio Carrino and Carlos Escolano and José A. R. Fonollosa}, year={2023}, eprint={2309.17134}, archivePrefix={arXiv}, primaryClass={cs.CL} } ```