setu4993 kevinmzy commited on
Commit
cf79e36
1 Parent(s): 80d6d9a

Update README.md (#1)

Browse files

- Update README.md (82f019f8783240968cbe8ea54a80ae6fd080e994)


Co-authored-by: Zhuoyuan Mao <kevinmzy@users.noreply.huggingface.co>

Files changed (1) hide show
  1. README.md +13 -7
README.md CHANGED
@@ -212,12 +212,18 @@ Details about data, training, evaluation and performance metrics are available i
212
  ### BibTeX entry and citation info
213
 
214
  ```bibtex
215
- @misc{mao2023lealla,
216
- title={LEALLA: Learning Lightweight Language-agnostic Sentence Embeddings with Knowledge Distillation},
217
- author={Zhuoyuan Mao and Tetsuji Nakagawa},
218
- year={2023},
219
- eprint={2302.08387},
220
- archivePrefix={arXiv},
221
- primaryClass={cs.CL}
 
 
 
 
 
 
222
  }
223
  ```
 
212
  ### BibTeX entry and citation info
213
 
214
  ```bibtex
215
+ @inproceedings{mao-nakagawa-2023-lealla,
216
+ title = "{LEALLA}: Learning Lightweight Language-agnostic Sentence Embeddings with Knowledge Distillation",
217
+ author = "Mao, Zhuoyuan and
218
+ Nakagawa, Tetsuji",
219
+ booktitle = "Proceedings of the 17th Conference of the European Chapter of the Association for Computational Linguistics",
220
+ month = may,
221
+ year = "2023",
222
+ address = "Dubrovnik, Croatia",
223
+ publisher = "Association for Computational Linguistics",
224
+ url = "https://aclanthology.org/2023.eacl-main.138",
225
+ doi = "10.18653/v1/2023.eacl-main.138",
226
+ pages = "1886--1894",
227
+ abstract = "Large-scale language-agnostic sentence embedding models such as LaBSE (Feng et al., 2022) obtain state-of-the-art performance for parallel sentence alignment. However, these large-scale models can suffer from inference speed and computation overhead. This study systematically explores learning language-agnostic sentence embeddings with lightweight models. We demonstrate that a thin-deep encoder can construct robust low-dimensional sentence embeddings for 109 languages. With our proposed distillation methods, we achieve further improvements by incorporating knowledge from a teacher model. Empirical results on Tatoeba, United Nations, and BUCC show the effectiveness of our lightweight models. We release our lightweight language-agnostic sentence embedding models LEALLA on TensorFlow Hub.",
228
  }
229
  ```