--- tags: - bert license: cc-by-4.0 --- ## bert-mlm-medium is a medium size BERT Language Model with an **MLM** pre-training objective. For more details about the pre-training objective and the pre-training hyperparameters, please refer to [How does the pre-training objective affect what large language models learn about linguistic properties?](https://aclanthology.org/2022.acl-short.16/) ## License CC BY 4.0 ## Citation If you use this model, please cite the following paper: ``` @inproceedings{alajrami2022does, title={How does the pre-training objective affect what large language models learn about linguistic properties?}, author={Alajrami, Ahmed and Aletras, Nikolaos}, booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)}, pages={131--147}, year={2022} } ```