Edit model card

bert-sr-medium

A medium-size BERT Language Model with a shuffle + random pre-training objective. For more details about the pre-training objective and the pre-training hyperparameters, please refer to How does the pre-training objective affect what large language models learn about linguistic properties?

License

CC BY 4.0

Citation

If you use this model, please cite the following paper:

@inproceedings{alajrami2022does,
  title={How does the pre-training objective affect what large language models learn about linguistic properties?},
  author={Alajrami, Ahmed and Aletras, Nikolaos},
  booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)},
  pages={131--147},
  year={2022}
}
Downloads last month
11
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.