|
--- |
|
tags: |
|
- bert |
|
license: cc-by-4.0 |
|
--- |
|
## bert-ascii-base |
|
is a BERT base Language Model pre-trained by predicting the summation of the **ASCII** code values of the characters in a masked token as a pre-training objective. For more details about the pre-training objective and the pre-training hyperparameters, please refer to [How does the pre-training objective affect what large language models learn about linguistic properties?](https://arxiv.org/abs/2203.10415) |
|
|
|
## License |
|
CC BY 4.0 |
|
|
|
## Citation |
|
If you use this model, please cite the following paper: |
|
``` |
|
@inproceedings{alajrami2022does, |
|
title={How does the pre-training objective affect what large language models learn about linguistic properties?}, |
|
author={Alajrami, Ahmed and Aletras, Nikolaos}, |
|
booktitle={Proceedings of the 60th Annual Meeting of the Association for Computational Linguistics (Volume 2: Short Papers)}, |
|
pages={131--147}, |
|
year={2022} |
|
} |
|
``` |