roberta-base-als / README.md
patrickvonplaten's picture
Create README.md
fb1e3e0

This project pretrains a roberta-base on the Alemannic (als) data subset of the OSCAR corpus in JAX/Flax.

We will be using the masked-language modeling loss for pretraining.