File size: 740 Bytes
de89287
 
6ecb4b7
de89287
 
 
1
2
3
4
5
6
7
# RoBERTa base model for Hindi language

[Pretrained model on Hindi language](https://huggingface.co/flax-community/roberta-hindi)  using a masked language modeling (MLM) objective. Model is able to achieve competitive accuracy compared to pre-existing models on downstream tasks like NamedEntityRecognition and Classification. There are some MLM examples which show that there is a visible room for improvement, but this should serve well as a good base model for hindi languages & could be fine-tuned on specific datasets.

> This is part of the
[Flax/Jax Community Week](https://discuss.huggingface.co/t/pretrain-roberta-from-scratch-in-hindi/7091), organized by [HuggingFace](https://huggingface.co/) and TPU usage sponsored by Google.