Javanese GPT-2 Model
Pretrained GPT-2 small language model for Javanese.
Pretraining Corpus
gpt2-small-javanese-280M
model was pretrained on 280MB of text. Below is list of data we trained on
- Javanese Wikipedia.
- Sastra.org archive.
- Common Crawl Corpus.
- Downloads last month
- 16