Javanese GPT-2 Model

Pretrained GPT-2 small language model for Javanese.

Pretraining Corpus

gpt2-small-javanese-280M model was pretrained on 280MB of text. Below is list of data we trained on

  1. Javanese Wikipedia.
  2. Sastra.org archive.
  3. Common Crawl Corpus.
Downloads last month
16
Hosted inference API
Text Generation