nano_tagalog_models / README.md
922CA's picture
Update README.md
4ff4712 verified
|
raw
history blame
297 Bytes
metadata
license: cc-by-sa-4.0
datasets:
  - facebook/belebele

Pretrained toy model. Made with Andrej Karpathy's NanoGPT, ~2023. Trained on part of Tagalog portion of Belebele.

Parameters:

  • batch_size = 64
  • block_size = 256
  • n_layer = 8
  • n_head = 8
  • n_embd = 768

Everything else is left as is.