gpt-2-bookcorpus / README.md
arun-AiBharat's picture
Update README.md
625dfcd verified
metadata
license: mit
datasets:
  - bookcorpus/bookcorpus
language:
  - en
library_name: transformers
  • The GPT -2 model was trained on the BookCorpus dataset for 60K steps.
  • No position embedding was used (NoPE).
  • Here is the wandb report
  • This is for educational purposes only.