metadata
license: mit
datasets:
- bookcorpus/bookcorpus
language:
- en
library_name: transformers
- The GPT -2 model was trained on the BookCorpus dataset for 60K steps.
- No position embedding was used (NoPE).
- Here is the wandb report
- This is for educational purposes only.