gpt2wikipedia200 / README.md
raf6423's picture
Update README.md
09f4ead verified
metadata
library_name: transformers
tags: []

Model Details

  • Developed by: Rafael Espinosa Mena
  • Model type: GPT2 124M
  • Language(s) (NLP): English

Uses

Pre-Trained from scratch on 200 wikipedia articles and intended for fine tunning.

Training Details

Trained for 200 epochs on 200 wikipedia articles. Used a learning rate of 3e-5, and a sliding window approach with 1024 tokens per chunk, and a 200 token window.

Training Data

https://huggingface.co/datasets/wikipedia

Hardware

Trained on a single V100 GPU

Model Card Authors

Rafael Espinosa Mena rafaelespinosamena@gmail.com