--- datasets: - wikitext language: - en metrics: - perplexity --- ## Model Details GPT-2 Pretrained on Wikitext-103 (180M sentences) on 32GB V100 GPU for around 1.10L iterations. Val_loss vs train_loss: ![Loss curve](https://huggingface.co/himanshubeniwal/gpt2-wikitext103/resolve/main/sample.png) ### Model Description Perplexity: 22.87 ### Out-of-Scope Use Just a test model. Please don't expect good results.