gpt2-wikitext103 / README.md
himanshubeniwal's picture
Update README.md
f87c51b
|
raw
history blame
422 Bytes
---
datasets:
- wikitext
language:
- en
metrics:
- perplexity
---
## Model Details
GPT-2 Pretrained on Wikitext-103 (180M sentences) on 32GB V100 GPU for around 1.10L iterations.
Val_loss vs train_loss:
![Loss curve](https://huggingface.co/himanshubeniwal/gpt2-wikitext103/resolve/main/sample.png)
### Model Description
Perplexity: 22.87
### Out-of-Scope Use
Just a test model. Please don't expect good results.