metadata
language:
- en
tags:
- text generation
- pytorch
- causal-lm
- gpt_neox
license: mit
datasets:
- hoskinson-center/proof-pile
ProofGPT-v0.1
Model Description
ProofGPT-v0.1 is a 6.7B parameter language model based on the GPT-NeoX architecture and trained on the proof-pile (v1.1). We initiailized training with pythia-6.7b weights, a precursor to the pythia-6.9b model that has roughly equivalent performance.
Detailed evaluations coming soon :)