--- datasets: - EleutherAI/pile language: - en --- # Model Card This model is an Attention (Llama architecture) model pretrained on 30Bn tokens of the Pile corpus. ### Model Sources The model implementation and training code that produced the model are provided here: https://github.com/HazyResearch/based