![](https://cdn-avatars.huggingface.co/v1/production/uploads/60bccec062080d33f875cd0c/FDQwEj3IBUG_fa377Ej1U.jpeg)
BEE-spoke-data/smol_llama-101M-GQA
Text Generation
•
Updated
•
3.7k
•
23
small-scale pretraining experiments of mine
Note this is a mid-training checkpoint of what is now smol_llama-220M