--- license: mit --- A model trained using lilLM: https://github.com/CohleM/lilLM The model is ~300M parameters and trained on 9.5B tokens from OpenWebText: https://huggingface.co/datasets/Skylion007/openwebtext