Enabling High-Sparsity Foundational Llama Models with Efficient Pretraining and Deployment Paper • 2405.03594 • Published May 6 • 7
Uncovering mesa-optimization algorithms in Transformers Paper • 2309.05858 • Published Sep 11, 2023 • 12