When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models Paper • 2406.07368 • Published Jun 11, 2024 • 2
ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization Paper • 2406.05981 • Published Jun 10, 2024 • 12
ShiftAddLLM: Accelerating Pretrained LLMs via Post-Training Multiplication-Less Reparameterization Paper • 2406.05981 • Published Jun 10, 2024 • 12
ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer Paper • 2306.06446 • Published Jun 10, 2023 • 1
NetDistiller: Empowering Tiny Deep Learning via In-Situ Distillation Paper • 2310.19820 • Published Oct 24, 2023 • 1