Collections
Discover the best community collections!
Collections trending this week
-
MambaByte: Token-free Selective State Space Model
Paper • 2401.13660 • Published • 49 -
Mamba: Linear-Time Sequence Modeling with Selective State Spaces
Paper • 2312.00752 • Published • 138 -
MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts
Paper • 2401.04081 • Published • 70 -
hustvl/Vim-tiny
Updated • 19
-
BiLLM: Pushing the Limit of Post-Training Quantization for LLMs
Paper • 2402.04291 • Published • 48 -
KVQuant: Towards 10 Million Context Length LLM Inference with KV Cache Quantization
Paper • 2401.18079 • Published • 7 -
Towards Next-Level Post-Training Quantization of Hyper-Scale Transformers
Paper • 2402.08958 • Published • 3 -
OneBit: Towards Extremely Low-bit Large Language Models
Paper • 2402.11295 • Published • 22