Mamba-Shedder
Collection
Compressed Mamba Models via Mamba-Shedder (Mamba-Shedder: Post-Transformer Compression for Efficient Selective Structured State Space Models)
•
7 items
•
Updated
git clone https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning.git
cd Mamba-Shedder
python eval.py --model_path <path to model>
Refer to our code repository for the environment information to run this command.
Intel is committed to respecting human rights and avoiding causing or contributing to adverse impacts on human rights. See Intel’s Global Human Rights Principles. Intel’s products and software are intended only to be used in applications that do not cause or contribute to adverse impacts on human rights.
@inproceedings{munoz2025mambashedder,
title = {Mamba-Shedder: Post-Transformer Compression for Efficient Selective Structured State Space Models},
author = {Mu{\~n}oz, J. Pablo and Yuan, Jinjie and Jain, Nilesh},
booktitle = "Proceedings of the 2025 Annual Conference of the Nations of the Americas Chapter of the Association for Computational Linguistics (NAACL 2025)",
month = jun,
year = "2025",
address = "Albuquerque, New Mexico",
publisher = "Association for Computational Linguistics",
url = "",
}
This work builds upon work done by the State-Spaces team. Please see the following for additional citations of their work:
Repository: (state-spaces/mamba
Paper: Transformers are SSMs: Generalized Models and Efficient Algorithms Through Structured State Space Duality