--- language: en license: apache-2.0 --- # Shears Model Card: shears-mpt-7b-50-base The sparsified [MPT-7B](https://huggingface.co/mosaicml/mpt-7b) with 50% sparsity as a base model in [Shears](https://arxiv.org/abs/2404.10934). ## Model Sources - **Repository:** [https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning/tree/main/Shears](https://github.com/IntelLabs/Hardware-Aware-Automated-Machine-Learning/tree/main/Shears) - **Paper:** [Shears: Unstructured Sparsity with Neural Low-rank Adapter Search](https://arxiv.org/abs/2404.10934) ## Citation ```bash @article{munoz2024shears, title = {Shears: Unstructured Sparsity with Neural Low-rank Adapter Search}, author={J. Pablo Munoz and Jinjie Yuan and Nilesh Jain}, journal={The 2024 Annual Conference of the North American Chapter of the Association for Computational Linguistics (NAACL-2024)}, year={2024} } ``` ## Acknowledgement Thanks to the work Wanda ([paper](https://arxiv.org/abs/2306.11695), [code](https://github.com/locuslab/wanda)), which provides a simple but effective pruning approach. ## License Apache-2.0