EfficientLLM: Scalable Pruning-Aware Pretraining for Architecture-Agnostic Edge Language Models Paper • 2502.06663 • Published Feb 10 • 1
EfficientLLM: Pruning-Aware Pretraining Collection This is the models of our paper "EfficientLLM: Scalable Pruning-Aware Pretraining for Architecture-Agnostic Edge Language Models". • 3 items • Updated Feb 10
BiPFT: Binary Pre-trained Foundation Transformer with Low-rank Estimation of Binarization Residual Polynomials Paper • 2312.08937 • Published Dec 14, 2023
LM-Cocktail: Resilient Tuning of Language Models via Model Merging Paper • 2311.13534 • Published Nov 22, 2023 • 3