90% Sparse DistilBERT-Base (uncased) Prune OFA
This model is a result from our paper Prune Once for All: Sparse Pre-Trained Language Models presented in ENLSP NeurIPS Workshop 2021.
For further details on the model and its result, see our paper and our implementation available here.
- Downloads last month
This model can be loaded on the Inference API on-demand.