ofirzaf commited on
Commit
056596a
1 Parent(s): b0c4d29

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +3 -2
README.md CHANGED
@@ -3,6 +3,7 @@ language: en
3
  ---
4
  # 90% Sparse BERT-Large (uncased) Fine Tuned on SQuADv1.1
5
  This model is a result of fine-tuning a Prune OFA 90% sparse pre-trained BERT-Large combined with knowledge distillation.
6
- This model yields the following results on SQuADv1.1 development set:
 
7
 
8
- `{"exact_match": 83.56669820245979, "f1": 90.20829352733487}`
3
  ---
4
  # 90% Sparse BERT-Large (uncased) Fine Tuned on SQuADv1.1
5
  This model is a result of fine-tuning a Prune OFA 90% sparse pre-trained BERT-Large combined with knowledge distillation.
6
+ This model yields the following results on SQuADv1.1 development set:<br>
7
+ `{"exact_match": 83.56669820245979, "f1": 90.20829352733487}`
8
 
9
+ For further details see our paper, [Prune Once for All: Sparse Pre-Trained Language Models](https://arxiv.org/abs/2111.05754), and our open source implementation available [here](https://github.com/IntelLabs/Model-Compression-Research-Package/tree/main/research/prune-once-for-all).