xww033 commited on
Commit
0ec8a4e
1 Parent(s): d5aec00

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +2 -0
README.md CHANGED
@@ -14,6 +14,8 @@ The model performance on the test sets are:
14
  |NER-PMR-large (multi-task model)| 92.9 | 54.7| 87.8| 88.4|
15
 
16
  Note that the performance of RoBERTa-large and PMR-large are single-task fine-tuning, while NER-PMR-large is a multi-task fine-tuned model.
 
 
17
 
18
  ### How to use
19
  You can try the codes from [this repo](https://github.com/DAMO-NLP-SG/PMR/NER) for both training and inference.
 
14
  |NER-PMR-large (multi-task model)| 92.9 | 54.7| 87.8| 88.4|
15
 
16
  Note that the performance of RoBERTa-large and PMR-large are single-task fine-tuning, while NER-PMR-large is a multi-task fine-tuned model.
17
+ As it is fine-tuned on multiple datasets, we believe that NER-PMR-large has a better generalization capability to other NER tasks than PMR-large and RoBERTa-large.
18
+
19
 
20
  ### How to use
21
  You can try the codes from [this repo](https://github.com/DAMO-NLP-SG/PMR/NER) for both training and inference.