tnaumann commited on
Commit
98768ec
1 Parent(s): e39e47b

Renames PubMedELECTRA to BiomedELECTRA

Browse files
Files changed (1) hide show
  1. README.md +5 -3
README.md CHANGED
@@ -5,15 +5,17 @@ tags:
5
  license: mit
6
  ---
7
 
8
- ## PubMedELECTRA-large (abstracts only)
 
 
9
 
10
  Pretraining large neural language models, such as BERT and ELECTRA, has led to impressive gains on many natural language processing (NLP) tasks. However, most pretraining efforts focus on general domain corpora, such as newswire and Web. A prevailing assumption is that even domain-specific pretraining can benefit by starting from general-domain language models. [Recent work](https://arxiv.org/abs/2007.15779) shows that for domains with abundant unlabeled text, such as biomedicine, pretraining language models from scratch results in substantial gains over continual pretraining of general-domain language models. [Followup work](https://arxiv.org/abs/2112.07869) explores alternate pretraining strategies and the impact of these on performance on the BLURB benchmark.
11
 
12
- This PubMedELECTRA is pretrained from scratch using _abstracts_ from [PubMed](https://pubmed.ncbi.nlm.nih.gov/).
13
 
14
  ## Citation
15
 
16
- If you find PubMedELECTRA useful in your research, please cite the following paper:
17
 
18
  ```latex
19
  @misc{https://doi.org/10.48550/arxiv.2112.07869,
 
5
  license: mit
6
  ---
7
 
8
+ ## MSR BiomedELECTRA-large (abstracts only)
9
+
10
+ *NOTE: This model was previously named "PubMedELECTRA-large (abstracts only)".*
11
 
12
  Pretraining large neural language models, such as BERT and ELECTRA, has led to impressive gains on many natural language processing (NLP) tasks. However, most pretraining efforts focus on general domain corpora, such as newswire and Web. A prevailing assumption is that even domain-specific pretraining can benefit by starting from general-domain language models. [Recent work](https://arxiv.org/abs/2007.15779) shows that for domains with abundant unlabeled text, such as biomedicine, pretraining language models from scratch results in substantial gains over continual pretraining of general-domain language models. [Followup work](https://arxiv.org/abs/2112.07869) explores alternate pretraining strategies and the impact of these on performance on the BLURB benchmark.
13
 
14
+ This BiomedELECTRA is pretrained from scratch using _abstracts_ from [PubMed](https://pubmed.ncbi.nlm.nih.gov/).
15
 
16
  ## Citation
17
 
18
+ If you find BiomedELECTRA useful in your research, please cite the following paper:
19
 
20
  ```latex
21
  @misc{https://doi.org/10.48550/arxiv.2112.07869,