julien-c HF staff commited on
Commit
9c6a7f4
1 Parent(s): 66eaddd

Migrate model card from transformers-repo

Browse files

Read announcement at https://discuss.huggingface.co/t/announcement-all-model-cards-will-be-migrated-to-hf-co-model-repos/2755
Original file history: https://github.com/huggingface/transformers/commits/master/model_cards/seiya/oubiobert-base-uncased/README.md

Files changed (1) hide show
  1. README.md +51 -0
README.md ADDED
@@ -0,0 +1,51 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ tags:
3
+ - exbert
4
+ license: apache-2.0
5
+ ---
6
+
7
+ # ouBioBERT-Base, Uncased
8
+
9
+ Bidirectional Encoder Representations from Transformers for Biomedical Text Mining by Osaka University (ouBioBERT) is a language model based on the BERT-Base (Devlin, et al., 2019) architecture. We pre-trained ouBioBERT on PubMed abstracts from the PubMed baseline (ftp://ftp.ncbi.nlm.nih.gov/pubmed/baseline) via our method.
10
+
11
+ The details of the pre-training procedure can be found in Wada, et al. (2020).
12
+
13
+ ## Evaluation
14
+
15
+ We evaluated the performance of ouBioBERT in terms of the biomedical language understanding evaluation (BLUE) benchmark (Peng, et al., 2019). The numbers are mean (standard deviation) on five different random seeds.
16
+
17
+
18
+ | Dataset | Task Type | Score |
19
+ |:----------------|:-----------------------------|-------------:|
20
+ | MedSTS | Sentence similarity | 84.9 (0.6) |
21
+ | BIOSSES | Sentence similarity | 92.3 (0.8) |
22
+ | BC5CDR-disease | Named-entity recognition | 87.4 (0.1) |
23
+ | BC5CDR-chemical | Named-entity recognition | 93.7 (0.2) |
24
+ | ShARe/CLEFE | Named-entity recognition | 80.1 (0.4) |
25
+ | DDI | Relation extraction | 81.1 (1.5) |
26
+ | ChemProt | Relation extraction | 75.0 (0.3) |
27
+ | i2b2 2010 | Relation extraction | 74.0 (0.8) |
28
+ | HoC | Document classification | 86.4 (0.5) |
29
+ | MedNLI | Inference | 83.6 (0.7) |
30
+ | **Total** | Macro average of the scores |**83.8 (0.3)**|
31
+
32
+
33
+ ## Code for Fine-tuning
34
+ We made the source code for fine-tuning freely available at [our repository](https://github.com/sy-wada/blue_benchmark_with_transformers).
35
+
36
+ ## Citation
37
+
38
+ If you use our work in your research, please kindly cite the following paper:
39
+
40
+ ```bibtex
41
+ @misc{2005.07202,
42
+ Author = {Shoya Wada and Toshihiro Takeda and Shiro Manabe and Shozo Konishi and Jun Kamohara and Yasushi Matsumura},
43
+ Title = {A pre-training technique to localize medical BERT and enhance BioBERT},
44
+ Year = {2020},
45
+ Eprint = {arXiv:2005.07202},
46
+ }
47
+ ```
48
+
49
+ <a href="https://huggingface.co/exbert/?model=seiya/oubiobert-base-uncased&sentence=Coronavirus%20disease%20(COVID-19)%20is%20caused%20by%20SARS-COV2%20and%20represents%20the%20causative%20agent%20of%20a%20potentially%20fatal%20disease%20that%20is%20of%20great%20global%20public%20health%20concern.">
50
+ <img width="300px" src="https://cdn-media.huggingface.co/exbert/button.png">
51
+ </a>