File size: 3,020 Bytes
df1f0cd
f6e0fcd
 
 
 
df1f0cd
 
f6e0fcd
9408b89
f6e0fcd
 
 
 
 
df1f0cd
 
 
9408b89
df1f0cd
 
 
9408b89
df1f0cd
9408b89
f6e0fcd
df1f0cd
9408b89
 
 
 
 
 
 
 
 
 
 
 
 
df1f0cd
9408b89
 
 
 
 
 
 
 
 
 
 
 
df1f0cd
9408b89
df1f0cd
 
 
 
 
9408b89
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
df1f0cd
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
---
language: 
  - en
bigbio_language:
  - English
license: other
multilinguality: monolingual
bigbio_license_shortname: MIXED
pretty_name: BLURB
homepage: https://microsoft.github.io/BLURB/tasks.html
bigbio_pubmed: true
bigbio_public: true
bigbio_tasks:
  - NAMED_ENTITY_RECOGNITION
---


# Dataset Card for BLURB

## Dataset Description

- **Homepage:** https://microsoft.github.io/BLURB/tasks.html
- **Pubmed:** True
- **Public:** True
- **Tasks:** NER

BLURB is a collection of resources for biomedical natural language processing. 
In general domains, such as newswire and the Web, comprehensive benchmarks and 
leaderboards such as GLUE have greatly accelerated progress in open-domain NLP. 
In biomedicine, however, such resources are ostensibly scarce. In the past, 
there have been a plethora of shared tasks in biomedical NLP, such as 
BioCreative, BioNLP Shared Tasks, SemEval, and BioASQ, to name just a few. These 
efforts have played a significant role in fueling interest and progress by the 
research community, but they typically focus on individual tasks. The advent of 
neural language models, such as BERT provides a unifying foundation to leverage 
transfer learning from unlabeled text to support a wide range of NLP 
applications. To accelerate progress in biomedical pretraining strategies and 
task-specific methods, it is thus imperative to create a broad-coverage 
benchmark encompassing diverse biomedical tasks. 

Inspired by prior efforts toward this direction (e.g., BLUE), we have created 
BLURB (short for Biomedical Language Understanding and Reasoning Benchmark). 
BLURB comprises of a comprehensive benchmark for PubMed-based biomedical NLP 
applications, as well as a leaderboard for tracking progress by the community. 
BLURB includes thirteen publicly available datasets in six diverse tasks. To 
avoid placing undue emphasis on tasks with many available datasets, such as 
named entity recognition (NER), BLURB reports the macro average across all tasks 
as the main score. The BLURB leaderboard is model-agnostic. Any system capable 
of producing the test predictions using the same training and development data 
can participate. The main goal of BLURB is to lower the entry barrier in 
biomedical NLP and help accelerate progress in this vitally important field for 
positive societal and human impact.

This implementation contains a subset of 5 tasks as of 2022.10.06, with their original train, dev, and test splits.


## Citation Information

```
@article{gu2021domain,
    title = {
        Domain-specific language model pretraining for biomedical natural
        language processing
    },
    author = {
        Gu, Yu and Tinn, Robert and Cheng, Hao and Lucas, Michael and
        Usuyama, Naoto and Liu, Xiaodong and Naumann, Tristan and Gao,
        Jianfeng and Poon, Hoifung
    },
    year = 2021,
    journal = {ACM Transactions on Computing for Healthcare (HEALTH)},
    publisher = {ACM New York, NY},
    volume = 3,
    number = 1,
    pages = {1--23}
}
```