blurb / README.md
gabrielaltay's picture
update metadata in README (#1)
f6e0fcd
---
language:
- en
bigbio_language:
- English
license: other
multilinguality: monolingual
bigbio_license_shortname: MIXED
pretty_name: BLURB
homepage: https://microsoft.github.io/BLURB/tasks.html
bigbio_pubmed: true
bigbio_public: true
bigbio_tasks:
- NAMED_ENTITY_RECOGNITION
---
# Dataset Card for BLURB
## Dataset Description
- **Homepage:** https://microsoft.github.io/BLURB/tasks.html
- **Pubmed:** True
- **Public:** True
- **Tasks:** NER
BLURB is a collection of resources for biomedical natural language processing.
In general domains, such as newswire and the Web, comprehensive benchmarks and
leaderboards such as GLUE have greatly accelerated progress in open-domain NLP.
In biomedicine, however, such resources are ostensibly scarce. In the past,
there have been a plethora of shared tasks in biomedical NLP, such as
BioCreative, BioNLP Shared Tasks, SemEval, and BioASQ, to name just a few. These
efforts have played a significant role in fueling interest and progress by the
research community, but they typically focus on individual tasks. The advent of
neural language models, such as BERT provides a unifying foundation to leverage
transfer learning from unlabeled text to support a wide range of NLP
applications. To accelerate progress in biomedical pretraining strategies and
task-specific methods, it is thus imperative to create a broad-coverage
benchmark encompassing diverse biomedical tasks.
Inspired by prior efforts toward this direction (e.g., BLUE), we have created
BLURB (short for Biomedical Language Understanding and Reasoning Benchmark).
BLURB comprises of a comprehensive benchmark for PubMed-based biomedical NLP
applications, as well as a leaderboard for tracking progress by the community.
BLURB includes thirteen publicly available datasets in six diverse tasks. To
avoid placing undue emphasis on tasks with many available datasets, such as
named entity recognition (NER), BLURB reports the macro average across all tasks
as the main score. The BLURB leaderboard is model-agnostic. Any system capable
of producing the test predictions using the same training and development data
can participate. The main goal of BLURB is to lower the entry barrier in
biomedical NLP and help accelerate progress in this vitally important field for
positive societal and human impact.
This implementation contains a subset of 5 tasks as of 2022.10.06, with their original train, dev, and test splits.
## Citation Information
```
@article{gu2021domain,
title = {
Domain-specific language model pretraining for biomedical natural
language processing
},
author = {
Gu, Yu and Tinn, Robert and Cheng, Hao and Lucas, Michael and
Usuyama, Naoto and Liu, Xiaodong and Naumann, Tristan and Gao,
Jianfeng and Poon, Hoifung
},
year = 2021,
journal = {ACM Transactions on Computing for Healthcare (HEALTH)},
publisher = {ACM New York, NY},
volume = 3,
number = 1,
pages = {1--23}
}
```