--- widget: - text: "low lung volumes, [MASK] pulmonary vascularity." tags: - fill-mask - pytorch - transformers - bert - biobert - radbert - language-model - uncased - radiology - biomedical datasets: - wikipedia - bookscorpus - pubmed - radreports language: - en license: mit --- RadBERT was continuously pre-trained on radiology reports from a BioBERT initialization. ## Citation ```bibtex @article{chambon_cook_langlotz_2022, title={Improved fine-tuning of in-domain transformer model for inferring COVID-19 presence in multi-institutional radiology reports}, DOI={10.1007/s10278-022-00714-8}, journal={Journal of Digital Imaging}, author={Chambon, Pierre and Cook, Tessa S. and Langlotz, Curtis P.}, year={2022} } ```