surgicberta / README.md
marcobombieri's picture
Update README.md
413356f
metadata
license: cc-by-nc-nd-4.0
widget:
  - text: hemithyroidectomy is the removal of half of the ⟨mask⟩ gland.
    example_title: Anatomical target
  - text: >-
      Li-Fraumeni ⟨mask⟩ is a hereditary tumor with autosomal dominant
      inheritance.
    example_title: Surgical terminology

SurgicBERTa

SurgicBERTa is a language model based on RoBERTa-base (Liu et al., 2019) architecture. We adapted RoBERTa-base to different surgical textbooks and academic papers via continued pretraining. This amount to about 7 million words and 300k surgical sentences. We used the full text of the books and papers in training, not just abstracts. Specific details of the adaptive pretraining procedure and evaluation tasks can be found in the paper below cited.

Citation

If using this model, please cite the following paper:

@article{bombieri_et_al_surgical_srl_2022,
title = {Machine understanding surgical actions from intervention procedure textbooks},
journal = {Computers in Biology and Medicine},
pages = {106415},
year = {2022},
issn = {0010-4825},
doi = {https://doi.org/10.1016/j.compbiomed.2022.106415},
url = {https://www.sciencedirect.com/science/article/pii/S0010482522011234},
author = {Marco Bombieri and Marco Rospocher and Simone Paolo Ponzetto and Paolo Fiorini},
keywords = {Semantic role labeling, Surgical data science, Procedural knowledge, Information extraction, Natural language processing}
}