Edit model card

Model Description

ClinicalDistilBERT-i2b2-2010 is a lightweight BERT-based model developed by fine-tuning ClinicalDistilBERT on the i2b2-2010 dataset for clinical Named Entity Recognition (NER). It is specifically designed to recognise entities from three categories: problem, treatment, and test.

Architecture

The architecture of this model remains the same as the ClinicalDistilBERT model. The size of the hidden dimension and the embedding layer are both set to 768. The vocabulary size is 28996. The number of transformer layers is 6, and the expansion rate of the feed-forward layer is 4. Overall, this model contains approximately 65 million parameters.

Use Cases

This model is suited for clinical NER and for medical tasks that require identification and classification of problems, treatments, and tests.

Citation

If you use this model, please consider citing the following paper:

@article{rohanian2023lightweight,
  title={Lightweight transformers for clinical natural language processing},
  author={Rohanian, Omid and Nouriborji, Mohammadmahdi and Jauncey, Hannah and Kouchaki, Samaneh and Nooralahzadeh, Farhad and Clifton, Lei and Merson, Laura and Clifton, David A and ISARIC Clinical Characterisation Group and others},
  journal={Natural Language Engineering},
  pages={1--28},
  year={2023},
  publisher={Cambridge University Press}
}
Downloads last month
2,658