ClinicalBERT
This model card describes the ClinicalBERT model, which was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed.
Pretraining Data
The ClinicalBERT model was trained on a large multicenter dataset with a large corpus of 1.2B words of diverse diseases we constructed. For more details, see here.
Model Pretraining
Pretraining Procedures
The training code can be found here and the model was trained on four A100 GPU. Model parameters were initialized with xxx.
Pretraining Hyperparameters
We used a batch size of xx, a maximum sequence length of xx, and a learning rate of xx for pre-training our models. The model was trained for xx steps. The dup factor for duplicating input data with different masks was set to 5. All other default parameters were used (xxx).
How to use the model
Load the model via the transformers library:
from transformers import AutoTokenizer, AutoModel
tokenizer = AutoTokenizer.from_pretrained("kimpty/ClinicalBERT")
model = AutoModel.from_pretrained("kimpty/ClinicalBERT")
More Information
Refer to the paper xxx.
Questions?
Post a Github issue on the xxx repo or email xxx with any questions.