Indahgalaputri/PubMedAbstract2M-BERT

This model is a continued pre-trained version of google-bert/bert-base-uncased

Model description

This BERT-based model has been continuously trained using the Masked Language Modeling (MLM) objective on 2 million samples from the English column of the vi_pubmed

Intended uses & limitations

This model is developed as part of an undergraduate thesis to compare the performance of general language models with domain-specific biomedical language models for Clinical NER applications. It is an experimental model, and its performance may not be optimal. Therefore, I recommend using biobert-v1.1 or PubMedBERT for better results.

Training and evaluation data

The model is trained on 2 million English text samples from vi_pubmed, using a 90:10 split ratio for training and evaluation.

Training procedure

The training procedure follows the tutorial provided by Hugging Face, with the batch size adjusted to 64. Training was conducted on an NVIDIA RTX 4070 GPU.

Training hyperparameters

The following hyperparameters were used during training:

  • optimizer: {'inner_optimizer': {'class_name': 'AdamWeightDecay', 'config': {'name': 'AdamWeightDecay', 'learning_rate': 2e-05, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-07, 'amsgrad': False, 'weight_decay_rate': 0.01}}, 'dynamic': True, 'initial_scale': 32768.0, 'dynamic_growth_steps': 2000}
  • training_precision: mixed_float16

Framework versions

  • Transformers 4.31.0
  • TensorFlow 2.10.1
  • Datasets 3.0.0
  • Tokenizers 0.13.3
Downloads last month
36
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.

Model tree for Indahgalaputri/PubMedAbstract2M-BERT

Finetuned
(3518)
this model

Dataset used to train Indahgalaputri/PubMedAbstract2M-BERT