Back to all models

Inference API has been turned off for this model.

Monthly model downloads

jcblaise/bert-tagalog-base-uncased-WWM jcblaise/bert-tagalog-base-uncased-WWM
last 30 days



Contributed by

jcblaise Jan Christian Blaise Cruz
13 models

How to use this model directly from the πŸ€—/transformers library:

Copy to clipboard
from transformers import AutoTokenizer, AutoModel tokenizer = AutoTokenizer.from_pretrained("jcblaise/bert-tagalog-base-uncased-WWM") model = AutoModel.from_pretrained("jcblaise/bert-tagalog-base-uncased-WWM")

BERT Tagalog Base Uncased (Whole Word Masking)

Tagalog version of BERT trained on a large preprocessed text corpus scraped and sourced from the internet. This model is part of a larger research project. We open-source the model to allow greater usage within the Filipino NLP community. This particular version uses whole word masking.


The model can be loaded and used in both PyTorch and TensorFlow through the HuggingFace Transformers package.

from transformers import TFAutoModel, AutoModel, AutoTokenizer

# TensorFlow
model = TFAutoModel.from_pretrained('jcblaise/bert-tagalog-base-uncased-WWM', from_pt=True)
tokenizer = AutoTokenizer.from_pretrained('jcblaise/bert-tagalog-base-uncased-WWM', do_lower_case=True)

# PyTorch
model = AutoModel.from_pretrained('jcblaise/bert-tagalog-base-uncased-WWM')
tokenizer = AutoTokenizer.from_pretrained('jcblaise/bert-tagalog-base-uncased-WWM', do_lower_case=True)

Finetuning scripts and other utilities we use for our projects can be found in our centralized repository at


All model details and training setups can be found in our papers. If you use our model or find it useful in your projects, please cite our work:

  title={{Localization of Fake News Detection via Multitask Transfer Learning}},
  author={Cruz, Jan Christian Blaise and Tan, Julianne Agatha and Cheng, Charibeth},
  booktitle={Proceedings of The 12th Language Resources and Evaluation Conference},

  title={Establishing Baselines for Text Classification in Low-Resource Languages},
  author={Cruz, Jan Christian Blaise and Cheng, Charibeth},
  journal={arXiv preprint arXiv:2005.02068},

  title={Evaluating Language Model Finetuning Techniques for Low-resource Languages},
  author={Cruz, Jan Christian Blaise and Cheng, Charibeth},
  journal={arXiv preprint arXiv:1907.00409},

Data and Other Resources

Data used to train this model as well as other benchmark datasets in Filipino can be found in my website at


If you have questions, concerns, or if you just want to chat about NLP and low-resource languages in general, you may reach me through my work email at