Edit model card

Filipino Sentence RoBERTa

We finetuned RoBERTa Tagalog Base (finetuned on COHFIE) on NewsPH-NLI to learn to encode filipino/tagalog sentences to sentence embeddings. We used sentence-transformers to finetune the model. All model details, training setups, and corpus details can be found in this paper: Automatic WordNet Construction using Word Sense Induction through Sentence Embeddings.

Intended uses & limitations

The intended use of this model is to extract sentence embeddings which will be used for clustering. This model may not be safe for use in production since we did not examine it for biases. Please use it with caution.

How to use

Using this model is easier when you have sentence-transformers installed:

pip install -U sentence-transformers

Here is how to use this model to encode sentences to sentence embeddings using SentenceTransformer:

from sentence_transformers import SentenceTransformer

model = SentenceTransformer("danjohnvelasco/filipino-sentence-roberta-v1")
sentence_list = ["sentence 1", "sentence 2", "sentence 3"]
sentence_embeddings = model.encode(sentence_list)

BibTeX entry and citation info

If you use this model, please cite our work:

  doi = {10.48550/ARXIV.2204.03251},
  url = {https://arxiv.org/abs/2204.03251},
  author = {Velasco, Dan John and Alba, Axel and Pelagio, Trisha Gail and Ramirez, Bryce Anthony and Cruz, Jan Christian Blaise and Cheng, Charibeth},
  keywords = {Computation and Language (cs.CL), FOS: Computer and information sciences, FOS: Computer and information sciences},
  title = {Automatic WordNet Construction using Word Sense Induction through Sentence Embeddings},
  publisher = {arXiv},
  year = {2022},
  copyright = {Creative Commons Attribution 4.0 International}
Downloads last month
Hosted inference API
This model can be loaded on the Inference API on-demand.

Dataset used to train danjohnvelasco/filipino-sentence-roberta-v1