SW-v1 / README.md
eolang's picture
Update README.md
c34691f
metadata
datasets:
  - xnli
language:
  - sw
library_name: transformers
examples: null
widget:
  - text: Joe Bidden ni rais wa [MASK].
    example_title: Sentence 1
  - text: Tumefanya mabadiliko muhimu [MASK] sera zetu za faragha na vidakuzi
    example_title: Sentence 2
  - text: Mtoto anaweza kupoteza [MASK] kabisa
    example_title: Sentence 3

SW

Model description

This is a transformers model pre-trained on a large corpus of Swahili data in a self-supervised fashion. This means it was pre-trained on the raw texts only, with no humans labeling them in any way (which is why it can use lots of publicly available data) with an automatic process to generate inputs and labels from those texts. More precisely, it was pre-trained with one objective:

  • Masked language modeling (MLM): taking a sentence, the model randomly masks 15% of the words in the input then run the entire masked sentence through the model and has to predict the masked words. This is different from traditional recurrent neural networks (RNNs) that usually see the terms one after the other, or from autoregressive models like GPT which internally masks the future tokens. It allows the model to learn a bidirectional representation of the sentence.

This way, the model learns an inner representation of the Swahili language that can then be used to extract features useful for downstream tasks e.g.

  • Named Entity Recognition (Token Classification)
  • Text Classification

The model is based on the Orginal BERT UNCASED which can be found on google-research/bert readme

Intended uses & limitations

You can use the raw model for masked language modeling, but it's primarily intended to be fine-tuned on a downstream task.

How to use

You can use this model directly with a pipeline for masked language modeling:

Tokenizer

from transformers import AutoTokenizer, AutoModelForMaskedLM

tokenizer = AutoTokenizer.from_pretrained("eolang/SW-v1")

text = "Hii ni tovuti ya idhaa ya Kiswahili ya BBC ambayo hukuletea habari na makala kutoka Afrika na kote duniani kwa lugha ya Kiswahili."
encoded_input = tokenizer(text, return_tensors='pt')
output = model(**encoded_input)
print(output)

Fill Mask Model

from transformers import AutoTokenizer, AutoModelForMaskedLM
from transformers import pipeline

tokenizer = AutoTokenizer.from_pretrained("eolang/SW-v1")
model = AutoModelForMaskedLM.from_pretrained("eolang/SW-v1")

fill_mask = pipeline("fill-mask", model=model, tokenizer=tokenizer)
sample_text = "Tumefanya mabadiliko muhimu [MASK] sera zetu za faragha na vidakuzi"

for prediction in fill_mask(sample_text):
    print(f"{prediction['sequence']}, confidence: {prediction['score']}")

Limitations and Bias

Even if the training data used for this model could be reasonably neutral, this model can have biased predictions. This is something I'm still working on improving. Feel free to share suggestions/comments via Discussions