ayoubkirouane's picture
Update README.md
3ea52aa
metadata
license: apache-2.0
base_model: distilroberta-base
tags:
  - generated_from_trainer
model-index:
  - name: distilroberta-base-finetuned-wikitext2
    results: []
datasets:
  - wikitext
language:
  - en
pipeline_tag: fill-mask

Overview

  • Model Name: FILL-MAsk-RoBERTa-base
  • Task: Masked Language Modeling (FILL-MAsk)
  • Dataset: WikiText2

Model Description

FILL-MAsk-RoBERTa-base is a distilled version of the RoBERTa-base model, designed for the Masked Language Modeling task. This model follows a similar training procedure as DistilBERT, resulting in a smaller model with 6 layers, 768 dimensions, and 12 attention heads. It contains a total of 82 million parameters, making it more lightweight compared to the original RoBERTa-base, which has 125 million parameters. On average, DistilRoBERTa is approximately twice as fast as RoBERTa-base.

Usage

FILL-MAsk-RoBERTa-base can be used for both direct and downstream tasks. It is suitable for masked language modeling tasks, where tokens are masked, and the model must predict the masked tokens. It is also intended to be fine-tuned on downstream tasks such as sequence classification, token classification, or question answering. Users can explore the Hugging Face Model Hub to find fine-tuned versions of this model for specific tasks of interest.

Limitations

  • Bias: Significant research has explored bias and fairness issues with language models. Predictions generated by this model may contain biases, including harmful stereotypes related to protected classes, identity characteristics, and sensitive social and occupational groups.

  • Fairness: It's essential to be aware of fairness considerations when using this model and to ensure that its predictions do not contribute to unfair or harmful outcomes.

  • Ethical Use: Users are encouraged to use this model ethically and responsibly, taking into account the potential for bias and ensuring that it does not generate harmful or offensive content.

  • Usage :

# Use a pipeline as a high-level helper
from transformers import pipeline

pipe = pipeline("fill-mask", model="ayoubkirouane/FILL-MAsk-RoBERTa-base")
result = pipe("The capital of Algeria is <mask>.")
print(result)