--- license: mit datasets: - wikitext language: - en ---
This repository demonstrates how to use Hugging Face Transformers library and TensorFlow to solve a masked language modeling task using DistilBERT. Specifically, we will use the pretrained "distilbert-base-cased" model to predict a missing word in a sentence from the "wikitext-2-raw-v1" dataset.
The goal of this project is to predict a missing word in a sentence using the pretrained "distilbert-base-cased" model. The model should take a sentence with a masked token and output the most probable word to fill in the masked token.
Here are the necessary libraries and modules:
The algorithmic approach to solving this problem is outlined below:
Run the provided Python script to perform masked language modeling with DistilBERT on the given dataset. The script will output the most probable predicted token for the masked position in the sentence.
This project is licensed under the MIT License. See the LICENSE file for more information.