# Fill-Mask PyTorch Model (Camembert) This model is a `fill-mask` model that was trained using the PyTorch framework and the Hugging Face Transformers library. It was utilized in Hugging Face's NLP course as an introductory model. ## Model Description This model uses the `camembert` architecture, a variant of the RoBERTa model adapted for French. It's designed for the fill-mask task, where a portion of input text is masked and the model predicts the missing token. ## Features - **PyTorch**: The model was implemented and trained using the PyTorch deep learning framework, which allows for dynamic computation graphs and is known for its flexibility and efficiency. - **Safetensors**: The model utilizes Safetensors, a Python library that provides safer operations for PyTorch Tensors. - **Transformers**: The model was built using the Hugging Face Transformers library, a state-of-the-art NLP library that provides thousands of pre-trained models and easy-to-use implementations of transformer architectures. - **AutoTrain Compatible**: This model is compatible with Hugging Face's AutoTrain, a tool that automates the training of transformer models. ## Usage ```python from transformers import CamembertForMaskedLM, CamembertTokenizer tokenizer = CamembertTokenizer.from_pretrained('model-name') model = CamembertForMaskedLM.from_pretrained('model-name') inputs = tokenizer("Le camembert est .", return_tensors='pt') outputs = model(**inputs) predictions = outputs.logits predicted_index = torch.argmax(predictions[0, mask_position]).item() predicted_token = tokenizer.convert_ids_to_tokens([predicted_index])[0]