JavaBERT
A BERT-like model pretrained on Java software code.
Training Data
The model is trained on 2,998,345 Java files retrieved from Open Source projects on GitHub.
Training Objective
MLM (Masked Language Model) objective was used to train this model.
Usage
from transformers import pipeline
pipe = pipeline('fill-mask', model='CAUKiel/JavaBERT')
output = pipe(CODE) # Replace with Java code; Use '[MASK]' to mask tokens/words in the code.