JavaBERT-uncased / README.md
neltds's picture
Create README.md
f34e776
|
raw
history blame
473 Bytes

JavaBERT

A BERT-like model pretrained on Java software code.

Training Data

The model is trained on 2,998,345 Java files retrieved from Open Source projects on GitHub.

Training Objective

MLM (Masked Language Model) objective was used to train this model.

Usage

from transformers import pipeline
pipe = pipeline('fill-mask', model='CAUKiel/JavaBERT')
output = pipe(CODE) # Replace with Java code; Use '[MASK]' to mask tokens/words in the code.