JavaBERT-uncased / README.md
neltds's picture
Added text example to README file
fa5b5a8
|
raw
history blame
681 Bytes
metadata
language:
  - java
  - code
license: apache-2.0
widget:
  - text: >-
      public [MASK] isOdd(Integer a){if (a % 2 == 0) {return "even";} else
      {return "odd";}}

JavaBERT

A BERT-like model pretrained on Java software code.

Training Data

The model is trained on 2,998,345 Java files retrieved from open source projects on GitHub. A bert-base-cased tokenizer was used.

Training Objective

A MLM (Masked Language Model) objective was used to train this model.

Usage

from transformers import pipeline
pipe = pipeline('fill-mask', model='CAUKiel/JavaBERT')
output = pipe(CODE) # Replace with Java code; Use '[MASK]' to mask tokens/words in the code.