code-mt5-base / README.md
jackal1586's picture
Update README.md
a424f3c

Tokenizer

We trained our tokenizer using sentencepiece's unigram tokenizer. Then loaded the tokenizer as MT5TokenizerFast.

Model

We used MT5-base model.

Datasets

We used Code Search Net's dataset and some scrapped data from internet to train the model. We maintained a list of datasets where each dataset had codes of same language.

Plots

Train loss

train loss

Evaluation loss

eval loss

Evaluation accuracy

eval accuracy

Learning rate

learning rate

Fine tuning (WIP)

We fine tuned the model with CodeXGLUE code-to-code-trans dataset, and scrapper data.