Edit model card

codebert-gpt2-commitgen

This model is a fine-tuned version on dataset provided in the paper titled "Towards Automatic Generation of Short Summaries of Commits" by Siyuan Jiang and Collin McMillan. Heres are the links

Paper :https://arxiv.org/abs/1708.09492 Data : https://sjiang1.github.io/commitgen

Model description

This is a sequence2sequence model with microsoft/codebert-base as encoder and gpt2 as decoder. Givena gitdiff file, this model can generate a short commit message summarizing the change.

Intended uses & limitations

The intended use is to automate github commit message. One limitation to consider is that the model can generate a summary of changes, but is only confined to type of change and might not be able to provide details about the change or output specific keywords related to change.

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • lr_scheduler_warmup_steps: 2000
  • num_epochs: 3

Training results

  • global_step=4521
  • training_loss=3.55994465065804
  • train_runtime: 3300.0492
  • train_samples_per_second: 21.919
  • train_steps_per_second: 1.37
  • total_flos: 1.062667587499776e+16
  • train_loss: 3.55994465065804

Framework versions

  • Transformers 4.25.1
  • Pytorch 1.13.0+cu116
  • Tokenizers 0.13.2
Downloads last month
9