CodeTrans model for program synthesis
Table of Contents
- Model Details
- How to Get Started With the Model
- Uses
- Risks, Limitations and Biases
- Training
- Evaluation
- Environmental Impact
- Citation Information
Model Details
- Model Description: This CodeTrans model is based on the
t5-small
model. It has its own SentencePiece vocabulary model. It used transfer-learning pre-training on 7 unsupervised datasets in the software development domain. It is then fine-tuned on the program synthesis task for the lisp inspired DSL code. - Developed by: Ahmed Elnaggar,Wei Ding
- Model Type: Summarization
- Language(s): English
- License: Unknown
- Resources for more information:
How to Get Started With the Model
Here is how to use this model to generate lisp inspired DSL code using Transformers SummarizationPipeline:
from transformers import AutoTokenizer, AutoModelWithLMHead, SummarizationPipeline
pipeline = SummarizationPipeline(
model=AutoModelWithLMHead.from_pretrained("SEBIS/code_trans_t5_small_program_synthese_transfer_learning_finetune"),
tokenizer=AutoTokenizer.from_pretrained("SEBIS/code_trans_t5_small_program_synthese_transfer_learning_finetune", skip_special_tokens=True),
device=0
)
tokenized_code = "you are given an array of numbers a and a number b , compute the difference of elements in a and b"
pipeline([tokenized_code])
Run this example in colab notebook.
Training data
The supervised training tasks datasets can be downloaded on Link
Uses
Direct Use
The model could be used to generate lisp inspired DSL code given the human language description tasks.
Risks, Limitations and Biases
As detailed in this model’s publication, this model makes use of the data-set One Billion Word Language Model Benchmark corpus in order to gather the self-supervised English data samples.
Significant research has explored bias and fairness issues with language models (see, e.g., Sheng et al. (2021) and Bender et al. (2021)). As such, it should be noted that language models that are pretrained from text corpus such as the One Billion Word Word Language Model Benchmark corpus have been further explored (e.g by Ngo, Helen & Araújo et al(2021) reports that the One Billion Word Word Language Model Benchmark corpus
“generate text in the linguistic style of news, without any grounding in the real world. In addition to potential harms from models which are inadvertently optimized for generating fake news.”
The aforementioned publication continues to warn that the One Billion Word Word Language Model Benchmark corpus
contains sentences which contain words commonly found on blocklists. While these sentences may have plausibly been used in expository contexts within the article, the destructive sentence-level preprocessing and shuffling applied to lm1b [One Billion Word Word Language Model Benchmark corpus] removes all long-range structure from the text and makes it infeasible to track the context and intent of individual examples.
Ngo, Helen & Araújo et al(2021)
Training
Training Data
The supervised training tasks datasets can be downloaded on Link
The authors provide additionally notes about the vocabulary used, in the associated paper:
We used the SentencePiece model (Kudo, 2018) to construct the vocabulary for this research, as well as to decode and encode the input/output.
Training procedure
Preprocessing
Transfer-learning Pretraining
The model was trained on a single TPU Pod V3-8 for 500,000 steps in total, using sequence length 512 (batch size 4096). It has a total of approximately 220M parameters and was trained using the encoder-decoder architecture. The optimizer used is AdaFactor with inverse square root learning rate schedule for pre-training.
Fine-tuning
This model was then fine-tuned on a single TPU Pod V2-8 for 5,000 steps in total, using sequence length 512 (batch size 256), using only the dataset only containing lisp inspired DSL data.
Evaluation
Results
For the code documentation tasks, different models achieves the following results on different programming languages (in BLEU score):
Test results :
Language / Model | LISP |
---|---|
CodeTrans-ST-Small | 89.43 |
CodeTrans-ST-Base | 89.65 |
CodeTrans-TF-Small | 90.30 |
CodeTrans-TF-Base | 90.24 |
CodeTrans-TF-Large | 90.21 |
CodeTrans-MT-Small | 82.88 |
CodeTrans-MT-Base | 86.99 |
CodeTrans-MT-Large | 90.27 |
CodeTrans-MT-TF-Small | 90.31 |
CodeTrans-MT-TF-Base | 90.30 |
CodeTrans-MT-TF-Large | 90.17 |
State of the art | 85.80 |
Environmental Impact
Carbon emissions can be estimated using the Machine Learning Impact calculator presented in Lacoste et al. (2019). We present the hardware type based on the associated paper.
Hardware Type: Nvidia RTX 8000 GPUs
Hours used: Unknown
Cloud Provider: GCC TPU v2-8 and v3-8.
Compute Region: Unknown
Carbon Emitted: Unknown
Citation Information
@misc{elnaggar2021codetrans,
title={CodeTrans: Towards Cracking the Language of Silicon's Code Through Self-Supervised Deep Learning and High Performance Computing},
author={Ahmed Elnaggar and Wei Ding and Llion Jones and Tom Gibbs and Tamas Feher and Christoph Angerer and Silvia Severini and Florian Matthes and Burkhard Rost},
year={2021},
eprint={2104.02443},
archivePrefix={arXiv},
primaryClass={cs.SE}
}
- Downloads last month
- 14