--- license: mit arxiv: 2302.04026 pipeline_tag: fill-mask tags: - code --- # C-BERT MLM ## Exploring Software Naturalness through Neural Language Models ## Overview This model is the unofficial HuggingFace version of "[C-BERT](http://arxiv.org/abs/2302.04026)" with just the masked language modeling head for pretraining. The weights come from "[An Empirical Comparison of Pre-Trained Models of Source Code](http://arxiv.org/abs/2302.04026)". Please cite the authors if you use this in an academic setting.