# BERTOverflow ## Model description We pre-trained BERT-base model on 152 million sentences from the StackOverflow's 10 year archive. #### How to use ```python from transformers import * import torch tokenizer = AutoTokenizer.from_pretrained("jeniya/BERTOverflow") model = AutoModel.from_pretrained("jeniya/BERTOverflow") ``` ### BibTeX entry and citation info ```bibtex @inproceedings{tabassum2020code, title={Code and Named Entity Recognition in StackOverflow}, author={Tabassum, Jeniya and Maddela, Mounica and Xu, Wei and Ritter, Alan }, booktitle = {Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics (ACL)}, year = {2020}, } ```