--- language: da tags: - danish - bert - masked-lm - botxo license: cc-by-4.0 datasets: - common_crawl - wikipedia - dindebat.dk - hestenettet.dk - danish OpenSubtitles pipeline_tag: fill-mask widget: - text: "København er [MASK] i Danmark." --- # Danish BERT (version 2, uncased) by [BotXO.ai](https://www.botxo.ai/) All credit goes to [BotXO.ai](https://www.botxo.ai/) who developed Danish BERT. For data and training details see their [GitHub repository](https://github.com/botxo/nordic_bert) or [this article](https://www.botxo.ai/en/blog/danish-bert-model/). It is both available in TensorFlow and Pytorch format. The original TensorFlow version can be downloaded using [this link](https://www.dropbox.com/s/19cjaoqvv2jicq9/danish_bert_uncased_v2.zip?dl=1). Here is an example on how to load Danish BERT in PyTorch using the [🤗Transformers](https://github.com/huggingface/transformers) library: ```python from transformers import AutoTokenizer, AutoModelForPreTraining tokenizer = AutoTokenizer.from_pretrained("Maltehb/danish-bert-botxo") model = AutoModelForPreTraining.from_pretrained("Maltehb/danish-bert-botxo") ```