metadata
language: da
license: cc-by-4.0
Danish ConvBERT small (cased)
ConvBERT model pretrained on a custom Danish corpus (~17.5gb). For details regarding data sources and training procedure, along with benchmarks on downstream tasks, go to: https://github.com/sarnikowski/danish_transformers
Usage
from transformers import ConvBertTokenizer, ConvBertModel
tokenizer = ConvBertTokenizer.from_pretrained("sarnikowski/convbert-small-da-cased")
model = ConvBertModel.from_pretrained("sarnikowski/convbert-small-da-cased")
Questions?
If you have any questions feel free to open an issue on the danish_transformers repository, or send an email to p.sarnikowski@gmail.com