Turkish Language Models with Huggingface's Transformers

As R&D Team at Loodos, we release cased and uncased versions of most recent language models for Turkish. More details about pretrained models and evaluations on downstream tasks can be found here (our repo).

Turkish BERT-Base (cased)

This is BERT-Base model which has 12 encoder layers with 768 hidden layer size trained on cased Turkish dataset.

Usage

Using AutoModel and AutoTokenizer from Transformers, you can import the model as described below.

from transformers import AutoModel, AutoTokenizer

tokenizer = AutoTokenizer.from_pretrained("loodos/bert-base-turkish-cased")

model = AutoModel.from_pretrained("loodos/bert-base-turkish-cased")

Details and Contact

You contact us to ask a question, open an issue or give feedback via our github repo.

Acknowledgments

Many thanks to TFRC Team for providing us cloud TPUs on Tensorflow Research Cloud to train our models.

Downloads last month
15
Hosted inference API

Unable to determine this model’s pipeline type. Check the docs .