KB-BERT for NER

Mixed cased and uncased data

This model is based on KB-BERT and was fine-tuned on the SUCX 3.0 - NER corpus, using the simple tags and partially lowercased data. For this model we used a variation of the data that did not use BIO-encoding to differentiate between the beginnings (B), and insides (I) of named entity tags.

The model was trained on the training data only, with the best model chosen by its performance on the validation data. You find more information about the model and the performance on our blog: https://kb-labb.github.io/posts/2022-02-07-sucx3_ner

Downloads last month
39
Safetensors
Model size
124M params
Tensor type
I64
·
F32
·
Inference Examples
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Dataset used to train KBLab/bert-base-swedish-lowermix-reallysimple-ner

Collection including KBLab/bert-base-swedish-lowermix-reallysimple-ner