stefan-it's picture
readme: add initial version of model card (#2)
789f744
|
raw
history blame
3.85 kB
---
language: de
license: mit
tags:
- flair
- token-classification
- sequence-tagger-model
base_model: deepset/gbert-base
widget:
- text: PASt ( KvD ) - Polizeipräsidium Westhessen [ Newsroom ] Wiesbaden ( ots )
- Am Sonntag , den 27.01.2019 führte die Autobahnpolizei Wiesbaden in Zusammenarbeit
mit der Präsidialwache in der Zeit von 11:00 - 16:00 Uhr eine Geschwindigkeitsmessung
in der Baustelle der A66 am Wiesbadener Kreuz durch .
---
# Fine-tuned Flair Model on German MobIE Dataset with AutoTrain
This Flair model was fine-tuned on the
[German MobIE](https://aclanthology.org/2021.konvens-1.22/)
NER Dataset using GBERT Base as backbone LM and the 🚀 [AutoTrain](https://github.com/huggingface/autotrain-advanced)
library.
## Dataset
The [German MobIE](https://github.com/DFKI-NLP/MobIE) dataset is a German-language dataset, which is human-annotated
with 20 coarse- and fine-grained entity types and entity linking information for geographically linkable entities. The
dataset consists of 3,232 social media texts and traffic reports with 91K tokens, and contains 20.5K annotated
entities, 13.1K of which are linked to a knowledge base.
The following named entities are annotated:
* `location-stop`
* `trigger`
* `organization-company`
* `location-city`
* `location`
* `event-cause`
* `location-street`
* `time`
* `date`
* `number`
* `duration`
* `organization`
* `person`
* `set`
* `distance`
* `disaster-type`
* `money`
* `org-position`
* `percent`
## Fine-Tuning
The latest [Flair version](https://github.com/flairNLP/flair/tree/42ea3f6854eba04387c38045f160c18bdaac07dc) is used for
fine-tuning. Additionally, the model is trained with the
[FLERT (Schweter and Akbik (2020)](https://arxiv.org/abs/2011.06993) approach, because the MobIE dataset thankfully
comes with document boundary information marker.
A hyper-parameter search over the following parameters with 5 different seeds per configuration is performed:
* Batch Sizes: [`16`]
* Learning Rates: [`5e-05`, `3e-05`]
All models are trained with the awesome [AutoTrain Advanced](https://github.com/huggingface/autotrain-advanced) from
Hugging Face. More details can be found in this [repository](https://github.com/stefan-it/autotrain-flair-mobie).
## Results
A hyper-parameter search with 5 different seeds per configuration is performed and micro F1-score on development set
is reported:
| Configuration | Seed 1 | Seed 2 | Seed 3 | Seed 4 | Seed 5 | Average |
|--------------------|-------------|-------------|-------------|-----------------|--------------|-----------------|
| `bs16-e10-lr5e-05` | [0.8446][1] | [0.8495][2] | [0.8455][3] | [0.8419][4] | [0.8476][5] | 0.8458 ± 0.0029 |
| `bs16-e10-lr3e-05` | [0.8392][6] | [0.8445][7] | [0.8495][8] | [**0.8381**][9] | [0.8449][10] | 0.8432 ± 0.0046 |
[1]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr5e-05-1
[2]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr5e-05-2
[3]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr5e-05-3
[4]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr5e-05-4
[5]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr5e-05-5
[6]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr3e-05-1
[7]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr3e-05-2
[8]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr3e-05-3
[9]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr3e-05-4
[10]: https://hf.co/stefan-it/autotrain-flair-mobie-gbert_base-bs16-e10-lr3e-05-5
The result in bold shows the performance of this model.
Additionally, the Flair [training log](training.log) and [TensorBoard logs](tensorboard) are also uploaded to the model
hub.