hjb
commited on
Commit
•
c791f42
1
Parent(s):
34c5503
Update README
Browse files
README.md
CHANGED
@@ -16,7 +16,7 @@ metrics:
|
|
16 |
---
|
17 |
|
18 |
# Ælæctra - Finetuned for Named Entity Recognition on the [DaNE dataset](https://danlp.alexandra.dk/304bd159d5de/datasets/ddt.zip) (Hvingelby et al., 2020).
|
19 |
-
**Ælæctra** is a Danish Transformer-based language model created to enhance the variety of Danish NLP resources with a more efficient model compared to previous state-of-the-art (SOTA) models.
|
20 |
|
21 |
Ælæctra was pretrained with the ELECTRA-Small (Clark et al., 2020) pretraining approach by using the Danish Gigaword Corpus (Strømberg-Derczynski et al., 2020) and evaluated on Named Entity Recognition (NER) tasks. Since NER only presents a limited picture of Ælæctra's capabilities I am very interested in further evaluations. Therefore, if you employ it for any task, feel free to hit me up your findings!
|
22 |
|
@@ -27,8 +27,8 @@ Here is an example on how to load the finetuned Ælæctra-cased model for Named
|
|
27 |
```python
|
28 |
from transformers import AutoTokenizer, AutoModelForTokenClassification
|
29 |
|
30 |
-
tokenizer = AutoTokenizer.from_pretrained("Maltehb/-l-
|
31 |
-
model = AutoModelForTokenClassification.from_pretrained("Maltehb/-l-
|
32 |
```
|
33 |
|
34 |
### Evaluation of current Danish Language Models
|
|
|
16 |
---
|
17 |
|
18 |
# Ælæctra - Finetuned for Named Entity Recognition on the [DaNE dataset](https://danlp.alexandra.dk/304bd159d5de/datasets/ddt.zip) (Hvingelby et al., 2020).
|
19 |
+
**Ælæctra** is a Danish Transformer-based language model created to enhance the variety of Danish NLP resources with a more efficient model compared to previous state-of-the-art (SOTA) models.
|
20 |
|
21 |
Ælæctra was pretrained with the ELECTRA-Small (Clark et al., 2020) pretraining approach by using the Danish Gigaword Corpus (Strømberg-Derczynski et al., 2020) and evaluated on Named Entity Recognition (NER) tasks. Since NER only presents a limited picture of Ælæctra's capabilities I am very interested in further evaluations. Therefore, if you employ it for any task, feel free to hit me up your findings!
|
22 |
|
|
|
27 |
```python
|
28 |
from transformers import AutoTokenizer, AutoModelForTokenClassification
|
29 |
|
30 |
+
tokenizer = AutoTokenizer.from_pretrained("Maltehb/-l-ctra-danish-electra-small-cased-ner-dane")
|
31 |
+
model = AutoModelForTokenClassification.from_pretrained("Maltehb/-l-ctra-danish-electra-small-cased-ner-dane")
|
32 |
```
|
33 |
|
34 |
### Evaluation of current Danish Language Models
|