File size: 1,761 Bytes
368da34
6cdd2d5
368da34
 
4efd183
ad283e8
 
 
 
4efd183
6cdd2d5
ad283e8
4efd183
 
 
 
 
ad283e8
 
 
 
 
 
4efd183
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
---
language: ar
license: apache-2.0
datasets:
- AQMAR
- ANERcorp
embeddings:
- GloVe
- Flair
---
# Arabic NER Model using Flair Embeddings
Training was conducted over 94 epochs, using a linear decaying learning rate of 2e-05, starting from 0.225 and a batch size of 32 with GloVe and Flair forward and backward embeddings.

Results:
- F1-score (micro) 0.8666
- F1-score (macro) 0.8488

|      | tp  | fp | fn | precision | recall | class-F1 |
|------|-----|----|----|-----------|--------|----------|
| LOC  | 539 | 51 | 68 | 0.9136    | 0.8880 | 0.9006   |
| MISC | 408 | 57 | 89 | 0.8774    | 0.8209 | 0.8482   |
| ORG  | 167 | 43 | 64 | 0.7952    | 0.7229 | 0.7574   |
| PER  | 501 | 65 | 60 | 0.8852    | 0.8930 | 0.8891   |

---

```
2020-10-27 12:05:47,801 Model: "SequenceTagger(
  (embeddings): StackedEmbeddings(
    (list_embedding_0): WordEmbeddings('glove')
    (list_embedding_1): FlairEmbeddings(
      (lm): LanguageModel(
        (drop): Dropout(p=0.1, inplace=False)
        (encoder): Embedding(7125, 100)
        (rnn): LSTM(100, 2048)
        (decoder): Linear(in_features=2048, out_features=7125, bias=True)
      )
    )
    (list_embedding_2): FlairEmbeddings(
      (lm): LanguageModel(
        (drop): Dropout(p=0.1, inplace=False)
        (encoder): Embedding(7125, 100)
        (rnn): LSTM(100, 2048)
        (decoder): Linear(in_features=2048, out_features=7125, bias=True)
      )
    )
  )
  (word_dropout): WordDropout(p=0.05)
  (locked_dropout): LockedDropout(p=0.5)
  (embedding2nn): Linear(in_features=4196, out_features=4196, bias=True)
  (rnn): LSTM(4196, 256, batch_first=True, bidirectional=True)
  (linear): Linear(in_features=512, out_features=15, bias=True)
  (beta): 1.0
  (weights): None
  (weight_tensor) None
  
 ```