gunghio commited on
Commit
eb28cbe
1 Parent(s): 2d7fb58

update model

Browse files
Files changed (6) hide show
  1. README.md +10 -10
  2. config.json +2 -3
  3. pytorch_model.bin +2 -2
  4. tokenizer.json +0 -0
  5. tokenizer_config.json +1 -1
  6. vocab.txt +0 -0
README.md CHANGED
@@ -6,21 +6,21 @@ metrics:
6
  - accuracy
7
 
8
  model-index:
9
- - name: distilbert-base-multilingual-cased-finetuned-conll2003-ner
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
- # distilbert-base-multilingual-cased-finetuned-conll2003-ner
16
 
17
  This model was trained from scratch on an unkown dataset.
18
  It achieves the following results on the evaluation set:
19
- - Loss: 0.0635
20
- - Precision: 0.9269
21
- - Recall: 0.9337
22
- - F1: 0.9303
23
- - Accuracy: 0.9835
24
 
25
  ## Model description
26
 
@@ -51,9 +51,9 @@ The following hyperparameters were used during training:
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
53
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
54
- | 0.2338 | 1.0 | 878 | 0.0753 | 0.9188 | 0.9089 | 0.9138 | 0.9795 |
55
- | 0.0541 | 2.0 | 1756 | 0.0681 | 0.9362 | 0.9278 | 0.9320 | 0.9830 |
56
- | 0.031 | 3.0 | 2634 | 0.0635 | 0.9269 | 0.9337 | 0.9303 | 0.9835 |
57
 
58
 
59
  ### Framework versions
 
6
  - accuracy
7
 
8
  model-index:
9
+ - name: gunghio/distilbert-base-multilingual-cased-finetuned-conll2003-ner
10
  ---
11
 
12
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
13
  should probably proofread and complete it, then remove this comment. -->
14
 
15
+ # gunghio/distilbert-base-multilingual-cased-finetuned-conll2003-ner
16
 
17
  This model was trained from scratch on an unkown dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 0.0484
20
+ - Precision: 0.9340
21
+ - Recall: 0.9413
22
+ - F1: 0.9376
23
+ - Accuracy: 0.9875
24
 
25
  ## Model description
26
 
 
51
 
52
  | Training Loss | Epoch | Step | Validation Loss | Precision | Recall | F1 | Accuracy |
53
  |:-------------:|:-----:|:----:|:---------------:|:---------:|:------:|:------:|:--------:|
54
+ | 0.1931 | 1.0 | 878 | 0.0518 | 0.9146 | 0.9276 | 0.9210 | 0.9852 |
55
+ | 0.0389 | 2.0 | 1756 | 0.0470 | 0.9261 | 0.9389 | 0.9325 | 0.9870 |
56
+ | 0.0228 | 3.0 | 2634 | 0.0484 | 0.9340 | 0.9413 | 0.9376 | 0.9875 |
57
 
58
 
59
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "distilbert-base-multilingual-cased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForTokenClassification"
@@ -35,12 +35,11 @@
35
  "model_type": "distilbert",
36
  "n_heads": 12,
37
  "n_layers": 6,
38
- "output_past": true,
39
  "pad_token_id": 0,
40
  "qa_dropout": 0.1,
41
  "seq_classif_dropout": 0.2,
42
  "sinusoidal_pos_embds": false,
43
  "tie_weights_": true,
44
  "transformers_version": "4.6.1",
45
- "vocab_size": 119547
46
  }
 
1
  {
2
+ "_name_or_path": "distilbert-base-uncased",
3
  "activation": "gelu",
4
  "architectures": [
5
  "DistilBertForTokenClassification"
 
35
  "model_type": "distilbert",
36
  "n_heads": 12,
37
  "n_layers": 6,
 
38
  "pad_token_id": 0,
39
  "qa_dropout": 0.1,
40
  "seq_classif_dropout": 0.2,
41
  "sinusoidal_pos_embds": false,
42
  "tie_weights_": true,
43
  "transformers_version": "4.6.1",
44
+ "vocab_size": 30522
45
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:790c1d8f79c4a1b09b3836462f3be2a7a741284977d99a353812d499693d61c7
3
- size 539006405
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:76b8f43b516f5f0340d896894e4aa0f3f903752bac218e5c8d1b8477a853a7c3
3
+ size 265521605
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-multilingual-cased"}
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "distilbert-base-uncased"}
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff