syedyusufali commited on
Commit
ae8652c
1 Parent(s): 7c70021

Training in progress epoch 0

Browse files
Files changed (6) hide show
  1. README.md +6 -8
  2. config.json +6 -6
  3. tf_model.h5 +2 -2
  4. tokenizer.json +0 -0
  5. tokenizer_config.json +1 -1
  6. vocab.txt +0 -0
README.md CHANGED
@@ -12,11 +12,11 @@ probably proofread and complete it, then remove this comment. -->
12
 
13
  # syedyusufali/bert-finetuned-ner
14
 
15
- This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
- - Train Loss: 0.1530
18
- - Validation Loss: 0.1800
19
- - Epoch: 2
20
 
21
  ## Model description
22
 
@@ -35,16 +35,14 @@ More information needed
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 663, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
39
  - training_precision: mixed_float16
40
 
41
  ### Training results
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
- | 0.2587 | 0.1929 | 0 |
46
- | 0.1748 | 0.1819 | 1 |
47
- | 0.1530 | 0.1800 | 2 |
48
 
49
 
50
  ### Framework versions
 
12
 
13
  # syedyusufali/bert-finetuned-ner
14
 
15
+ This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Train Loss: 0.2808
18
+ - Validation Loss: 0.1717
19
+ - Epoch: 0
20
 
21
  ## Model description
22
 
 
35
  ### Training hyperparameters
36
 
37
  The following hyperparameters were used during training:
38
+ - optimizer: {'inner_optimizer': {'class_name': 'AdamWeightDecay', 'config': {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 1017, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}}, 'dynamic': True, 'initial_scale': 32768.0, 'dynamic_growth_steps': 2000}
39
  - training_precision: mixed_float16
40
 
41
  ### Training results
42
 
43
  | Train Loss | Validation Loss | Epoch |
44
  |:----------:|:---------------:|:-----:|
45
+ | 0.2808 | 0.1717 | 0 |
 
 
46
 
47
 
48
  ### Framework versions
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bert-base-cased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
@@ -11,14 +11,14 @@
11
  "hidden_size": 768,
12
  "id2label": {
13
  "0": "O",
14
- "1": "B",
15
- "2": "I"
16
  },
17
  "initializer_range": 0.02,
18
  "intermediate_size": 3072,
19
  "label2id": {
20
- "B": "1",
21
- "I": "2",
22
  "O": "0"
23
  },
24
  "layer_norm_eps": 1e-12,
@@ -31,5 +31,5 @@
31
  "transformers_version": "4.18.0",
32
  "type_vocab_size": 2,
33
  "use_cache": true,
34
- "vocab_size": 28996
35
  }
 
1
  {
2
+ "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
 
11
  "hidden_size": 768,
12
  "id2label": {
13
  "0": "O",
14
+ "1": "B-Disease",
15
+ "2": "I-Disease"
16
  },
17
  "initializer_range": 0.02,
18
  "intermediate_size": 3072,
19
  "label2id": {
20
+ "B-Disease": "1",
21
+ "I-Disease": "2",
22
  "O": "0"
23
  },
24
  "layer_norm_eps": 1e-12,
 
31
  "transformers_version": "4.18.0",
32
  "type_vocab_size": 2,
33
  "use_cache": true,
34
+ "vocab_size": 30522
35
  }
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab0ce6bc03dd46d2f21df51025f05c38f4be6b5432fd99939fab026a772105a7
3
- size 431173652
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3bd615c26e473086a5a3c8817478292101f4e7c47e6274296baa65f6e48981d9
3
+ size 435861524
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"do_lower_case": false, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-cased", "tokenizer_class": "BertTokenizer"}
 
1
+ {"do_lower_case": true, "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": "[MASK]", "tokenize_chinese_chars": true, "strip_accents": null, "from_pt": true, "model_max_length": 512, "special_tokens_map_file": null, "name_or_path": "bert-base-uncased", "tokenizer_class": "BertTokenizer"}
vocab.txt CHANGED
The diff for this file is too large to render. See raw diff