imtiaz114 commited on
Commit
3eba8e1
1 Parent(s): e72d3b7

Training in progress epoch 0

Browse files
Files changed (7) hide show
  1. README.md +7 -14
  2. config.json +34 -29
  3. special_tokens_map.json +5 -5
  4. spiece.model +3 -0
  5. tf_model.h5 +2 -2
  6. tokenizer.json +2 -2
  7. tokenizer_config.json +12 -8
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: mit
3
- base_model: xlm-roberta-base
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
@@ -13,11 +13,11 @@ probably proofread and complete it, then remove this comment. -->
13
 
14
  # imtiaz114/mbert-ner-baseline-1
15
 
16
- This model is a fine-tuned version of [xlm-roberta-base](https://huggingface.co/xlm-roberta-base) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 0.0470
19
- - Validation Loss: 0.2044
20
- - Epoch: 7
21
 
22
  ## Model description
23
 
@@ -36,21 +36,14 @@ More information needed
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 12736, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.1}
40
  - training_precision: mixed_float16
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
- | 0.2960 | 0.1926 | 0 |
47
- | 0.1732 | 0.1713 | 1 |
48
- | 0.1326 | 0.1692 | 2 |
49
- | 0.1070 | 0.1702 | 3 |
50
- | 0.0849 | 0.1813 | 4 |
51
- | 0.0686 | 0.1918 | 5 |
52
- | 0.0564 | 0.1961 | 6 |
53
- | 0.0470 | 0.2044 | 7 |
54
 
55
 
56
  ### Framework versions
 
1
  ---
2
  license: mit
3
+ base_model: ai4bharat/indic-bert
4
  tags:
5
  - generated_from_keras_callback
6
  model-index:
 
13
 
14
  # imtiaz114/mbert-ner-baseline-1
15
 
16
+ This model is a fine-tuned version of [ai4bharat/indic-bert](https://huggingface.co/ai4bharat/indic-bert) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 0.4884
19
+ - Validation Loss: 0.3742
20
+ - Epoch: 0
21
 
22
  ## Model description
23
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 3184, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.1}
40
  - training_precision: mixed_float16
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
+ | 0.4884 | 0.3742 | 0 |
 
 
 
 
 
 
 
47
 
48
 
49
  ### Framework versions
config.json CHANGED
@@ -1,45 +1,50 @@
1
  {
2
- "_name_or_path": "xlm-roberta-base",
3
  "architectures": [
4
- "XLMRobertaForTokenClassification"
5
  ],
6
- "attention_probs_dropout_prob": 0.1,
7
- "bos_token_id": 0,
8
- "classifier_dropout": null,
9
- "eos_token_id": 2,
 
 
 
10
  "hidden_act": "gelu",
11
- "hidden_dropout_prob": 0.1,
12
  "hidden_size": 768,
13
  "id2label": {
14
- "0": "I-PER",
15
- "1": "O",
16
- "2": "I-ORG",
17
- "3": "I-LOC",
18
- "4": "B-LOC",
19
- "5": "B-PER",
20
- "6": "B-ORG"
21
  },
22
  "initializer_range": 0.02,
 
23
  "intermediate_size": 3072,
24
  "label2id": {
25
- "B-LOC": 4,
26
- "B-ORG": 6,
27
- "B-PER": 5,
28
- "I-LOC": 3,
29
- "I-ORG": 2,
30
- "I-PER": 0,
31
- "O": 1
32
  },
33
- "layer_norm_eps": 1e-05,
34
- "max_position_embeddings": 514,
35
- "model_type": "xlm-roberta",
 
36
  "num_attention_heads": 12,
 
37
  "num_hidden_layers": 12,
38
- "output_past": true,
39
- "pad_token_id": 1,
40
  "position_embedding_type": "absolute",
41
  "transformers_version": "4.31.0.dev0",
42
- "type_vocab_size": 1,
43
- "use_cache": true,
44
- "vocab_size": 250002
45
  }
 
1
  {
2
+ "_name_or_path": "ai4bharat/indic-bert",
3
  "architectures": [
4
+ "AlbertForTokenClassification"
5
  ],
6
+ "attention_probs_dropout_prob": 0,
7
+ "bos_token_id": 2,
8
+ "classifier_dropout_prob": 0.1,
9
+ "down_scale_factor": 1,
10
+ "embedding_size": 128,
11
+ "eos_token_id": 3,
12
+ "gap_size": 0,
13
  "hidden_act": "gelu",
14
+ "hidden_dropout_prob": 0,
15
  "hidden_size": 768,
16
  "id2label": {
17
+ "0": "B-PER",
18
+ "1": "I-LOC",
19
+ "2": "I-PER",
20
+ "3": "O",
21
+ "4": "I-ORG",
22
+ "5": "B-ORG",
23
+ "6": "B-LOC"
24
  },
25
  "initializer_range": 0.02,
26
+ "inner_group_num": 1,
27
  "intermediate_size": 3072,
28
  "label2id": {
29
+ "B-LOC": 6,
30
+ "B-ORG": 5,
31
+ "B-PER": 0,
32
+ "I-LOC": 1,
33
+ "I-ORG": 4,
34
+ "I-PER": 2,
35
+ "O": 3
36
  },
37
+ "layer_norm_eps": 1e-12,
38
+ "max_position_embeddings": 512,
39
+ "model_type": "albert",
40
+ "net_structure_type": 0,
41
  "num_attention_heads": 12,
42
+ "num_hidden_groups": 1,
43
  "num_hidden_layers": 12,
44
+ "num_memory_blocks": 0,
45
+ "pad_token_id": 0,
46
  "position_embedding_type": "absolute",
47
  "transformers_version": "4.31.0.dev0",
48
+ "type_vocab_size": 2,
49
+ "vocab_size": 200000
 
50
  }
special_tokens_map.json CHANGED
@@ -1,15 +1,15 @@
1
  {
2
- "bos_token": "<s>",
3
- "cls_token": "<s>",
4
- "eos_token": "</s>",
5
  "mask_token": {
6
- "content": "<mask>",
7
  "lstrip": true,
8
  "normalized": false,
9
  "rstrip": false,
10
  "single_word": false
11
  },
12
  "pad_token": "<pad>",
13
- "sep_token": "</s>",
14
  "unk_token": "<unk>"
15
  }
 
1
  {
2
+ "bos_token": "[CLS]",
3
+ "cls_token": "[CLS]",
4
+ "eos_token": "[SEP]",
5
  "mask_token": {
6
+ "content": "[MASK]",
7
  "lstrip": true,
8
  "normalized": false,
9
  "rstrip": false,
10
  "single_word": false
11
  },
12
  "pad_token": "<pad>",
13
+ "sep_token": "[SEP]",
14
  "unk_token": "<unk>"
15
  }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3a1173c2b6e144a02c001e289a05b5dbefddf247c50d4dcf42633158b2968fcb
3
+ size 5646064
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e2ec9c4ad944351229ad58b4ee10aae1b310347826bc2cf30c61f55b3564e7d
3
- size 1110121552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7fbd569667e98157f97a2a7669a9886aec87224e068f93ab24102fb369d3fa3c
3
+ size 131473068
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f2c509a525eb51aebb33fb59c24ee923c1d4c1db23c3ae81fe05ccf354084f7b
3
- size 17082758
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:431ec72a485a382d7e3a6063af5f54249b36e4870e0c8d079918e79cd0a4f8f1
3
+ size 15285606
tokenizer_config.json CHANGED
@@ -1,19 +1,23 @@
1
  {
2
- "bos_token": "<s>",
3
  "clean_up_tokenization_spaces": true,
4
- "cls_token": "<s>",
5
- "eos_token": "</s>",
 
 
6
  "mask_token": {
7
  "__type": "AddedToken",
8
- "content": "<mask>",
9
  "lstrip": true,
10
- "normalized": true,
11
  "rstrip": false,
12
  "single_word": false
13
  },
14
- "model_max_length": 512,
15
  "pad_token": "<pad>",
16
- "sep_token": "</s>",
17
- "tokenizer_class": "XLMRobertaTokenizer",
 
 
18
  "unk_token": "<unk>"
19
  }
 
1
  {
2
+ "bos_token": "[CLS]",
3
  "clean_up_tokenization_spaces": true,
4
+ "cls_token": "[CLS]",
5
+ "do_lower_case": true,
6
+ "eos_token": "[SEP]",
7
+ "keep_accents": false,
8
  "mask_token": {
9
  "__type": "AddedToken",
10
+ "content": "[MASK]",
11
  "lstrip": true,
12
+ "normalized": false,
13
  "rstrip": false,
14
  "single_word": false
15
  },
16
+ "model_max_length": 1000000000000000019884624838656,
17
  "pad_token": "<pad>",
18
+ "remove_space": true,
19
+ "sep_token": "[SEP]",
20
+ "sp_model_kwargs": {},
21
+ "tokenizer_class": "AlbertTokenizer",
22
  "unk_token": "<unk>"
23
  }