tr-aravindan commited on
Commit
ad24589
·
1 Parent(s): 7207f89

Training in progress epoch 0

Browse files
Files changed (4) hide show
  1. README.md +9 -11
  2. config.json +1 -1
  3. tf_model.h5 +1 -1
  4. tokenizer_config.json +42 -0
README.md CHANGED
@@ -15,9 +15,9 @@ probably proofread and complete it, then remove this comment. -->
15
 
16
  This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - Train Loss: 0.0277
19
- - Validation Loss: 0.0521
20
- - Epoch: 2
21
 
22
  ## Model description
23
 
@@ -36,21 +36,19 @@ More information needed
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
- - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 2634, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
40
  - training_precision: mixed_float16
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
- | 0.1720 | 0.0629 | 0 |
47
- | 0.0472 | 0.0573 | 1 |
48
- | 0.0277 | 0.0521 | 2 |
49
 
50
 
51
  ### Framework versions
52
 
53
- - Transformers 4.31.0
54
- - TensorFlow 2.12.0
55
- - Datasets 2.14.1
56
- - Tokenizers 0.13.3
 
15
 
16
  This model is a fine-tuned version of [bert-base-cased](https://huggingface.co/bert-base-cased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Train Loss: 0.1673
19
+ - Validation Loss: 0.0670
20
+ - Epoch: 0
21
 
22
  ## Model description
23
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - optimizer: {'name': 'AdamWeightDecay', 'learning_rate': {'module': 'keras.optimizers.schedules', 'class_name': 'PolynomialDecay', 'config': {'initial_learning_rate': 2e-05, 'decay_steps': 2634, 'end_learning_rate': 0.0, 'power': 1.0, 'cycle': False, 'name': None}, 'registered_name': None}, 'decay': 0.0, 'beta_1': 0.9, 'beta_2': 0.999, 'epsilon': 1e-08, 'amsgrad': False, 'weight_decay_rate': 0.01}
40
  - training_precision: mixed_float16
41
 
42
  ### Training results
43
 
44
  | Train Loss | Validation Loss | Epoch |
45
  |:----------:|:---------------:|:-----:|
46
+ | 0.1673 | 0.0670 | 0 |
 
 
47
 
48
 
49
  ### Framework versions
50
 
51
+ - Transformers 4.35.0
52
+ - TensorFlow 2.14.0
53
+ - Datasets 2.14.6
54
+ - Tokenizers 0.14.1
config.json CHANGED
@@ -40,7 +40,7 @@
40
  "num_hidden_layers": 12,
41
  "pad_token_id": 0,
42
  "position_embedding_type": "absolute",
43
- "transformers_version": "4.31.0",
44
  "type_vocab_size": 2,
45
  "use_cache": true,
46
  "vocab_size": 28996
 
40
  "num_hidden_layers": 12,
41
  "pad_token_id": 0,
42
  "position_embedding_type": "absolute",
43
+ "transformers_version": "4.35.0",
44
  "type_vocab_size": 2,
45
  "use_cache": true,
46
  "vocab_size": 28996
tf_model.h5 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:64a91c1c7e7d1dd44fcad12f83273b27231955ae5c06896c93209b10b7362082
3
  size 431179756
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73855d8d73d772688e8d47b0f3be236703d406c2c7023f961245590324e96882
3
  size 431179756
tokenizer_config.json CHANGED
@@ -1,4 +1,46 @@
1
  {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
  "clean_up_tokenization_spaces": true,
3
  "cls_token": "[CLS]",
4
  "do_lower_case": false,
 
1
  {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "[PAD]",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "100": {
12
+ "content": "[UNK]",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "101": {
20
+ "content": "[CLS]",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "102": {
28
+ "content": "[SEP]",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "103": {
36
+ "content": "[MASK]",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ }
43
+ },
44
  "clean_up_tokenization_spaces": true,
45
  "cls_token": "[CLS]",
46
  "do_lower_case": false,