Shresthadev403 commited on
Commit
1f707cb
1 Parent(s): 4c0f394

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- license: apache-2.0
3
- base_model: bert-base-uncased
4
  tags:
5
  - generated_from_trainer
6
  model-index:
@@ -13,14 +11,14 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # ner-bert-ingredients
15
 
16
- This model is a fine-tuned version of [bert-base-uncased](https://huggingface.co/bert-base-uncased) on an unknown dataset.
17
  It achieves the following results on the evaluation set:
18
- - eval_loss: 4.5379
19
- - eval_runtime: 0.0188
20
- - eval_samples_per_second: 53.052
21
- - eval_steps_per_second: 53.052
22
- - epoch: 30.0
23
- - step: 30
24
 
25
  ## Model description
26
 
 
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  model-index:
 
11
 
12
  # ner-bert-ingredients
13
 
14
+ This model was trained from scratch on an unknown dataset.
15
  It achieves the following results on the evaluation set:
16
+ - eval_loss: 4.6987
17
+ - eval_runtime: 0.0184
18
+ - eval_samples_per_second: 54.299
19
+ - eval_steps_per_second: 54.299
20
+ - epoch: 50.0
21
+ - step: 50
22
 
23
  ## Model description
24
 
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "bert-base-uncased",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
 
1
  {
2
+ "_name_or_path": "/kaggle/working/ner-bert-ingredients",
3
  "architectures": [
4
  "BertForTokenClassification"
5
  ],
logs/events.out.tfevents.1705982894.70e47a1f5afe.42.10 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a8cf1a326acf37230827be10bd766abddfc1f7f139188f0edb355399d001477f
3
- size 7868
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4160a39b5645f9319c76ff3d3cadb1a08c17e33ead50552c98de67ee0265f294
3
+ size 8288
logs/events.out.tfevents.1705983035.70e47a1f5afe.42.11 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:40e3106771c36c1a648980573eab6d7576766a4a44e6df90f9c4ecf149a8fcd5
3
+ size 7086
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ca597349ea03bf594dd26319fcd2e7cd81cf40596d0f2c852834fc2111d0504b
3
  size 435756040
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:45d04dad0a7c2b4a89b02a87f0f119ab92b550cd7e094ff234eb79f7dd90d2a7
3
  size 435756040
special_tokens_map.json CHANGED
@@ -1,7 +1,37 @@
1
  {
2
- "cls_token": "[CLS]",
3
- "mask_token": "[MASK]",
4
- "pad_token": "[PAD]",
5
- "sep_token": "[SEP]",
6
- "unk_token": "[UNK]"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  }
 
1
  {
2
+ "cls_token": {
3
+ "content": "[CLS]",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "mask_token": {
10
+ "content": "[MASK]",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "pad_token": {
17
+ "content": "[PAD]",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ },
23
+ "sep_token": {
24
+ "content": "[SEP]",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "unk_token": {
31
+ "content": "[UNK]",
32
+ "lstrip": false,
33
+ "normalized": false,
34
+ "rstrip": false,
35
+ "single_word": false
36
+ }
37
  }
tokenizer_config.json CHANGED
@@ -45,11 +45,18 @@
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
 
48
  "model_max_length": 512,
 
49
  "pad_token": "[PAD]",
 
 
50
  "sep_token": "[SEP]",
 
51
  "strip_accents": null,
52
  "tokenize_chinese_chars": true,
53
  "tokenizer_class": "BertTokenizer",
 
 
54
  "unk_token": "[UNK]"
55
  }
 
45
  "cls_token": "[CLS]",
46
  "do_lower_case": true,
47
  "mask_token": "[MASK]",
48
+ "max_length": 512,
49
  "model_max_length": 512,
50
+ "pad_to_multiple_of": null,
51
  "pad_token": "[PAD]",
52
+ "pad_token_type_id": 0,
53
+ "padding_side": "right",
54
  "sep_token": "[SEP]",
55
+ "stride": 0,
56
  "strip_accents": null,
57
  "tokenize_chinese_chars": true,
58
  "tokenizer_class": "BertTokenizer",
59
+ "truncation_side": "right",
60
+ "truncation_strategy": "longest_first",
61
  "unk_token": "[UNK]"
62
  }
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 4.4464287757873535,
3
  "best_model_checkpoint": "ner-bert-ingredients/checkpoint-10",
4
- "epoch": 30.0,
5
  "eval_steps": 10,
6
- "global_step": 30,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -49,6 +49,34 @@
49
  "eval_samples_per_second": 53.052,
50
  "eval_steps_per_second": 53.052,
51
  "step": 30
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
52
  }
53
  ],
54
  "logging_steps": 10,
@@ -56,7 +84,7 @@
56
  "num_input_tokens_seen": 0,
57
  "num_train_epochs": 50,
58
  "save_steps": 10,
59
- "total_flos": 19713692873160.0,
60
  "train_batch_size": 16,
61
  "trial_name": null,
62
  "trial_params": null
 
1
  {
2
  "best_metric": 4.4464287757873535,
3
  "best_model_checkpoint": "ner-bert-ingredients/checkpoint-10",
4
+ "epoch": 50.0,
5
  "eval_steps": 10,
6
+ "global_step": 50,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
49
  "eval_samples_per_second": 53.052,
50
  "eval_steps_per_second": 53.052,
51
  "step": 30
52
+ },
53
+ {
54
+ "epoch": 40.0,
55
+ "learning_rate": 1e-05,
56
+ "loss": 0.0764,
57
+ "step": 40
58
+ },
59
+ {
60
+ "epoch": 40.0,
61
+ "eval_loss": 4.603794574737549,
62
+ "eval_runtime": 0.0208,
63
+ "eval_samples_per_second": 48.022,
64
+ "eval_steps_per_second": 48.022,
65
+ "step": 40
66
+ },
67
+ {
68
+ "epoch": 50.0,
69
+ "learning_rate": 4e-05,
70
+ "loss": 0.063,
71
+ "step": 50
72
+ },
73
+ {
74
+ "epoch": 50.0,
75
+ "eval_loss": 4.698660850524902,
76
+ "eval_runtime": 0.0184,
77
+ "eval_samples_per_second": 54.299,
78
+ "eval_steps_per_second": 54.299,
79
+ "step": 50
80
  }
81
  ],
82
  "logging_steps": 10,
 
84
  "num_input_tokens_seen": 0,
85
  "num_train_epochs": 50,
86
  "save_steps": 10,
87
+ "total_flos": 32856154788600.0,
88
  "train_batch_size": 16,
89
  "trial_name": null,
90
  "trial_params": null