sreejith8100 commited on
Commit
87ffd57
1 Parent(s): 6d5e72d

End of training

Browse files
README.md CHANGED
@@ -1,6 +1,4 @@
1
  ---
2
- license: mit
3
- base_model: naver-clova-ix/donut-base
4
  tags:
5
  - generated_from_trainer
6
  datasets:
@@ -15,9 +13,9 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # donut-base-sroie
17
 
18
- This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 9.6913
21
 
22
  ## Model description
23
 
@@ -42,18 +40,37 @@ The following hyperparameters were used during training:
42
  - seed: 42
43
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
44
  - lr_scheduler_type: linear
45
- - num_epochs: 6
46
 
47
  ### Training results
48
 
49
  | Training Loss | Epoch | Step | Validation Loss |
50
  |:-------------:|:-----:|:----:|:---------------:|
51
- | 18.9088 | 1.0 | 3 | 18.3757 |
52
- | 14.9604 | 2.0 | 6 | 12.0340 |
53
- | 12.7281 | 3.0 | 9 | 11.9031 |
54
- | 10.6777 | 4.0 | 12 | 9.9294 |
55
- | 10.1939 | 5.0 | 15 | 10.0405 |
56
- | 9.9764 | 6.0 | 18 | 9.6913 |
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
57
 
58
 
59
  ### Framework versions
 
1
  ---
 
 
2
  tags:
3
  - generated_from_trainer
4
  datasets:
 
13
 
14
  # donut-base-sroie
15
 
16
+ This model was trained from scratch on the imagefolder dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 5.7016
19
 
20
  ## Model description
21
 
 
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
+ - num_epochs: 25
44
 
45
  ### Training results
46
 
47
  | Training Loss | Epoch | Step | Validation Loss |
48
  |:-------------:|:-----:|:----:|:---------------:|
49
+ | 9.5672 | 1.0 | 3 | 8.6900 |
50
+ | 9.2165 | 2.0 | 6 | 8.1679 |
51
+ | 8.6254 | 3.0 | 9 | 8.8826 |
52
+ | 8.5961 | 4.0 | 12 | 7.9094 |
53
+ | 8.0975 | 5.0 | 15 | 7.6390 |
54
+ | 8.0688 | 6.0 | 18 | 7.3710 |
55
+ | 7.8179 | 7.0 | 21 | 7.1756 |
56
+ | 7.4056 | 8.0 | 24 | 7.1199 |
57
+ | 7.2254 | 9.0 | 27 | 6.8964 |
58
+ | 7.1395 | 10.0 | 30 | 6.8406 |
59
+ | 7.2451 | 11.0 | 33 | 6.5694 |
60
+ | 6.8052 | 12.0 | 36 | 6.7802 |
61
+ | 6.8587 | 13.0 | 39 | 6.3983 |
62
+ | 6.6803 | 14.0 | 42 | 6.3614 |
63
+ | 6.5832 | 15.0 | 45 | 6.2045 |
64
+ | 6.3016 | 16.0 | 48 | 6.3187 |
65
+ | 6.3266 | 17.0 | 51 | 6.1394 |
66
+ | 6.2154 | 18.0 | 54 | 5.9751 |
67
+ | 6.0545 | 19.0 | 57 | 5.9170 |
68
+ | 6.0753 | 20.0 | 60 | 5.9828 |
69
+ | 6.0034 | 21.0 | 63 | 5.9600 |
70
+ | 6.0545 | 22.0 | 66 | 5.8052 |
71
+ | 5.8055 | 23.0 | 69 | 5.7358 |
72
+ | 5.7705 | 24.0 | 72 | 5.6998 |
73
+ | 5.8007 | 25.0 | 75 | 5.7016 |
74
 
75
 
76
  ### Framework versions
generation_config.json CHANGED
@@ -1,8 +1,10 @@
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
 
4
  "eos_token_id": 2,
5
  "forced_eos_token_id": 2,
 
6
  "pad_token_id": 1,
7
  "transformers_version": "4.34.0"
8
  }
 
1
  {
2
  "_from_model_config": true,
3
  "bos_token_id": 0,
4
+ "decoder_start_token_id": 0,
5
  "eos_token_id": 2,
6
  "forced_eos_token_id": 2,
7
+ "max_length": 512,
8
  "pad_token_id": 1,
9
  "transformers_version": "4.34.0"
10
  }
preprocessor_config.json CHANGED
@@ -19,8 +19,8 @@
19
  "processor_class": "DonutProcessor",
20
  "resample": 2,
21
  "rescale_factor": 0.00392156862745098,
22
- "size": [
23
- 800,
24
- 1200
25
- ]
26
  }
 
19
  "processor_class": "DonutProcessor",
20
  "resample": 2,
21
  "rescale_factor": 0.00392156862745098,
22
+ "size": {
23
+ "height": 1200,
24
+ "width": 800
25
+ }
26
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:a1afa8c8cba314946c29f1a854b129e38f31efbc8f7c3f1bc734be8a720a03d5
3
  size 810743577
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ee195ff3b9c28b6512294616a99331bd2bff815135213f4a34377c79f4baa30d
3
  size 810743577
special_tokens_map.json CHANGED
@@ -1,5 +1,10 @@
1
  {
2
  "additional_special_tokens": [
 
 
 
 
 
3
  "<s_iitcdip>",
4
  "<s_synthdog>",
5
  "<s_Ignore_9>",
@@ -84,7 +89,6 @@
84
  "</s_mother_nationalty_1>",
85
  "<s_child_middle_name_2>",
86
  "<s_Ignore_149>",
87
- "<s>",
88
  "<s_Ignore_82>",
89
  "</s_Ignore_59>",
90
  "<s_father_religion_2>",
@@ -150,7 +154,6 @@
150
  "<s_Ignore_111>",
151
  "<s_Ignore_92>",
152
  "<s_Ignore_61>",
153
- "</s>",
154
  "</s_Ignore_112>",
155
  "</s_Ignore_84>",
156
  "</s_Ignore_63>",
 
1
  {
2
  "additional_special_tokens": [
3
+ "<s>",
4
+ "<pad>",
5
+ "</s>",
6
+ "<unk>",
7
+ "<mask>",
8
  "<s_iitcdip>",
9
  "<s_synthdog>",
10
  "<s_Ignore_9>",
 
89
  "</s_mother_nationalty_1>",
90
  "<s_child_middle_name_2>",
91
  "<s_Ignore_149>",
 
92
  "<s_Ignore_82>",
93
  "</s_Ignore_59>",
94
  "<s_father_religion_2>",
 
154
  "<s_Ignore_111>",
155
  "<s_Ignore_92>",
156
  "<s_Ignore_61>",
 
157
  "</s_Ignore_112>",
158
  "</s_Ignore_84>",
159
  "</s_Ignore_63>",
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -2,9 +2,9 @@
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<s>",
5
- "lstrip": true,
6
  "normalized": false,
7
- "rstrip": true,
8
  "single_word": false,
9
  "special": true
10
  },
@@ -18,9 +18,9 @@
18
  },
19
  "2": {
20
  "content": "</s>",
21
- "lstrip": true,
22
  "normalized": false,
23
- "rstrip": true,
24
  "single_word": false,
25
  "special": true
26
  },
@@ -35,7 +35,7 @@
35
  "57521": {
36
  "content": "<mask>",
37
  "lstrip": true,
38
- "normalized": true,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
@@ -3122,6 +3122,11 @@
3122
  }
3123
  },
3124
  "additional_special_tokens": [
 
 
 
 
 
3125
  "<s_iitcdip>",
3126
  "<s_synthdog>",
3127
  "<s_Ignore_9>",
@@ -3206,7 +3211,6 @@
3206
  "</s_mother_nationalty_1>",
3207
  "<s_child_middle_name_2>",
3208
  "<s_Ignore_149>",
3209
- "<s>",
3210
  "<s_Ignore_82>",
3211
  "</s_Ignore_59>",
3212
  "<s_father_religion_2>",
@@ -3272,7 +3276,6 @@
3272
  "<s_Ignore_111>",
3273
  "<s_Ignore_92>",
3274
  "<s_Ignore_61>",
3275
- "</s>",
3276
  "</s_Ignore_112>",
3277
  "</s_Ignore_84>",
3278
  "</s_Ignore_63>",
@@ -3514,11 +3517,18 @@
3514
  "cls_token": "<s>",
3515
  "eos_token": "</s>",
3516
  "mask_token": "<mask>",
 
3517
  "model_max_length": 1000000000000000019884624838656,
 
3518
  "pad_token": "<pad>",
 
 
3519
  "processor_class": "DonutProcessor",
3520
  "sep_token": "</s>",
3521
  "sp_model_kwargs": {},
 
3522
  "tokenizer_class": "XLMRobertaTokenizer",
 
 
3523
  "unk_token": "<unk>"
3524
  }
 
2
  "added_tokens_decoder": {
3
  "0": {
4
  "content": "<s>",
5
+ "lstrip": false,
6
  "normalized": false,
7
+ "rstrip": false,
8
  "single_word": false,
9
  "special": true
10
  },
 
18
  },
19
  "2": {
20
  "content": "</s>",
21
+ "lstrip": false,
22
  "normalized": false,
23
+ "rstrip": false,
24
  "single_word": false,
25
  "special": true
26
  },
 
35
  "57521": {
36
  "content": "<mask>",
37
  "lstrip": true,
38
+ "normalized": false,
39
  "rstrip": false,
40
  "single_word": false,
41
  "special": true
 
3122
  }
3123
  },
3124
  "additional_special_tokens": [
3125
+ "<s>",
3126
+ "<pad>",
3127
+ "</s>",
3128
+ "<unk>",
3129
+ "<mask>",
3130
  "<s_iitcdip>",
3131
  "<s_synthdog>",
3132
  "<s_Ignore_9>",
 
3211
  "</s_mother_nationalty_1>",
3212
  "<s_child_middle_name_2>",
3213
  "<s_Ignore_149>",
 
3214
  "<s_Ignore_82>",
3215
  "</s_Ignore_59>",
3216
  "<s_father_religion_2>",
 
3276
  "<s_Ignore_111>",
3277
  "<s_Ignore_92>",
3278
  "<s_Ignore_61>",
 
3279
  "</s_Ignore_112>",
3280
  "</s_Ignore_84>",
3281
  "</s_Ignore_63>",
 
3517
  "cls_token": "<s>",
3518
  "eos_token": "</s>",
3519
  "mask_token": "<mask>",
3520
+ "max_length": 512,
3521
  "model_max_length": 1000000000000000019884624838656,
3522
+ "pad_to_multiple_of": null,
3523
  "pad_token": "<pad>",
3524
+ "pad_token_type_id": 0,
3525
+ "padding_side": "right",
3526
  "processor_class": "DonutProcessor",
3527
  "sep_token": "</s>",
3528
  "sp_model_kwargs": {},
3529
+ "stride": 0,
3530
  "tokenizer_class": "XLMRobertaTokenizer",
3531
+ "truncation_side": "right",
3532
+ "truncation_strategy": "longest_first",
3533
  "unk_token": "<unk>"
3534
  }