davelotito commited on
Commit
893e9cd
1 Parent(s): 73aa20a

End of training

Browse files
README.md CHANGED
@@ -1,13 +1,13 @@
1
  ---
2
  license: mit
3
- base_model: naver-clova-ix/donut-base
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
  - imagefolder
8
  metrics:
9
  - bleu
10
- - rouge
11
  model-index:
12
  - name: donut-base-sroie-v2
13
  results: []
@@ -18,18 +18,17 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # donut-base-sroie-v2
20
 
21
- This model is a fine-tuned version of [naver-clova-ix/donut-base](https://huggingface.co/naver-clova-ix/donut-base) on the imagefolder dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.2767
24
- - Bleu: 0.0208
25
- - Precisions: [0.9802904564315352, 0.9578246392896781, 0.9331742243436754, 0.9045161290322581]
26
- - Brevity Penalty: 0.0221
27
- - Length Ratio: 0.2077
28
- - Translation Length: 964
29
- - Reference Length: 4641
30
- - Rouge: 0.2990
31
- - Gen Len: 20.0
32
- - Meteor: 0.0936
33
 
34
  ## Model description
35
 
@@ -61,17 +60,17 @@ The following hyperparameters were used during training:
61
 
62
  ### Training results
63
 
64
- | Training Loss | Epoch | Step | Validation Loss | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Rouge | Gen Len | Meteor |
65
- |:-------------:|:-----:|:----:|:---------------:|:------:|:--------------------------------------------------------------------------------:|:---------------:|:------------:|:------------------:|:----------------:|:------:|:-------:|:------:|
66
- | No log | 0.99 | 62 | 0.3836 | 0.0204 | [0.9853709508881923, 0.9686800894854586, 0.950661853188929, 0.9296875] | 0.0213 | 0.2062 | 957 | 4641 | 0.3040 | 20.0 | 0.0939 |
67
- | 0.0493 | 2.0 | 125 | 0.2654 | 0.0206 | [0.9823100936524454, 0.9621380846325167, 0.9413173652694611, 0.917098445595855] | 0.0217 | 0.2071 | 961 | 4641 | 0.2996 | 20.0 | 0.0939 |
68
- | 0.0493 | 2.99 | 187 | 0.2727 | 0.0207 | [0.9823284823284824, 0.9621802002224694, 0.9401913875598086, 0.9146183699870634] | 0.0218 | 0.2073 | 962 | 4641 | 0.3007 | 20.0 | 0.0937 |
69
- | 0.0734 | 3.97 | 248 | 0.2767 | 0.0208 | [0.9802904564315352, 0.9578246392896781, 0.9331742243436754, 0.9045161290322581] | 0.0221 | 0.2077 | 964 | 4641 | 0.2990 | 20.0 | 0.0936 |
70
 
71
 
72
  ### Framework versions
73
 
74
- - Transformers 4.39.0.dev0
75
- - Pytorch 2.2.1+cu121
76
- - Datasets 2.17.1
77
  - Tokenizers 0.15.2
 
1
  ---
2
  license: mit
3
+ base_model: davelotito/donut-base-sroie
4
  tags:
5
  - generated_from_trainer
6
  datasets:
7
  - imagefolder
8
  metrics:
9
  - bleu
10
+ - wer
11
  model-index:
12
  - name: donut-base-sroie-v2
13
  results: []
 
18
 
19
  # donut-base-sroie-v2
20
 
21
+ This model is a fine-tuned version of [davelotito/donut-base-sroie](https://huggingface.co/davelotito/donut-base-sroie) on the imagefolder dataset.
22
  It achieves the following results on the evaluation set:
23
+ - Loss: 0.4355
24
+ - Bleu: 0.8879
25
+ - Precisions: [0.943646408839779, 0.9119229045271179, 0.8854285064787452, 0.860009225092251]
26
+ - Brevity Penalty: 0.9868
27
+ - Length Ratio: 0.9869
28
+ - Translation Length: 4525
29
+ - Reference Length: 4585
30
+ - Cer: 0.0857
31
+ - Wer: 0.2978
 
32
 
33
  ## Model description
34
 
 
60
 
61
  ### Training results
62
 
63
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Precisions | Brevity Penalty | Length Ratio | Translation Length | Reference Length | Cer | Wer |
64
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:--------------------------------------------------------------------------------:|:---------------:|:------------:|:------------------:|:----------------:|:------:|:------:|
65
+ | No log | 0.99 | 62 | 0.4638 | 0.8823 | [0.9399823477493381, 0.9044528977399866, 0.8772128915115751, 0.8514851485148515] | 0.9884 | 0.9884 | 4532 | 4585 | 0.0912 | 0.3085 |
66
+ | 0.0043 | 2.0 | 125 | 0.4421 | 0.8853 | [0.9405155320555189, 0.9059428060768543, 0.8794470881486517, 0.8537931034482759] | 0.9899 | 0.9900 | 4539 | 4585 | 0.0889 | 0.3050 |
67
+ | 0.0043 | 2.99 | 187 | 0.4328 | 0.8904 | [0.9399122807017544, 0.9068267734044919, 0.8809201623815968, 0.8558682223747426] | 0.9945 | 0.9945 | 4560 | 4585 | 0.0842 | 0.2939 |
68
+ | 0.0106 | 3.97 | 248 | 0.4355 | 0.8879 | [0.943646408839779, 0.9119229045271179, 0.8854285064787452, 0.860009225092251] | 0.9868 | 0.9869 | 4525 | 4585 | 0.0857 | 0.2978 |
69
 
70
 
71
  ### Framework versions
72
 
73
+ - Transformers 4.40.0.dev0
74
+ - Pytorch 2.1.0
75
+ - Datasets 2.18.0
76
  - Tokenizers 0.15.2
generation_config.json CHANGED
@@ -1,7 +1,9 @@
1
  {
2
  "bos_token_id": 0,
 
3
  "eos_token_id": 2,
4
  "forced_eos_token_id": 2,
 
5
  "pad_token_id": 1,
6
- "transformers_version": "4.39.0.dev0"
7
  }
 
1
  {
2
  "bos_token_id": 0,
3
+ "decoder_start_token_id": 0,
4
  "eos_token_id": 2,
5
  "forced_eos_token_id": 2,
6
+ "max_length": 512,
7
  "pad_token_id": 1,
8
+ "transformers_version": "4.40.0.dev0"
9
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:03ef1f024d1be6aead61edbdc17056f08496d889d819cfe6c633b4d0da05f677
3
  size 809070744
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7504e3f46a16625564b7490aeef33fe0ae152340809953544ddb9ac2000e2631
3
  size 809070744
preprocessor_config.json CHANGED
@@ -37,8 +37,8 @@
37
  "processor_class": "DonutProcessor",
38
  "resample": 2,
39
  "rescale_factor": 0.00392156862745098,
40
- "size": [
41
- 720,
42
- 960
43
- ]
44
  }
 
37
  "processor_class": "DonutProcessor",
38
  "resample": 2,
39
  "rescale_factor": 0.00392156862745098,
40
+ "size": {
41
+ "height": 960,
42
+ "width": 720
43
+ }
44
  }
special_tokens_map.json CHANGED
@@ -1,19 +1,7 @@
1
  {
2
  "additional_special_tokens": [
3
- {
4
- "content": "<s>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "</s>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- }
17
  ],
18
  "bos_token": {
19
  "content": "<s>",
 
1
  {
2
  "additional_special_tokens": [
3
+ "</s>",
4
+ "<s>"
 
 
 
 
 
 
 
 
 
 
 
 
5
  ],
6
  "bos_token": {
7
  "content": "<s>",
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -66,19 +66,26 @@
66
  }
67
  },
68
  "additional_special_tokens": [
69
- "<s>",
70
- "</s>"
71
  ],
72
  "bos_token": "<s>",
73
  "clean_up_tokenization_spaces": true,
74
  "cls_token": "<s>",
75
  "eos_token": "</s>",
76
  "mask_token": "<mask>",
 
77
  "model_max_length": 1000000000000000019884624838656,
 
78
  "pad_token": "<pad>",
 
 
79
  "processor_class": "DonutProcessor",
80
  "sep_token": "</s>",
81
  "sp_model_kwargs": {},
 
82
  "tokenizer_class": "XLMRobertaTokenizer",
 
 
83
  "unk_token": "<unk>"
84
  }
 
66
  }
67
  },
68
  "additional_special_tokens": [
69
+ "</s>",
70
+ "<s>"
71
  ],
72
  "bos_token": "<s>",
73
  "clean_up_tokenization_spaces": true,
74
  "cls_token": "<s>",
75
  "eos_token": "</s>",
76
  "mask_token": "<mask>",
77
+ "max_length": 512,
78
  "model_max_length": 1000000000000000019884624838656,
79
+ "pad_to_multiple_of": null,
80
  "pad_token": "<pad>",
81
+ "pad_token_type_id": 0,
82
+ "padding_side": "right",
83
  "processor_class": "DonutProcessor",
84
  "sep_token": "</s>",
85
  "sp_model_kwargs": {},
86
+ "stride": 0,
87
  "tokenizer_class": "XLMRobertaTokenizer",
88
+ "truncation_side": "right",
89
+ "truncation_strategy": "longest_first",
90
  "unk_token": "<unk>"
91
  }