channudam commited on
Commit
ac3318f
1 Parent(s): 3b249e3

End of training

Browse files
README.md CHANGED
@@ -1,27 +1,21 @@
1
  ---
2
  tags:
3
  - generated_from_trainer
4
- - trocr
5
  model-index:
6
  - name: khmer-trocr-base-printed
7
  results: []
8
- license: mit
9
- language:
10
- - km
11
- library_name: transformers
12
- pipeline_tag: image-to-text
13
  ---
14
 
15
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
16
  should probably proofread and complete it, then remove this comment. -->
17
 
18
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/rayranger/huggingface/runs/hpngjitd)
19
  # khmer-trocr-base-printed
20
 
21
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 0.4627
24
- - Cer: 0.7359
25
 
26
  ## Model description
27
 
@@ -46,24 +40,23 @@ The following hyperparameters were used during training:
46
  - seed: 42
47
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
48
  - lr_scheduler_type: linear
49
- - num_epochs: 3
50
  - mixed_precision_training: Native AMP
51
 
52
  ### Training results
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Cer |
55
  |:-------------:|:------:|:-----:|:---------------:|:------:|
56
- | 1.0771 | 0.2612 | 1000 | 1.0644 | 0.8631 |
57
- | 1.0182 | 0.5223 | 2000 | 0.9849 | 0.8791 |
58
- | 0.9409 | 0.7835 | 3000 | 0.9387 | 0.8704 |
59
- | 0.9053 | 1.0447 | 4000 | 0.8917 | 0.8646 |
60
- | 0.8428 | 1.3058 | 5000 | 0.8280 | 0.8600 |
61
- | 0.729 | 1.5670 | 6000 | 0.7200 | 0.8192 |
62
- | 0.6011 | 1.8282 | 7000 | 0.6203 | 0.7893 |
63
- | 0.5268 | 2.0893 | 8000 | 0.5568 | 0.7621 |
64
- | 0.4819 | 2.3505 | 9000 | 0.5150 | 0.7554 |
65
- | 0.4362 | 2.6116 | 10000 | 0.4799 | 0.7412 |
66
- | 0.4555 | 2.8728 | 11000 | 0.4627 | 0.7359 |
67
 
68
 
69
  ### Framework versions
@@ -71,4 +64,4 @@ The following hyperparameters were used during training:
71
  - Transformers 4.42.3
72
  - Pytorch 2.1.2
73
  - Datasets 2.20.0
74
- - Tokenizers 0.19.1
 
1
  ---
2
  tags:
3
  - generated_from_trainer
 
4
  model-index:
5
  - name: khmer-trocr-base-printed
6
  results: []
 
 
 
 
 
7
  ---
8
 
9
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
10
  should probably proofread and complete it, then remove this comment. -->
11
 
12
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/rayranger/huggingface/runs/hddo3082)
13
  # khmer-trocr-base-printed
14
 
15
  This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
16
  It achieves the following results on the evaluation set:
17
+ - Loss: 0.1980
18
+ - Cer: 0.5955
19
 
20
  ## Model description
21
 
 
40
  - seed: 42
41
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
42
  - lr_scheduler_type: linear
43
+ - num_epochs: 5
44
  - mixed_precision_training: Native AMP
45
 
46
  ### Training results
47
 
48
  | Training Loss | Epoch | Step | Validation Loss | Cer |
49
  |:-------------:|:------:|:-----:|:---------------:|:------:|
50
+ | 1.1417 | 0.4876 | 1000 | 1.0816 | 0.8567 |
51
+ | 1.031 | 0.9751 | 2000 | 0.9824 | 0.8652 |
52
+ | 0.9015 | 1.4627 | 3000 | 0.8875 | 0.8421 |
53
+ | 0.7111 | 1.9503 | 4000 | 0.6645 | 0.7871 |
54
+ | 0.5049 | 2.4378 | 5000 | 0.4831 | 0.7234 |
55
+ | 0.4108 | 2.9254 | 6000 | 0.3594 | 0.6712 |
56
+ | 0.2876 | 3.4130 | 7000 | 0.3076 | 0.6458 |
57
+ | 0.2163 | 3.9005 | 8000 | 0.2418 | 0.6214 |
58
+ | 0.1862 | 4.3881 | 9000 | 0.2119 | 0.5998 |
59
+ | 0.1933 | 4.8757 | 10000 | 0.1980 | 0.5955 |
 
60
 
61
 
62
  ### Framework versions
 
64
  - Transformers 4.42.3
65
  - Pytorch 2.1.2
66
  - Datasets 2.20.0
67
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -3,7 +3,7 @@
3
  "VisionEncoderDecoderModel"
4
  ],
5
  "decoder": {
6
- "_name_or_path": "channudam/khmer-roberta-base",
7
  "add_cross_attention": true,
8
  "architectures": [
9
  "RobertaForMaskedLM"
@@ -85,7 +85,7 @@
85
  "decoder_start_token_id": 0,
86
  "early_stopping": true,
87
  "encoder": {
88
- "_name_or_path": "facebook/deit-base-distilled-patch16-224",
89
  "add_cross_attention": false,
90
  "architectures": [
91
  "DeiTForImageClassificationWithTeacher"
@@ -1112,7 +1112,7 @@
1112
  "998": "ear, spike, capitulum",
1113
  "999": "toilet tissue, toilet paper, bathroom tissue"
1114
  },
1115
- "image_size": 224,
1116
  "initializer_range": 0.02,
1117
  "intermediate_size": 3072,
1118
  "is_decoder": false,
@@ -2161,7 +2161,7 @@
2161
  "eos_token_id": 264,
2162
  "is_encoder_decoder": true,
2163
  "length_penalty": 2.0,
2164
- "max_length": 128,
2165
  "model_type": "vision-encoder-decoder",
2166
  "no_repeat_ngram_size": 3,
2167
  "num_beams": 4,
 
3
  "VisionEncoderDecoderModel"
4
  ],
5
  "decoder": {
6
+ "_name_or_path": "/kaggle/input/khmerrobertamlm/KhmerRobertaMLM",
7
  "add_cross_attention": true,
8
  "architectures": [
9
  "RobertaForMaskedLM"
 
85
  "decoder_start_token_id": 0,
86
  "early_stopping": true,
87
  "encoder": {
88
+ "_name_or_path": "facebook/deit-base-distilled-patch16-384",
89
  "add_cross_attention": false,
90
  "architectures": [
91
  "DeiTForImageClassificationWithTeacher"
 
1112
  "998": "ear, spike, capitulum",
1113
  "999": "toilet tissue, toilet paper, bathroom tissue"
1114
  },
1115
+ "image_size": 384,
1116
  "initializer_range": 0.02,
1117
  "intermediate_size": 3072,
1118
  "is_decoder": false,
 
2161
  "eos_token_id": 264,
2162
  "is_encoder_decoder": true,
2163
  "length_penalty": 2.0,
2164
+ "max_length": 145,
2165
  "model_type": "vision-encoder-decoder",
2166
  "no_repeat_ngram_size": 3,
2167
  "num_beams": 4,
generation_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "early_stopping": true,
5
  "eos_token_id": 264,
6
  "length_penalty": 2.0,
7
- "max_length": 128,
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 1,
 
4
  "early_stopping": true,
5
  "eos_token_id": 264,
6
  "length_penalty": 2.0,
7
+ "max_length": 145,
8
  "no_repeat_ngram_size": 3,
9
  "num_beams": 4,
10
  "pad_token_id": 1,
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:55cbbf1b978618866b4a4f2ecdd35c982f1fb5c1a3607bc7f09bfcea817f06b2
3
- size 607168248
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:91c126cc91712d1c91445786293cf542cbd0ba9ac2d40c950a5d640885803615
3
+ size 608335608
tokenizer_config.json CHANGED
@@ -48,8 +48,8 @@
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
- "max_len": 128,
52
- "model_max_length": 128,
53
  "pad_token": "<pad>",
54
  "sep_token": "</s>",
55
  "tokenizer_class": "RobertaTokenizer",
 
48
  "eos_token": "</s>",
49
  "errors": "replace",
50
  "mask_token": "<mask>",
51
+ "max_len": 150,
52
+ "model_max_length": 150,
53
  "pad_token": "<pad>",
54
  "sep_token": "</s>",
55
  "tokenizer_class": "RobertaTokenizer",
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b4d9857cda86526b31e08dfdbae7f5e3b6f0f0537ae19f0407f84c6ebd8097b9
3
  size 5240
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:016c3c28dbc67f7a85bbc1e4279553dae73e15da096e6a19b0186beb0a625b8f
3
  size 5240