pritamdeka commited on
Commit
02abbe2
1 Parent(s): feb67c8

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,75 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ - bgc
5
+ license: apache-2.0
6
+ base_model: google/mt5-base
7
+ tags:
8
+ - generated_from_trainer
9
+ metrics:
10
+ - bleu
11
+ model-index:
12
+ - name: tst-translation
13
+ results: []
14
+ ---
15
+
16
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
17
+ should probably proofread and complete it, then remove this comment. -->
18
+
19
+ # tst-translation
20
+
21
+ This model is a fine-tuned version of [google/mt5-base](https://huggingface.co/google/mt5-base) on an unknown dataset.
22
+ It achieves the following results on the evaluation set:
23
+ - Loss: 3.1206
24
+ - Bleu: 7.9168
25
+ - Gen Len: 51.4945
26
+
27
+ ## Model description
28
+
29
+ More information needed
30
+
31
+ ## Intended uses & limitations
32
+
33
+ More information needed
34
+
35
+ ## Training and evaluation data
36
+
37
+ More information needed
38
+
39
+ ## Training procedure
40
+
41
+ ### Training hyperparameters
42
+
43
+ The following hyperparameters were used during training:
44
+ - learning_rate: 0.0005
45
+ - train_batch_size: 8
46
+ - eval_batch_size: 16
47
+ - seed: 42
48
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
+ - lr_scheduler_type: linear
50
+ - num_epochs: 20.0
51
+
52
+ ### Training results
53
+
54
+ | Training Loss | Epoch | Step | Validation Loss | Bleu | Gen Len |
55
+ |:-------------:|:-------:|:----:|:---------------:|:------:|:--------:|
56
+ | 4.5126 | 1.5748 | 200 | 2.7474 | 1.0891 | 120.8315 |
57
+ | 2.4414 | 3.1496 | 400 | 2.5120 | 5.0172 | 54.7623 |
58
+ | 1.724 | 4.7244 | 600 | 2.4090 | 5.2115 | 61.985 |
59
+ | 1.2755 | 6.2992 | 800 | 2.5519 | 7.0203 | 46.1414 |
60
+ | 0.97 | 7.8740 | 1000 | 2.5975 | 7.005 | 56.5266 |
61
+ | 0.7251 | 9.4488 | 1200 | 2.7918 | 7.6235 | 52.6841 |
62
+ | 0.584 | 11.0236 | 1400 | 2.8952 | 7.3273 | 49.9659 |
63
+ | 0.4358 | 12.5984 | 1600 | 3.1206 | 7.9168 | 51.4945 |
64
+ | 0.3619 | 14.1732 | 1800 | 3.2645 | 7.9096 | 50.5517 |
65
+ | 0.2933 | 15.7480 | 2000 | 3.3971 | 7.9015 | 49.6169 |
66
+ | 0.2447 | 17.3228 | 2200 | 3.5039 | 7.8441 | 49.6911 |
67
+ | 0.2151 | 18.8976 | 2400 | 3.5556 | 7.8198 | 50.1153 |
68
+
69
+
70
+ ### Framework versions
71
+
72
+ - Transformers 4.43.0.dev0
73
+ - Pytorch 2.3.0+cu121
74
+ - Datasets 2.20.0
75
+ - Tokenizers 0.19.1
all_results.json ADDED
@@ -0,0 +1,16 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_bleu": 7.9168,
4
+ "eval_gen_len": 51.4945,
5
+ "eval_loss": 3.120616912841797,
6
+ "eval_runtime": 243.0735,
7
+ "eval_samples": 997,
8
+ "eval_samples_per_second": 4.102,
9
+ "eval_steps_per_second": 0.259,
10
+ "total_flos": 2811538357714944.0,
11
+ "train_loss": 1.096170819650485,
12
+ "train_runtime": 5288.9286,
13
+ "train_samples": 1012,
14
+ "train_samples_per_second": 3.827,
15
+ "train_steps_per_second": 0.48
16
+ }
config.json ADDED
@@ -0,0 +1,33 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "google/mt5-base",
3
+ "architectures": [
4
+ "MT5ForConditionalGeneration"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 768,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "gelu_new",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "gated-gelu",
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": true,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "model_type": "mt5",
20
+ "num_decoder_layers": 12,
21
+ "num_heads": 12,
22
+ "num_layers": 12,
23
+ "output_past": true,
24
+ "pad_token_id": 0,
25
+ "relative_attention_max_distance": 128,
26
+ "relative_attention_num_buckets": 32,
27
+ "tie_word_embeddings": false,
28
+ "tokenizer_class": "T5Tokenizer",
29
+ "torch_dtype": "float32",
30
+ "transformers_version": "4.43.0.dev0",
31
+ "use_cache": true,
32
+ "vocab_size": 250112
33
+ }
eval_results.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "eval_bleu": 7.9168,
4
+ "eval_gen_len": 51.4945,
5
+ "eval_loss": 3.120616912841797,
6
+ "eval_runtime": 243.0735,
7
+ "eval_samples": 997,
8
+ "eval_samples_per_second": 4.102,
9
+ "eval_steps_per_second": 0.259
10
+ }
generation_config.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "decoder_start_token_id": 0,
3
+ "eos_token_id": 1,
4
+ "pad_token_id": 0,
5
+ "transformers_version": "4.43.0.dev0"
6
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8c734d7e90d50a365b3d46a6da0ae953f66eb09d0c201617acdc158141af850f
3
+ size 2329638768
runs/Jul09_12-21-52_015fb6e6cd9b/events.out.tfevents.1720527722.015fb6e6cd9b.16973.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f26470163952b9cab53d10a0d0945f78052db1ee95e16f81af3c1fa0b2a48e5f
3
+ size 12574
runs/Jul09_12-21-52_015fb6e6cd9b/events.out.tfevents.1720533399.015fb6e6cd9b.16973.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8913331e3eef92b0b7170bb369225b8c1a1c0453cc2d26a676915573a7a122c0
3
+ size 458
special_tokens_map.json ADDED
@@ -0,0 +1,23 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "eos_token": {
3
+ "content": "</s>",
4
+ "lstrip": false,
5
+ "normalized": false,
6
+ "rstrip": false,
7
+ "single_word": false
8
+ },
9
+ "pad_token": {
10
+ "content": "<pad>",
11
+ "lstrip": false,
12
+ "normalized": false,
13
+ "rstrip": false,
14
+ "single_word": false
15
+ },
16
+ "unk_token": {
17
+ "content": "<unk>",
18
+ "lstrip": false,
19
+ "normalized": false,
20
+ "rstrip": false,
21
+ "single_word": false
22
+ }
23
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65c2d7defb6472fada8a935bb364ae3433f7451780c8a59ab6b3cfbaadb32608
3
+ size 16349930
tokenizer_config.json ADDED
@@ -0,0 +1,839 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": true,
3
+ "added_tokens_decoder": {
4
+ "0": {
5
+ "content": "<pad>",
6
+ "lstrip": false,
7
+ "normalized": false,
8
+ "rstrip": false,
9
+ "single_word": false,
10
+ "special": true
11
+ },
12
+ "1": {
13
+ "content": "</s>",
14
+ "lstrip": false,
15
+ "normalized": false,
16
+ "rstrip": false,
17
+ "single_word": false,
18
+ "special": true
19
+ },
20
+ "2": {
21
+ "content": "<unk>",
22
+ "lstrip": false,
23
+ "normalized": false,
24
+ "rstrip": false,
25
+ "single_word": false,
26
+ "special": true
27
+ },
28
+ "250000": {
29
+ "content": "▁<extra_id_99>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": false
35
+ },
36
+ "250001": {
37
+ "content": "▁<extra_id_98>",
38
+ "lstrip": false,
39
+ "normalized": false,
40
+ "rstrip": false,
41
+ "single_word": false,
42
+ "special": false
43
+ },
44
+ "250002": {
45
+ "content": "▁<extra_id_97>",
46
+ "lstrip": false,
47
+ "normalized": false,
48
+ "rstrip": false,
49
+ "single_word": false,
50
+ "special": false
51
+ },
52
+ "250003": {
53
+ "content": "▁<extra_id_96>",
54
+ "lstrip": false,
55
+ "normalized": false,
56
+ "rstrip": false,
57
+ "single_word": false,
58
+ "special": false
59
+ },
60
+ "250004": {
61
+ "content": "▁<extra_id_95>",
62
+ "lstrip": false,
63
+ "normalized": false,
64
+ "rstrip": false,
65
+ "single_word": false,
66
+ "special": false
67
+ },
68
+ "250005": {
69
+ "content": "▁<extra_id_94>",
70
+ "lstrip": false,
71
+ "normalized": false,
72
+ "rstrip": false,
73
+ "single_word": false,
74
+ "special": false
75
+ },
76
+ "250006": {
77
+ "content": "▁<extra_id_93>",
78
+ "lstrip": false,
79
+ "normalized": false,
80
+ "rstrip": false,
81
+ "single_word": false,
82
+ "special": false
83
+ },
84
+ "250007": {
85
+ "content": "▁<extra_id_92>",
86
+ "lstrip": false,
87
+ "normalized": false,
88
+ "rstrip": false,
89
+ "single_word": false,
90
+ "special": false
91
+ },
92
+ "250008": {
93
+ "content": "▁<extra_id_91>",
94
+ "lstrip": false,
95
+ "normalized": false,
96
+ "rstrip": false,
97
+ "single_word": false,
98
+ "special": false
99
+ },
100
+ "250009": {
101
+ "content": "▁<extra_id_90>",
102
+ "lstrip": false,
103
+ "normalized": false,
104
+ "rstrip": false,
105
+ "single_word": false,
106
+ "special": false
107
+ },
108
+ "250010": {
109
+ "content": "▁<extra_id_89>",
110
+ "lstrip": false,
111
+ "normalized": false,
112
+ "rstrip": false,
113
+ "single_word": false,
114
+ "special": false
115
+ },
116
+ "250011": {
117
+ "content": "▁<extra_id_88>",
118
+ "lstrip": false,
119
+ "normalized": false,
120
+ "rstrip": false,
121
+ "single_word": false,
122
+ "special": false
123
+ },
124
+ "250012": {
125
+ "content": "▁<extra_id_87>",
126
+ "lstrip": false,
127
+ "normalized": false,
128
+ "rstrip": false,
129
+ "single_word": false,
130
+ "special": false
131
+ },
132
+ "250013": {
133
+ "content": "▁<extra_id_86>",
134
+ "lstrip": false,
135
+ "normalized": false,
136
+ "rstrip": false,
137
+ "single_word": false,
138
+ "special": false
139
+ },
140
+ "250014": {
141
+ "content": "▁<extra_id_85>",
142
+ "lstrip": false,
143
+ "normalized": false,
144
+ "rstrip": false,
145
+ "single_word": false,
146
+ "special": false
147
+ },
148
+ "250015": {
149
+ "content": "▁<extra_id_84>",
150
+ "lstrip": false,
151
+ "normalized": false,
152
+ "rstrip": false,
153
+ "single_word": false,
154
+ "special": false
155
+ },
156
+ "250016": {
157
+ "content": "▁<extra_id_83>",
158
+ "lstrip": false,
159
+ "normalized": false,
160
+ "rstrip": false,
161
+ "single_word": false,
162
+ "special": false
163
+ },
164
+ "250017": {
165
+ "content": "▁<extra_id_82>",
166
+ "lstrip": false,
167
+ "normalized": false,
168
+ "rstrip": false,
169
+ "single_word": false,
170
+ "special": false
171
+ },
172
+ "250018": {
173
+ "content": "▁<extra_id_81>",
174
+ "lstrip": false,
175
+ "normalized": false,
176
+ "rstrip": false,
177
+ "single_word": false,
178
+ "special": false
179
+ },
180
+ "250019": {
181
+ "content": "▁<extra_id_80>",
182
+ "lstrip": false,
183
+ "normalized": false,
184
+ "rstrip": false,
185
+ "single_word": false,
186
+ "special": false
187
+ },
188
+ "250020": {
189
+ "content": "▁<extra_id_79>",
190
+ "lstrip": false,
191
+ "normalized": false,
192
+ "rstrip": false,
193
+ "single_word": false,
194
+ "special": false
195
+ },
196
+ "250021": {
197
+ "content": "▁<extra_id_78>",
198
+ "lstrip": false,
199
+ "normalized": false,
200
+ "rstrip": false,
201
+ "single_word": false,
202
+ "special": false
203
+ },
204
+ "250022": {
205
+ "content": "▁<extra_id_77>",
206
+ "lstrip": false,
207
+ "normalized": false,
208
+ "rstrip": false,
209
+ "single_word": false,
210
+ "special": false
211
+ },
212
+ "250023": {
213
+ "content": "▁<extra_id_76>",
214
+ "lstrip": false,
215
+ "normalized": false,
216
+ "rstrip": false,
217
+ "single_word": false,
218
+ "special": false
219
+ },
220
+ "250024": {
221
+ "content": "▁<extra_id_75>",
222
+ "lstrip": false,
223
+ "normalized": false,
224
+ "rstrip": false,
225
+ "single_word": false,
226
+ "special": false
227
+ },
228
+ "250025": {
229
+ "content": "▁<extra_id_74>",
230
+ "lstrip": false,
231
+ "normalized": false,
232
+ "rstrip": false,
233
+ "single_word": false,
234
+ "special": false
235
+ },
236
+ "250026": {
237
+ "content": "▁<extra_id_73>",
238
+ "lstrip": false,
239
+ "normalized": false,
240
+ "rstrip": false,
241
+ "single_word": false,
242
+ "special": false
243
+ },
244
+ "250027": {
245
+ "content": "▁<extra_id_72>",
246
+ "lstrip": false,
247
+ "normalized": false,
248
+ "rstrip": false,
249
+ "single_word": false,
250
+ "special": false
251
+ },
252
+ "250028": {
253
+ "content": "▁<extra_id_71>",
254
+ "lstrip": false,
255
+ "normalized": false,
256
+ "rstrip": false,
257
+ "single_word": false,
258
+ "special": false
259
+ },
260
+ "250029": {
261
+ "content": "▁<extra_id_70>",
262
+ "lstrip": false,
263
+ "normalized": false,
264
+ "rstrip": false,
265
+ "single_word": false,
266
+ "special": false
267
+ },
268
+ "250030": {
269
+ "content": "▁<extra_id_69>",
270
+ "lstrip": false,
271
+ "normalized": false,
272
+ "rstrip": false,
273
+ "single_word": false,
274
+ "special": false
275
+ },
276
+ "250031": {
277
+ "content": "▁<extra_id_68>",
278
+ "lstrip": false,
279
+ "normalized": false,
280
+ "rstrip": false,
281
+ "single_word": false,
282
+ "special": false
283
+ },
284
+ "250032": {
285
+ "content": "▁<extra_id_67>",
286
+ "lstrip": false,
287
+ "normalized": false,
288
+ "rstrip": false,
289
+ "single_word": false,
290
+ "special": false
291
+ },
292
+ "250033": {
293
+ "content": "▁<extra_id_66>",
294
+ "lstrip": false,
295
+ "normalized": false,
296
+ "rstrip": false,
297
+ "single_word": false,
298
+ "special": false
299
+ },
300
+ "250034": {
301
+ "content": "▁<extra_id_65>",
302
+ "lstrip": false,
303
+ "normalized": false,
304
+ "rstrip": false,
305
+ "single_word": false,
306
+ "special": false
307
+ },
308
+ "250035": {
309
+ "content": "▁<extra_id_64>",
310
+ "lstrip": false,
311
+ "normalized": false,
312
+ "rstrip": false,
313
+ "single_word": false,
314
+ "special": false
315
+ },
316
+ "250036": {
317
+ "content": "▁<extra_id_63>",
318
+ "lstrip": false,
319
+ "normalized": false,
320
+ "rstrip": false,
321
+ "single_word": false,
322
+ "special": false
323
+ },
324
+ "250037": {
325
+ "content": "▁<extra_id_62>",
326
+ "lstrip": false,
327
+ "normalized": false,
328
+ "rstrip": false,
329
+ "single_word": false,
330
+ "special": false
331
+ },
332
+ "250038": {
333
+ "content": "▁<extra_id_61>",
334
+ "lstrip": false,
335
+ "normalized": false,
336
+ "rstrip": false,
337
+ "single_word": false,
338
+ "special": false
339
+ },
340
+ "250039": {
341
+ "content": "▁<extra_id_60>",
342
+ "lstrip": false,
343
+ "normalized": false,
344
+ "rstrip": false,
345
+ "single_word": false,
346
+ "special": false
347
+ },
348
+ "250040": {
349
+ "content": "▁<extra_id_59>",
350
+ "lstrip": false,
351
+ "normalized": false,
352
+ "rstrip": false,
353
+ "single_word": false,
354
+ "special": false
355
+ },
356
+ "250041": {
357
+ "content": "▁<extra_id_58>",
358
+ "lstrip": false,
359
+ "normalized": false,
360
+ "rstrip": false,
361
+ "single_word": false,
362
+ "special": false
363
+ },
364
+ "250042": {
365
+ "content": "▁<extra_id_57>",
366
+ "lstrip": false,
367
+ "normalized": false,
368
+ "rstrip": false,
369
+ "single_word": false,
370
+ "special": false
371
+ },
372
+ "250043": {
373
+ "content": "▁<extra_id_56>",
374
+ "lstrip": false,
375
+ "normalized": false,
376
+ "rstrip": false,
377
+ "single_word": false,
378
+ "special": false
379
+ },
380
+ "250044": {
381
+ "content": "▁<extra_id_55>",
382
+ "lstrip": false,
383
+ "normalized": false,
384
+ "rstrip": false,
385
+ "single_word": false,
386
+ "special": false
387
+ },
388
+ "250045": {
389
+ "content": "▁<extra_id_54>",
390
+ "lstrip": false,
391
+ "normalized": false,
392
+ "rstrip": false,
393
+ "single_word": false,
394
+ "special": false
395
+ },
396
+ "250046": {
397
+ "content": "▁<extra_id_53>",
398
+ "lstrip": false,
399
+ "normalized": false,
400
+ "rstrip": false,
401
+ "single_word": false,
402
+ "special": false
403
+ },
404
+ "250047": {
405
+ "content": "▁<extra_id_52>",
406
+ "lstrip": false,
407
+ "normalized": false,
408
+ "rstrip": false,
409
+ "single_word": false,
410
+ "special": false
411
+ },
412
+ "250048": {
413
+ "content": "▁<extra_id_51>",
414
+ "lstrip": false,
415
+ "normalized": false,
416
+ "rstrip": false,
417
+ "single_word": false,
418
+ "special": false
419
+ },
420
+ "250049": {
421
+ "content": "▁<extra_id_50>",
422
+ "lstrip": false,
423
+ "normalized": false,
424
+ "rstrip": false,
425
+ "single_word": false,
426
+ "special": false
427
+ },
428
+ "250050": {
429
+ "content": "���<extra_id_49>",
430
+ "lstrip": false,
431
+ "normalized": false,
432
+ "rstrip": false,
433
+ "single_word": false,
434
+ "special": false
435
+ },
436
+ "250051": {
437
+ "content": "▁<extra_id_48>",
438
+ "lstrip": false,
439
+ "normalized": false,
440
+ "rstrip": false,
441
+ "single_word": false,
442
+ "special": false
443
+ },
444
+ "250052": {
445
+ "content": "▁<extra_id_47>",
446
+ "lstrip": false,
447
+ "normalized": false,
448
+ "rstrip": false,
449
+ "single_word": false,
450
+ "special": false
451
+ },
452
+ "250053": {
453
+ "content": "▁<extra_id_46>",
454
+ "lstrip": false,
455
+ "normalized": false,
456
+ "rstrip": false,
457
+ "single_word": false,
458
+ "special": false
459
+ },
460
+ "250054": {
461
+ "content": "▁<extra_id_45>",
462
+ "lstrip": false,
463
+ "normalized": false,
464
+ "rstrip": false,
465
+ "single_word": false,
466
+ "special": false
467
+ },
468
+ "250055": {
469
+ "content": "▁<extra_id_44>",
470
+ "lstrip": false,
471
+ "normalized": false,
472
+ "rstrip": false,
473
+ "single_word": false,
474
+ "special": false
475
+ },
476
+ "250056": {
477
+ "content": "▁<extra_id_43>",
478
+ "lstrip": false,
479
+ "normalized": false,
480
+ "rstrip": false,
481
+ "single_word": false,
482
+ "special": false
483
+ },
484
+ "250057": {
485
+ "content": "▁<extra_id_42>",
486
+ "lstrip": false,
487
+ "normalized": false,
488
+ "rstrip": false,
489
+ "single_word": false,
490
+ "special": false
491
+ },
492
+ "250058": {
493
+ "content": "▁<extra_id_41>",
494
+ "lstrip": false,
495
+ "normalized": false,
496
+ "rstrip": false,
497
+ "single_word": false,
498
+ "special": false
499
+ },
500
+ "250059": {
501
+ "content": "▁<extra_id_40>",
502
+ "lstrip": false,
503
+ "normalized": false,
504
+ "rstrip": false,
505
+ "single_word": false,
506
+ "special": false
507
+ },
508
+ "250060": {
509
+ "content": "▁<extra_id_39>",
510
+ "lstrip": false,
511
+ "normalized": false,
512
+ "rstrip": false,
513
+ "single_word": false,
514
+ "special": false
515
+ },
516
+ "250061": {
517
+ "content": "▁<extra_id_38>",
518
+ "lstrip": false,
519
+ "normalized": false,
520
+ "rstrip": false,
521
+ "single_word": false,
522
+ "special": false
523
+ },
524
+ "250062": {
525
+ "content": "▁<extra_id_37>",
526
+ "lstrip": false,
527
+ "normalized": false,
528
+ "rstrip": false,
529
+ "single_word": false,
530
+ "special": false
531
+ },
532
+ "250063": {
533
+ "content": "▁<extra_id_36>",
534
+ "lstrip": false,
535
+ "normalized": false,
536
+ "rstrip": false,
537
+ "single_word": false,
538
+ "special": false
539
+ },
540
+ "250064": {
541
+ "content": "▁<extra_id_35>",
542
+ "lstrip": false,
543
+ "normalized": false,
544
+ "rstrip": false,
545
+ "single_word": false,
546
+ "special": false
547
+ },
548
+ "250065": {
549
+ "content": "▁<extra_id_34>",
550
+ "lstrip": false,
551
+ "normalized": false,
552
+ "rstrip": false,
553
+ "single_word": false,
554
+ "special": false
555
+ },
556
+ "250066": {
557
+ "content": "▁<extra_id_33>",
558
+ "lstrip": false,
559
+ "normalized": false,
560
+ "rstrip": false,
561
+ "single_word": false,
562
+ "special": false
563
+ },
564
+ "250067": {
565
+ "content": "▁<extra_id_32>",
566
+ "lstrip": false,
567
+ "normalized": false,
568
+ "rstrip": false,
569
+ "single_word": false,
570
+ "special": false
571
+ },
572
+ "250068": {
573
+ "content": "▁<extra_id_31>",
574
+ "lstrip": false,
575
+ "normalized": false,
576
+ "rstrip": false,
577
+ "single_word": false,
578
+ "special": false
579
+ },
580
+ "250069": {
581
+ "content": "▁<extra_id_30>",
582
+ "lstrip": false,
583
+ "normalized": false,
584
+ "rstrip": false,
585
+ "single_word": false,
586
+ "special": false
587
+ },
588
+ "250070": {
589
+ "content": "▁<extra_id_29>",
590
+ "lstrip": false,
591
+ "normalized": false,
592
+ "rstrip": false,
593
+ "single_word": false,
594
+ "special": false
595
+ },
596
+ "250071": {
597
+ "content": "▁<extra_id_28>",
598
+ "lstrip": false,
599
+ "normalized": false,
600
+ "rstrip": false,
601
+ "single_word": false,
602
+ "special": false
603
+ },
604
+ "250072": {
605
+ "content": "▁<extra_id_27>",
606
+ "lstrip": false,
607
+ "normalized": false,
608
+ "rstrip": false,
609
+ "single_word": false,
610
+ "special": false
611
+ },
612
+ "250073": {
613
+ "content": "▁<extra_id_26>",
614
+ "lstrip": false,
615
+ "normalized": false,
616
+ "rstrip": false,
617
+ "single_word": false,
618
+ "special": false
619
+ },
620
+ "250074": {
621
+ "content": "▁<extra_id_25>",
622
+ "lstrip": false,
623
+ "normalized": false,
624
+ "rstrip": false,
625
+ "single_word": false,
626
+ "special": false
627
+ },
628
+ "250075": {
629
+ "content": "▁<extra_id_24>",
630
+ "lstrip": false,
631
+ "normalized": false,
632
+ "rstrip": false,
633
+ "single_word": false,
634
+ "special": false
635
+ },
636
+ "250076": {
637
+ "content": "▁<extra_id_23>",
638
+ "lstrip": false,
639
+ "normalized": false,
640
+ "rstrip": false,
641
+ "single_word": false,
642
+ "special": false
643
+ },
644
+ "250077": {
645
+ "content": "▁<extra_id_22>",
646
+ "lstrip": false,
647
+ "normalized": false,
648
+ "rstrip": false,
649
+ "single_word": false,
650
+ "special": false
651
+ },
652
+ "250078": {
653
+ "content": "▁<extra_id_21>",
654
+ "lstrip": false,
655
+ "normalized": false,
656
+ "rstrip": false,
657
+ "single_word": false,
658
+ "special": false
659
+ },
660
+ "250079": {
661
+ "content": "▁<extra_id_20>",
662
+ "lstrip": false,
663
+ "normalized": false,
664
+ "rstrip": false,
665
+ "single_word": false,
666
+ "special": false
667
+ },
668
+ "250080": {
669
+ "content": "▁<extra_id_19>",
670
+ "lstrip": false,
671
+ "normalized": false,
672
+ "rstrip": false,
673
+ "single_word": false,
674
+ "special": false
675
+ },
676
+ "250081": {
677
+ "content": "▁<extra_id_18>",
678
+ "lstrip": false,
679
+ "normalized": false,
680
+ "rstrip": false,
681
+ "single_word": false,
682
+ "special": false
683
+ },
684
+ "250082": {
685
+ "content": "▁<extra_id_17>",
686
+ "lstrip": false,
687
+ "normalized": false,
688
+ "rstrip": false,
689
+ "single_word": false,
690
+ "special": false
691
+ },
692
+ "250083": {
693
+ "content": "▁<extra_id_16>",
694
+ "lstrip": false,
695
+ "normalized": false,
696
+ "rstrip": false,
697
+ "single_word": false,
698
+ "special": false
699
+ },
700
+ "250084": {
701
+ "content": "▁<extra_id_15>",
702
+ "lstrip": false,
703
+ "normalized": false,
704
+ "rstrip": false,
705
+ "single_word": false,
706
+ "special": false
707
+ },
708
+ "250085": {
709
+ "content": "▁<extra_id_14>",
710
+ "lstrip": false,
711
+ "normalized": false,
712
+ "rstrip": false,
713
+ "single_word": false,
714
+ "special": false
715
+ },
716
+ "250086": {
717
+ "content": "▁<extra_id_13>",
718
+ "lstrip": false,
719
+ "normalized": false,
720
+ "rstrip": false,
721
+ "single_word": false,
722
+ "special": false
723
+ },
724
+ "250087": {
725
+ "content": "▁<extra_id_12>",
726
+ "lstrip": false,
727
+ "normalized": false,
728
+ "rstrip": false,
729
+ "single_word": false,
730
+ "special": false
731
+ },
732
+ "250088": {
733
+ "content": "▁<extra_id_11>",
734
+ "lstrip": false,
735
+ "normalized": false,
736
+ "rstrip": false,
737
+ "single_word": false,
738
+ "special": false
739
+ },
740
+ "250089": {
741
+ "content": "▁<extra_id_10>",
742
+ "lstrip": false,
743
+ "normalized": false,
744
+ "rstrip": false,
745
+ "single_word": false,
746
+ "special": false
747
+ },
748
+ "250090": {
749
+ "content": "▁<extra_id_9>",
750
+ "lstrip": false,
751
+ "normalized": false,
752
+ "rstrip": false,
753
+ "single_word": false,
754
+ "special": false
755
+ },
756
+ "250091": {
757
+ "content": "▁<extra_id_8>",
758
+ "lstrip": false,
759
+ "normalized": false,
760
+ "rstrip": false,
761
+ "single_word": false,
762
+ "special": false
763
+ },
764
+ "250092": {
765
+ "content": "▁<extra_id_7>",
766
+ "lstrip": false,
767
+ "normalized": false,
768
+ "rstrip": false,
769
+ "single_word": false,
770
+ "special": false
771
+ },
772
+ "250093": {
773
+ "content": "▁<extra_id_6>",
774
+ "lstrip": false,
775
+ "normalized": false,
776
+ "rstrip": false,
777
+ "single_word": false,
778
+ "special": false
779
+ },
780
+ "250094": {
781
+ "content": "▁<extra_id_5>",
782
+ "lstrip": false,
783
+ "normalized": false,
784
+ "rstrip": false,
785
+ "single_word": false,
786
+ "special": false
787
+ },
788
+ "250095": {
789
+ "content": "▁<extra_id_4>",
790
+ "lstrip": false,
791
+ "normalized": false,
792
+ "rstrip": false,
793
+ "single_word": false,
794
+ "special": false
795
+ },
796
+ "250096": {
797
+ "content": "▁<extra_id_3>",
798
+ "lstrip": false,
799
+ "normalized": false,
800
+ "rstrip": false,
801
+ "single_word": false,
802
+ "special": false
803
+ },
804
+ "250097": {
805
+ "content": "▁<extra_id_2>",
806
+ "lstrip": false,
807
+ "normalized": false,
808
+ "rstrip": false,
809
+ "single_word": false,
810
+ "special": false
811
+ },
812
+ "250098": {
813
+ "content": "▁<extra_id_1>",
814
+ "lstrip": false,
815
+ "normalized": false,
816
+ "rstrip": false,
817
+ "single_word": false,
818
+ "special": false
819
+ },
820
+ "250099": {
821
+ "content": "▁<extra_id_0>",
822
+ "lstrip": false,
823
+ "normalized": false,
824
+ "rstrip": false,
825
+ "single_word": false,
826
+ "special": false
827
+ }
828
+ },
829
+ "additional_special_tokens": [],
830
+ "clean_up_tokenization_spaces": true,
831
+ "eos_token": "</s>",
832
+ "extra_ids": 0,
833
+ "legacy": true,
834
+ "model_max_length": 1000000000000000019884624838656,
835
+ "pad_token": "<pad>",
836
+ "sp_model_kwargs": {},
837
+ "tokenizer_class": "T5Tokenizer",
838
+ "unk_token": "<unk>"
839
+ }
train_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 20.0,
3
+ "total_flos": 2811538357714944.0,
4
+ "train_loss": 1.096170819650485,
5
+ "train_runtime": 5288.9286,
6
+ "train_samples": 1012,
7
+ "train_samples_per_second": 3.827,
8
+ "train_steps_per_second": 0.48
9
+ }
trainer_state.json ADDED
@@ -0,0 +1,246 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "best_metric": 7.9168,
3
+ "best_model_checkpoint": "/content/tst-translation/checkpoint-1600",
4
+ "epoch": 20.0,
5
+ "eval_steps": 200,
6
+ "global_step": 2540,
7
+ "is_hyper_param_search": false,
8
+ "is_local_process_zero": true,
9
+ "is_world_process_zero": true,
10
+ "log_history": [
11
+ {
12
+ "epoch": 1.574803149606299,
13
+ "grad_norm": 2.453336238861084,
14
+ "learning_rate": 0.00046062992125984255,
15
+ "loss": 4.5126,
16
+ "step": 200
17
+ },
18
+ {
19
+ "epoch": 1.574803149606299,
20
+ "eval_bleu": 1.0891,
21
+ "eval_gen_len": 120.8315,
22
+ "eval_loss": 2.747382402420044,
23
+ "eval_runtime": 505.8332,
24
+ "eval_samples_per_second": 1.971,
25
+ "eval_steps_per_second": 0.125,
26
+ "step": 200
27
+ },
28
+ {
29
+ "epoch": 3.1496062992125986,
30
+ "grad_norm": 2.0367627143859863,
31
+ "learning_rate": 0.00042125984251968504,
32
+ "loss": 2.4414,
33
+ "step": 400
34
+ },
35
+ {
36
+ "epoch": 3.1496062992125986,
37
+ "eval_bleu": 5.0172,
38
+ "eval_gen_len": 54.7623,
39
+ "eval_loss": 2.5120186805725098,
40
+ "eval_runtime": 340.8894,
41
+ "eval_samples_per_second": 2.925,
42
+ "eval_steps_per_second": 0.185,
43
+ "step": 400
44
+ },
45
+ {
46
+ "epoch": 4.724409448818898,
47
+ "grad_norm": 2.0367233753204346,
48
+ "learning_rate": 0.00038188976377952753,
49
+ "loss": 1.724,
50
+ "step": 600
51
+ },
52
+ {
53
+ "epoch": 4.724409448818898,
54
+ "eval_bleu": 5.2115,
55
+ "eval_gen_len": 61.985,
56
+ "eval_loss": 2.4089620113372803,
57
+ "eval_runtime": 368.5588,
58
+ "eval_samples_per_second": 2.705,
59
+ "eval_steps_per_second": 0.171,
60
+ "step": 600
61
+ },
62
+ {
63
+ "epoch": 6.299212598425197,
64
+ "grad_norm": 1.8066316843032837,
65
+ "learning_rate": 0.00034251968503937007,
66
+ "loss": 1.2755,
67
+ "step": 800
68
+ },
69
+ {
70
+ "epoch": 6.299212598425197,
71
+ "eval_bleu": 7.0203,
72
+ "eval_gen_len": 46.1414,
73
+ "eval_loss": 2.551859140396118,
74
+ "eval_runtime": 215.1683,
75
+ "eval_samples_per_second": 4.634,
76
+ "eval_steps_per_second": 0.293,
77
+ "step": 800
78
+ },
79
+ {
80
+ "epoch": 7.874015748031496,
81
+ "grad_norm": 1.8381917476654053,
82
+ "learning_rate": 0.0003031496062992126,
83
+ "loss": 0.97,
84
+ "step": 1000
85
+ },
86
+ {
87
+ "epoch": 7.874015748031496,
88
+ "eval_bleu": 7.005,
89
+ "eval_gen_len": 56.5266,
90
+ "eval_loss": 2.5974559783935547,
91
+ "eval_runtime": 308.9884,
92
+ "eval_samples_per_second": 3.227,
93
+ "eval_steps_per_second": 0.204,
94
+ "step": 1000
95
+ },
96
+ {
97
+ "epoch": 9.448818897637794,
98
+ "grad_norm": 1.8475762605667114,
99
+ "learning_rate": 0.0002637795275590551,
100
+ "loss": 0.7251,
101
+ "step": 1200
102
+ },
103
+ {
104
+ "epoch": 9.448818897637794,
105
+ "eval_bleu": 7.6235,
106
+ "eval_gen_len": 52.6841,
107
+ "eval_loss": 2.791808605194092,
108
+ "eval_runtime": 280.2999,
109
+ "eval_samples_per_second": 3.557,
110
+ "eval_steps_per_second": 0.225,
111
+ "step": 1200
112
+ },
113
+ {
114
+ "epoch": 11.023622047244094,
115
+ "grad_norm": 1.7481825351715088,
116
+ "learning_rate": 0.00022440944881889764,
117
+ "loss": 0.584,
118
+ "step": 1400
119
+ },
120
+ {
121
+ "epoch": 11.023622047244094,
122
+ "eval_bleu": 7.3273,
123
+ "eval_gen_len": 49.9659,
124
+ "eval_loss": 2.8952395915985107,
125
+ "eval_runtime": 205.6302,
126
+ "eval_samples_per_second": 4.849,
127
+ "eval_steps_per_second": 0.306,
128
+ "step": 1400
129
+ },
130
+ {
131
+ "epoch": 12.598425196850394,
132
+ "grad_norm": 1.5020047426223755,
133
+ "learning_rate": 0.00018503937007874016,
134
+ "loss": 0.4358,
135
+ "step": 1600
136
+ },
137
+ {
138
+ "epoch": 12.598425196850394,
139
+ "eval_bleu": 7.9168,
140
+ "eval_gen_len": 51.4945,
141
+ "eval_loss": 3.120616912841797,
142
+ "eval_runtime": 243.389,
143
+ "eval_samples_per_second": 4.096,
144
+ "eval_steps_per_second": 0.259,
145
+ "step": 1600
146
+ },
147
+ {
148
+ "epoch": 14.173228346456693,
149
+ "grad_norm": 1.6936888694763184,
150
+ "learning_rate": 0.00014566929133858267,
151
+ "loss": 0.3619,
152
+ "step": 1800
153
+ },
154
+ {
155
+ "epoch": 14.173228346456693,
156
+ "eval_bleu": 7.9096,
157
+ "eval_gen_len": 50.5517,
158
+ "eval_loss": 3.264512538909912,
159
+ "eval_runtime": 220.31,
160
+ "eval_samples_per_second": 4.525,
161
+ "eval_steps_per_second": 0.286,
162
+ "step": 1800
163
+ },
164
+ {
165
+ "epoch": 15.748031496062993,
166
+ "grad_norm": 1.4287927150726318,
167
+ "learning_rate": 0.0001062992125984252,
168
+ "loss": 0.2933,
169
+ "step": 2000
170
+ },
171
+ {
172
+ "epoch": 15.748031496062993,
173
+ "eval_bleu": 7.9015,
174
+ "eval_gen_len": 49.6169,
175
+ "eval_loss": 3.3970730304718018,
176
+ "eval_runtime": 195.9873,
177
+ "eval_samples_per_second": 5.087,
178
+ "eval_steps_per_second": 0.321,
179
+ "step": 2000
180
+ },
181
+ {
182
+ "epoch": 17.322834645669293,
183
+ "grad_norm": 1.6358137130737305,
184
+ "learning_rate": 6.692913385826773e-05,
185
+ "loss": 0.2447,
186
+ "step": 2200
187
+ },
188
+ {
189
+ "epoch": 17.322834645669293,
190
+ "eval_bleu": 7.8441,
191
+ "eval_gen_len": 49.6911,
192
+ "eval_loss": 3.5039305686950684,
193
+ "eval_runtime": 201.4971,
194
+ "eval_samples_per_second": 4.948,
195
+ "eval_steps_per_second": 0.313,
196
+ "step": 2200
197
+ },
198
+ {
199
+ "epoch": 18.89763779527559,
200
+ "grad_norm": 1.3624520301818848,
201
+ "learning_rate": 2.7559055118110236e-05,
202
+ "loss": 0.2151,
203
+ "step": 2400
204
+ },
205
+ {
206
+ "epoch": 18.89763779527559,
207
+ "eval_bleu": 7.8198,
208
+ "eval_gen_len": 50.1153,
209
+ "eval_loss": 3.5556399822235107,
210
+ "eval_runtime": 195.3596,
211
+ "eval_samples_per_second": 5.103,
212
+ "eval_steps_per_second": 0.322,
213
+ "step": 2400
214
+ },
215
+ {
216
+ "epoch": 20.0,
217
+ "step": 2540,
218
+ "total_flos": 2811538357714944.0,
219
+ "train_loss": 1.096170819650485,
220
+ "train_runtime": 5288.9286,
221
+ "train_samples_per_second": 3.827,
222
+ "train_steps_per_second": 0.48
223
+ }
224
+ ],
225
+ "logging_steps": 200,
226
+ "max_steps": 2540,
227
+ "num_input_tokens_seen": 0,
228
+ "num_train_epochs": 20,
229
+ "save_steps": 200,
230
+ "stateful_callbacks": {
231
+ "TrainerControl": {
232
+ "args": {
233
+ "should_epoch_stop": false,
234
+ "should_evaluate": false,
235
+ "should_log": false,
236
+ "should_save": true,
237
+ "should_training_stop": true
238
+ },
239
+ "attributes": {}
240
+ }
241
+ },
242
+ "total_flos": 2811538357714944.0,
243
+ "train_batch_size": 8,
244
+ "trial_name": null,
245
+ "trial_params": null
246
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e1abb3eb8b98db22a27b1cc92a1026ac087bd9375c5d0aaa99df0c75d8ed5d9a
3
+ size 5368