balaramas commited on
Commit
f21c78c
β€’
1 Parent(s): b1ff380

Training in progress, step 500

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
README.md ADDED
@@ -0,0 +1,47 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: csebuetnlp/mT5_m2o_hindi_crossSum
3
+ tags:
4
+ - generated_from_trainer
5
+ model-index:
6
+ - name: marian-t5
7
+ results: []
8
+ ---
9
+
10
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
+ should probably proofread and complete it, then remove this comment. -->
12
+
13
+ # marian-t5
14
+
15
+ This model is a fine-tuned version of [csebuetnlp/mT5_m2o_hindi_crossSum](https://huggingface.co/csebuetnlp/mT5_m2o_hindi_crossSum) on an unknown dataset.
16
+
17
+ ## Model description
18
+
19
+ More information needed
20
+
21
+ ## Intended uses & limitations
22
+
23
+ More information needed
24
+
25
+ ## Training and evaluation data
26
+
27
+ More information needed
28
+
29
+ ## Training procedure
30
+
31
+ ### Training hyperparameters
32
+
33
+ The following hyperparameters were used during training:
34
+ - learning_rate: 5e-05
35
+ - train_batch_size: 32
36
+ - eval_batch_size: 32
37
+ - seed: 42
38
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
39
+ - lr_scheduler_type: linear
40
+ - num_epochs: 3.0
41
+
42
+ ### Framework versions
43
+
44
+ - Transformers 4.37.2
45
+ - Pytorch 2.1.2+cu121
46
+ - Datasets 2.16.1
47
+ - Tokenizers 0.15.1
all_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "predict_gen_len": 52.7633955448525,
3
+ "predict_loss": 3.9636030197143555,
4
+ "predict_rouge1": 0.1906,
5
+ "predict_rouge2": 0.0401,
6
+ "predict_rougeL": 0.1886,
7
+ "predict_rougeLsum": 0.1886,
8
+ "predict_runtime": 234.4913,
9
+ "predict_samples": 1661,
10
+ "predict_samples_per_second": 7.083,
11
+ "predict_steps_per_second": 0.222
12
+ }
config.json ADDED
@@ -0,0 +1,220 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "csebuetnlp/mT5_m2o_hindi_crossSum",
3
+ "architectures": [
4
+ "MT5ForConditionalGeneration"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 768,
10
+ "decoder_start_token_id": 250021,
11
+ "dense_act_fn": "gelu_new",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "gated-gelu",
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": true,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "length_penalty": 0.6,
20
+ "max_length": 84,
21
+ "model_type": "mt5",
22
+ "num_beams": 4,
23
+ "num_decoder_layers": 12,
24
+ "num_heads": 12,
25
+ "num_layers": 12,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "relative_attention_max_distance": 128,
29
+ "relative_attention_num_buckets": 32,
30
+ "task_specific_params": {
31
+ "langid_map": {
32
+ "amharic": [
33
+ 35,
34
+ "\u2581<extra_id_64>"
35
+ ],
36
+ "arabic": [
37
+ 4,
38
+ "\u2581<extra_id_95>"
39
+ ],
40
+ "azerbaijani": [
41
+ 7,
42
+ "\u2581<extra_id_92>"
43
+ ],
44
+ "bengali": [
45
+ 42,
46
+ "\u2581<extra_id_57>"
47
+ ],
48
+ "burmese": [
49
+ 33,
50
+ "\u2581<extra_id_66>"
51
+ ],
52
+ "chinese_simplified": [
53
+ 40,
54
+ "\u2581<extra_id_59>"
55
+ ],
56
+ "chinese_traditional": [
57
+ 44,
58
+ "\u2581<extra_id_55>"
59
+ ],
60
+ "english": [
61
+ 30,
62
+ "\u2581<extra_id_69>"
63
+ ],
64
+ "french": [
65
+ 10,
66
+ "\u2581<extra_id_89>"
67
+ ],
68
+ "gujarati": [
69
+ 27,
70
+ "\u2581<extra_id_72>"
71
+ ],
72
+ "hausa": [
73
+ 43,
74
+ "\u2581<extra_id_56>"
75
+ ],
76
+ "hindi": [
77
+ 21,
78
+ "\u2581<extra_id_78>"
79
+ ],
80
+ "igbo": [
81
+ 9,
82
+ "\u2581<extra_id_90>"
83
+ ],
84
+ "indonesian": [
85
+ 1,
86
+ "\u2581<extra_id_98>"
87
+ ],
88
+ "japanese": [
89
+ 37,
90
+ "\u2581<extra_id_62>"
91
+ ],
92
+ "kirundi": [
93
+ 0,
94
+ "\u2581<extra_id_99>"
95
+ ],
96
+ "korean": [
97
+ 29,
98
+ "\u2581<extra_id_70>"
99
+ ],
100
+ "kyrgyz": [
101
+ 5,
102
+ "\u2581<extra_id_94>"
103
+ ],
104
+ "marathi": [
105
+ 13,
106
+ "\u2581<extra_id_86>"
107
+ ],
108
+ "nepali": [
109
+ 20,
110
+ "\u2581<extra_id_79>"
111
+ ],
112
+ "oromo": [
113
+ 41,
114
+ "\u2581<extra_id_58>"
115
+ ],
116
+ "pashto": [
117
+ 34,
118
+ "\u2581<extra_id_65>"
119
+ ],
120
+ "persian": [
121
+ 23,
122
+ "\u2581<extra_id_76>"
123
+ ],
124
+ "pidgin": [
125
+ 14,
126
+ "\u2581<extra_id_85>"
127
+ ],
128
+ "portuguese": [
129
+ 39,
130
+ "\u2581<extra_id_60>"
131
+ ],
132
+ "punjabi": [
133
+ 17,
134
+ "\u2581<extra_id_82>"
135
+ ],
136
+ "russian": [
137
+ 36,
138
+ "\u2581<extra_id_63>"
139
+ ],
140
+ "scottish_gaelic": [
141
+ 24,
142
+ "\u2581<extra_id_75>"
143
+ ],
144
+ "serbian_cyrillic": [
145
+ 28,
146
+ "\u2581<extra_id_71>"
147
+ ],
148
+ "serbian_latin": [
149
+ 11,
150
+ "\u2581<extra_id_88>"
151
+ ],
152
+ "sinhala": [
153
+ 31,
154
+ "\u2581<extra_id_68>"
155
+ ],
156
+ "somali": [
157
+ 19,
158
+ "\u2581<extra_id_80>"
159
+ ],
160
+ "spanish": [
161
+ 3,
162
+ "\u2581<extra_id_96>"
163
+ ],
164
+ "swahili": [
165
+ 18,
166
+ "\u2581<extra_id_81>"
167
+ ],
168
+ "tamil": [
169
+ 32,
170
+ "\u2581<extra_id_67>"
171
+ ],
172
+ "telugu": [
173
+ 22,
174
+ "\u2581<extra_id_77>"
175
+ ],
176
+ "thai": [
177
+ 6,
178
+ "\u2581<extra_id_93>"
179
+ ],
180
+ "tigrinya": [
181
+ 16,
182
+ "\u2581<extra_id_83>"
183
+ ],
184
+ "turkish": [
185
+ 15,
186
+ "\u2581<extra_id_84>"
187
+ ],
188
+ "ukrainian": [
189
+ 2,
190
+ "\u2581<extra_id_97>"
191
+ ],
192
+ "urdu": [
193
+ 38,
194
+ "\u2581<extra_id_61>"
195
+ ],
196
+ "uzbek": [
197
+ 8,
198
+ "\u2581<extra_id_91>"
199
+ ],
200
+ "vietnamese": [
201
+ 12,
202
+ "\u2581<extra_id_87>"
203
+ ],
204
+ "welsh": [
205
+ 26,
206
+ "\u2581<extra_id_73>"
207
+ ],
208
+ "yoruba": [
209
+ 25,
210
+ "\u2581<extra_id_74>"
211
+ ]
212
+ }
213
+ },
214
+ "tie_word_embeddings": false,
215
+ "tokenizer_class": "T5Tokenizer",
216
+ "torch_dtype": "float32",
217
+ "transformers_version": "4.37.2",
218
+ "use_cache": true,
219
+ "vocab_size": 250112
220
+ }
generated_predictions.txt ADDED
The diff for this file is too large to render. See raw diff
 
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4d34b7c51a8139e88535d3b00ea138a4e34561d3a493d899cf4901c175f38d55
3
+ size 2329638768
predict_results.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "predict_gen_len": 52.7633955448525,
3
+ "predict_loss": 3.9636030197143555,
4
+ "predict_rouge1": 0.1906,
5
+ "predict_rouge2": 0.0401,
6
+ "predict_rougeL": 0.1886,
7
+ "predict_rougeLsum": 0.1886,
8
+ "predict_runtime": 234.4913,
9
+ "predict_samples": 1661,
10
+ "predict_samples_per_second": 7.083,
11
+ "predict_steps_per_second": 0.222
12
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "▁<extra_id_64>",
4
+ "▁<extra_id_95>",
5
+ "▁<extra_id_92>",
6
+ "▁<extra_id_57>",
7
+ "▁<extra_id_66>",
8
+ "▁<extra_id_59>",
9
+ "▁<extra_id_55>",
10
+ "▁<extra_id_69>",
11
+ "▁<extra_id_89>",
12
+ "▁<extra_id_72>",
13
+ "▁<extra_id_56>",
14
+ "▁<extra_id_78>",
15
+ "▁<extra_id_90>",
16
+ "▁<extra_id_98>",
17
+ "▁<extra_id_62>",
18
+ "▁<extra_id_99>",
19
+ "▁<extra_id_70>",
20
+ "▁<extra_id_94>",
21
+ "▁<extra_id_86>",
22
+ "▁<extra_id_79>",
23
+ "▁<extra_id_58>",
24
+ "▁<extra_id_65>",
25
+ "▁<extra_id_76>",
26
+ "▁<extra_id_85>",
27
+ "▁<extra_id_60>",
28
+ "▁<extra_id_82>",
29
+ "▁<extra_id_63>",
30
+ "▁<extra_id_75>",
31
+ "▁<extra_id_71>",
32
+ "▁<extra_id_88>",
33
+ "▁<extra_id_68>",
34
+ "▁<extra_id_80>",
35
+ "▁<extra_id_96>",
36
+ "▁<extra_id_81>",
37
+ "▁<extra_id_67>",
38
+ "▁<extra_id_77>",
39
+ "▁<extra_id_93>",
40
+ "▁<extra_id_83>",
41
+ "▁<extra_id_84>",
42
+ "▁<extra_id_97>",
43
+ "▁<extra_id_61>",
44
+ "▁<extra_id_91>",
45
+ "▁<extra_id_87>",
46
+ "▁<extra_id_73>",
47
+ "▁<extra_id_74>"
48
+ ],
49
+ "eos_token": {
50
+ "content": "</s>",
51
+ "lstrip": false,
52
+ "normalized": false,
53
+ "rstrip": false,
54
+ "single_word": false
55
+ },
56
+ "pad_token": {
57
+ "content": "<pad>",
58
+ "lstrip": false,
59
+ "normalized": false,
60
+ "rstrip": false,
61
+ "single_word": false
62
+ },
63
+ "unk_token": {
64
+ "content": "<unk>",
65
+ "lstrip": false,
66
+ "normalized": false,
67
+ "rstrip": false,
68
+ "single_word": false
69
+ }
70
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:52d050cd36d20923108fd5761aac44d5e38f95a05def1228b74d65076a11c4ef
3
+ size 16339247
tokenizer_config.json ADDED
@@ -0,0 +1,444 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "added_tokens_decoder": {
3
+ "0": {
4
+ "content": "<pad>",
5
+ "lstrip": false,
6
+ "normalized": false,
7
+ "rstrip": false,
8
+ "single_word": false,
9
+ "special": true
10
+ },
11
+ "1": {
12
+ "content": "</s>",
13
+ "lstrip": false,
14
+ "normalized": false,
15
+ "rstrip": false,
16
+ "single_word": false,
17
+ "special": true
18
+ },
19
+ "2": {
20
+ "content": "<unk>",
21
+ "lstrip": false,
22
+ "normalized": false,
23
+ "rstrip": false,
24
+ "single_word": false,
25
+ "special": true
26
+ },
27
+ "250000": {
28
+ "content": "▁<extra_id_99>",
29
+ "lstrip": false,
30
+ "normalized": false,
31
+ "rstrip": false,
32
+ "single_word": false,
33
+ "special": true
34
+ },
35
+ "250001": {
36
+ "content": "▁<extra_id_98>",
37
+ "lstrip": false,
38
+ "normalized": false,
39
+ "rstrip": false,
40
+ "single_word": false,
41
+ "special": true
42
+ },
43
+ "250002": {
44
+ "content": "▁<extra_id_97>",
45
+ "lstrip": false,
46
+ "normalized": false,
47
+ "rstrip": false,
48
+ "single_word": false,
49
+ "special": true
50
+ },
51
+ "250003": {
52
+ "content": "▁<extra_id_96>",
53
+ "lstrip": false,
54
+ "normalized": false,
55
+ "rstrip": false,
56
+ "single_word": false,
57
+ "special": true
58
+ },
59
+ "250004": {
60
+ "content": "▁<extra_id_95>",
61
+ "lstrip": false,
62
+ "normalized": false,
63
+ "rstrip": false,
64
+ "single_word": false,
65
+ "special": true
66
+ },
67
+ "250005": {
68
+ "content": "▁<extra_id_94>",
69
+ "lstrip": false,
70
+ "normalized": false,
71
+ "rstrip": false,
72
+ "single_word": false,
73
+ "special": true
74
+ },
75
+ "250006": {
76
+ "content": "▁<extra_id_93>",
77
+ "lstrip": false,
78
+ "normalized": false,
79
+ "rstrip": false,
80
+ "single_word": false,
81
+ "special": true
82
+ },
83
+ "250007": {
84
+ "content": "▁<extra_id_92>",
85
+ "lstrip": false,
86
+ "normalized": false,
87
+ "rstrip": false,
88
+ "single_word": false,
89
+ "special": true
90
+ },
91
+ "250008": {
92
+ "content": "▁<extra_id_91>",
93
+ "lstrip": false,
94
+ "normalized": false,
95
+ "rstrip": false,
96
+ "single_word": false,
97
+ "special": true
98
+ },
99
+ "250009": {
100
+ "content": "▁<extra_id_90>",
101
+ "lstrip": false,
102
+ "normalized": false,
103
+ "rstrip": false,
104
+ "single_word": false,
105
+ "special": true
106
+ },
107
+ "250010": {
108
+ "content": "▁<extra_id_89>",
109
+ "lstrip": false,
110
+ "normalized": false,
111
+ "rstrip": false,
112
+ "single_word": false,
113
+ "special": true
114
+ },
115
+ "250011": {
116
+ "content": "▁<extra_id_88>",
117
+ "lstrip": false,
118
+ "normalized": false,
119
+ "rstrip": false,
120
+ "single_word": false,
121
+ "special": true
122
+ },
123
+ "250012": {
124
+ "content": "▁<extra_id_87>",
125
+ "lstrip": false,
126
+ "normalized": false,
127
+ "rstrip": false,
128
+ "single_word": false,
129
+ "special": true
130
+ },
131
+ "250013": {
132
+ "content": "▁<extra_id_86>",
133
+ "lstrip": false,
134
+ "normalized": false,
135
+ "rstrip": false,
136
+ "single_word": false,
137
+ "special": true
138
+ },
139
+ "250014": {
140
+ "content": "▁<extra_id_85>",
141
+ "lstrip": false,
142
+ "normalized": false,
143
+ "rstrip": false,
144
+ "single_word": false,
145
+ "special": true
146
+ },
147
+ "250015": {
148
+ "content": "▁<extra_id_84>",
149
+ "lstrip": false,
150
+ "normalized": false,
151
+ "rstrip": false,
152
+ "single_word": false,
153
+ "special": true
154
+ },
155
+ "250016": {
156
+ "content": "▁<extra_id_83>",
157
+ "lstrip": false,
158
+ "normalized": false,
159
+ "rstrip": false,
160
+ "single_word": false,
161
+ "special": true
162
+ },
163
+ "250017": {
164
+ "content": "▁<extra_id_82>",
165
+ "lstrip": false,
166
+ "normalized": false,
167
+ "rstrip": false,
168
+ "single_word": false,
169
+ "special": true
170
+ },
171
+ "250018": {
172
+ "content": "▁<extra_id_81>",
173
+ "lstrip": false,
174
+ "normalized": false,
175
+ "rstrip": false,
176
+ "single_word": false,
177
+ "special": true
178
+ },
179
+ "250019": {
180
+ "content": "▁<extra_id_80>",
181
+ "lstrip": false,
182
+ "normalized": false,
183
+ "rstrip": false,
184
+ "single_word": false,
185
+ "special": true
186
+ },
187
+ "250020": {
188
+ "content": "▁<extra_id_79>",
189
+ "lstrip": false,
190
+ "normalized": false,
191
+ "rstrip": false,
192
+ "single_word": false,
193
+ "special": true
194
+ },
195
+ "250021": {
196
+ "content": "▁<extra_id_78>",
197
+ "lstrip": false,
198
+ "normalized": false,
199
+ "rstrip": false,
200
+ "single_word": false,
201
+ "special": true
202
+ },
203
+ "250022": {
204
+ "content": "▁<extra_id_77>",
205
+ "lstrip": false,
206
+ "normalized": false,
207
+ "rstrip": false,
208
+ "single_word": false,
209
+ "special": true
210
+ },
211
+ "250023": {
212
+ "content": "▁<extra_id_76>",
213
+ "lstrip": false,
214
+ "normalized": false,
215
+ "rstrip": false,
216
+ "single_word": false,
217
+ "special": true
218
+ },
219
+ "250024": {
220
+ "content": "▁<extra_id_75>",
221
+ "lstrip": false,
222
+ "normalized": false,
223
+ "rstrip": false,
224
+ "single_word": false,
225
+ "special": true
226
+ },
227
+ "250025": {
228
+ "content": "▁<extra_id_74>",
229
+ "lstrip": false,
230
+ "normalized": false,
231
+ "rstrip": false,
232
+ "single_word": false,
233
+ "special": true
234
+ },
235
+ "250026": {
236
+ "content": "▁<extra_id_73>",
237
+ "lstrip": false,
238
+ "normalized": false,
239
+ "rstrip": false,
240
+ "single_word": false,
241
+ "special": true
242
+ },
243
+ "250027": {
244
+ "content": "▁<extra_id_72>",
245
+ "lstrip": false,
246
+ "normalized": false,
247
+ "rstrip": false,
248
+ "single_word": false,
249
+ "special": true
250
+ },
251
+ "250028": {
252
+ "content": "▁<extra_id_71>",
253
+ "lstrip": false,
254
+ "normalized": false,
255
+ "rstrip": false,
256
+ "single_word": false,
257
+ "special": true
258
+ },
259
+ "250029": {
260
+ "content": "▁<extra_id_70>",
261
+ "lstrip": false,
262
+ "normalized": false,
263
+ "rstrip": false,
264
+ "single_word": false,
265
+ "special": true
266
+ },
267
+ "250030": {
268
+ "content": "▁<extra_id_69>",
269
+ "lstrip": false,
270
+ "normalized": false,
271
+ "rstrip": false,
272
+ "single_word": false,
273
+ "special": true
274
+ },
275
+ "250031": {
276
+ "content": "▁<extra_id_68>",
277
+ "lstrip": false,
278
+ "normalized": false,
279
+ "rstrip": false,
280
+ "single_word": false,
281
+ "special": true
282
+ },
283
+ "250032": {
284
+ "content": "▁<extra_id_67>",
285
+ "lstrip": false,
286
+ "normalized": false,
287
+ "rstrip": false,
288
+ "single_word": false,
289
+ "special": true
290
+ },
291
+ "250033": {
292
+ "content": "▁<extra_id_66>",
293
+ "lstrip": false,
294
+ "normalized": false,
295
+ "rstrip": false,
296
+ "single_word": false,
297
+ "special": true
298
+ },
299
+ "250034": {
300
+ "content": "▁<extra_id_65>",
301
+ "lstrip": false,
302
+ "normalized": false,
303
+ "rstrip": false,
304
+ "single_word": false,
305
+ "special": true
306
+ },
307
+ "250035": {
308
+ "content": "▁<extra_id_64>",
309
+ "lstrip": false,
310
+ "normalized": false,
311
+ "rstrip": false,
312
+ "single_word": false,
313
+ "special": true
314
+ },
315
+ "250036": {
316
+ "content": "▁<extra_id_63>",
317
+ "lstrip": false,
318
+ "normalized": false,
319
+ "rstrip": false,
320
+ "single_word": false,
321
+ "special": true
322
+ },
323
+ "250037": {
324
+ "content": "▁<extra_id_62>",
325
+ "lstrip": false,
326
+ "normalized": false,
327
+ "rstrip": false,
328
+ "single_word": false,
329
+ "special": true
330
+ },
331
+ "250038": {
332
+ "content": "▁<extra_id_61>",
333
+ "lstrip": false,
334
+ "normalized": false,
335
+ "rstrip": false,
336
+ "single_word": false,
337
+ "special": true
338
+ },
339
+ "250039": {
340
+ "content": "▁<extra_id_60>",
341
+ "lstrip": false,
342
+ "normalized": false,
343
+ "rstrip": false,
344
+ "single_word": false,
345
+ "special": true
346
+ },
347
+ "250040": {
348
+ "content": "▁<extra_id_59>",
349
+ "lstrip": false,
350
+ "normalized": false,
351
+ "rstrip": false,
352
+ "single_word": false,
353
+ "special": true
354
+ },
355
+ "250041": {
356
+ "content": "▁<extra_id_58>",
357
+ "lstrip": false,
358
+ "normalized": false,
359
+ "rstrip": false,
360
+ "single_word": false,
361
+ "special": true
362
+ },
363
+ "250042": {
364
+ "content": "▁<extra_id_57>",
365
+ "lstrip": false,
366
+ "normalized": false,
367
+ "rstrip": false,
368
+ "single_word": false,
369
+ "special": true
370
+ },
371
+ "250043": {
372
+ "content": "▁<extra_id_56>",
373
+ "lstrip": false,
374
+ "normalized": false,
375
+ "rstrip": false,
376
+ "single_word": false,
377
+ "special": true
378
+ },
379
+ "250044": {
380
+ "content": "▁<extra_id_55>",
381
+ "lstrip": false,
382
+ "normalized": false,
383
+ "rstrip": false,
384
+ "single_word": false,
385
+ "special": true
386
+ }
387
+ },
388
+ "additional_special_tokens": [
389
+ "▁<extra_id_64>",
390
+ "▁<extra_id_95>",
391
+ "▁<extra_id_92>",
392
+ "▁<extra_id_57>",
393
+ "▁<extra_id_66>",
394
+ "▁<extra_id_59>",
395
+ "▁<extra_id_55>",
396
+ "▁<extra_id_69>",
397
+ "▁<extra_id_89>",
398
+ "▁<extra_id_72>",
399
+ "▁<extra_id_56>",
400
+ "▁<extra_id_78>",
401
+ "▁<extra_id_90>",
402
+ "▁<extra_id_98>",
403
+ "▁<extra_id_62>",
404
+ "▁<extra_id_99>",
405
+ "▁<extra_id_70>",
406
+ "▁<extra_id_94>",
407
+ "▁<extra_id_86>",
408
+ "▁<extra_id_79>",
409
+ "▁<extra_id_58>",
410
+ "▁<extra_id_65>",
411
+ "▁<extra_id_76>",
412
+ "▁<extra_id_85>",
413
+ "▁<extra_id_60>",
414
+ "▁<extra_id_82>",
415
+ "▁<extra_id_63>",
416
+ "▁<extra_id_75>",
417
+ "▁<extra_id_71>",
418
+ "▁<extra_id_88>",
419
+ "▁<extra_id_68>",
420
+ "▁<extra_id_80>",
421
+ "▁<extra_id_96>",
422
+ "▁<extra_id_81>",
423
+ "▁<extra_id_67>",
424
+ "▁<extra_id_77>",
425
+ "▁<extra_id_93>",
426
+ "▁<extra_id_83>",
427
+ "▁<extra_id_84>",
428
+ "▁<extra_id_97>",
429
+ "▁<extra_id_61>",
430
+ "οΏ½οΏ½οΏ½<extra_id_91>",
431
+ "▁<extra_id_87>",
432
+ "▁<extra_id_73>",
433
+ "▁<extra_id_74>"
434
+ ],
435
+ "clean_up_tokenization_spaces": true,
436
+ "eos_token": "</s>",
437
+ "extra_ids": 0,
438
+ "legacy": true,
439
+ "model_max_length": 1000000000000000019884624838656,
440
+ "pad_token": "<pad>",
441
+ "sp_model_kwargs": {},
442
+ "tokenizer_class": "T5Tokenizer",
443
+ "unk_token": "<unk>"
444
+ }
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1f5ec3235b6bee14048d05ae78f3c01dc21fb6957cbfb6f9b1c93178de1ea482
3
+ size 5048