xnohat commited on
Commit
26eeab1
1 Parent(s): 73d49d4

Upload folder using huggingface_hub

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
config.json ADDED
@@ -0,0 +1,219 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "csebuetnlp/mT5_m2m_crossSum_enhanced",
3
+ "architectures": [
4
+ "MT5ForConditionalGeneration"
5
+ ],
6
+ "classifier_dropout": 0.0,
7
+ "d_ff": 2048,
8
+ "d_kv": 64,
9
+ "d_model": 768,
10
+ "decoder_start_token_id": 0,
11
+ "dense_act_fn": "gelu_new",
12
+ "dropout_rate": 0.1,
13
+ "eos_token_id": 1,
14
+ "feed_forward_proj": "gated-gelu",
15
+ "initializer_factor": 1.0,
16
+ "is_encoder_decoder": true,
17
+ "is_gated_act": true,
18
+ "layer_norm_epsilon": 1e-06,
19
+ "length_penalty": 0.6,
20
+ "max_length": 84,
21
+ "model_type": "mt5",
22
+ "num_beams": 4,
23
+ "num_decoder_layers": 12,
24
+ "num_heads": 12,
25
+ "num_layers": 12,
26
+ "output_past": true,
27
+ "pad_token_id": 0,
28
+ "relative_attention_max_distance": 128,
29
+ "relative_attention_num_buckets": 32,
30
+ "task_specific_params": {
31
+ "langid_map": {
32
+ "amharic": [
33
+ 35,
34
+ "\u2581<extra_id_64>"
35
+ ],
36
+ "arabic": [
37
+ 4,
38
+ "\u2581<extra_id_95>"
39
+ ],
40
+ "azerbaijani": [
41
+ 7,
42
+ "\u2581<extra_id_92>"
43
+ ],
44
+ "bengali": [
45
+ 42,
46
+ "\u2581<extra_id_57>"
47
+ ],
48
+ "burmese": [
49
+ 33,
50
+ "\u2581<extra_id_66>"
51
+ ],
52
+ "chinese_simplified": [
53
+ 40,
54
+ "\u2581<extra_id_59>"
55
+ ],
56
+ "chinese_traditional": [
57
+ 44,
58
+ "\u2581<extra_id_55>"
59
+ ],
60
+ "english": [
61
+ 30,
62
+ "\u2581<extra_id_69>"
63
+ ],
64
+ "french": [
65
+ 10,
66
+ "\u2581<extra_id_89>"
67
+ ],
68
+ "gujarati": [
69
+ 27,
70
+ "\u2581<extra_id_72>"
71
+ ],
72
+ "hausa": [
73
+ 43,
74
+ "\u2581<extra_id_56>"
75
+ ],
76
+ "hindi": [
77
+ 21,
78
+ "\u2581<extra_id_78>"
79
+ ],
80
+ "igbo": [
81
+ 9,
82
+ "\u2581<extra_id_90>"
83
+ ],
84
+ "indonesian": [
85
+ 1,
86
+ "\u2581<extra_id_98>"
87
+ ],
88
+ "japanese": [
89
+ 37,
90
+ "\u2581<extra_id_62>"
91
+ ],
92
+ "kirundi": [
93
+ 0,
94
+ "\u2581<extra_id_99>"
95
+ ],
96
+ "korean": [
97
+ 29,
98
+ "\u2581<extra_id_70>"
99
+ ],
100
+ "kyrgyz": [
101
+ 5,
102
+ "\u2581<extra_id_94>"
103
+ ],
104
+ "marathi": [
105
+ 13,
106
+ "\u2581<extra_id_86>"
107
+ ],
108
+ "nepali": [
109
+ 20,
110
+ "\u2581<extra_id_79>"
111
+ ],
112
+ "oromo": [
113
+ 41,
114
+ "\u2581<extra_id_58>"
115
+ ],
116
+ "pashto": [
117
+ 34,
118
+ "\u2581<extra_id_65>"
119
+ ],
120
+ "persian": [
121
+ 23,
122
+ "\u2581<extra_id_76>"
123
+ ],
124
+ "pidgin": [
125
+ 14,
126
+ "\u2581<extra_id_85>"
127
+ ],
128
+ "portuguese": [
129
+ 39,
130
+ "\u2581<extra_id_60>"
131
+ ],
132
+ "punjabi": [
133
+ 17,
134
+ "\u2581<extra_id_82>"
135
+ ],
136
+ "russian": [
137
+ 36,
138
+ "\u2581<extra_id_63>"
139
+ ],
140
+ "scottish_gaelic": [
141
+ 24,
142
+ "\u2581<extra_id_75>"
143
+ ],
144
+ "serbian_cyrillic": [
145
+ 28,
146
+ "\u2581<extra_id_71>"
147
+ ],
148
+ "serbian_latin": [
149
+ 11,
150
+ "\u2581<extra_id_88>"
151
+ ],
152
+ "sinhala": [
153
+ 31,
154
+ "\u2581<extra_id_68>"
155
+ ],
156
+ "somali": [
157
+ 19,
158
+ "\u2581<extra_id_80>"
159
+ ],
160
+ "spanish": [
161
+ 3,
162
+ "\u2581<extra_id_96>"
163
+ ],
164
+ "swahili": [
165
+ 18,
166
+ "\u2581<extra_id_81>"
167
+ ],
168
+ "tamil": [
169
+ 32,
170
+ "\u2581<extra_id_67>"
171
+ ],
172
+ "telugu": [
173
+ 22,
174
+ "\u2581<extra_id_77>"
175
+ ],
176
+ "thai": [
177
+ 6,
178
+ "\u2581<extra_id_93>"
179
+ ],
180
+ "tigrinya": [
181
+ 16,
182
+ "\u2581<extra_id_83>"
183
+ ],
184
+ "turkish": [
185
+ 15,
186
+ "\u2581<extra_id_84>"
187
+ ],
188
+ "ukrainian": [
189
+ 2,
190
+ "\u2581<extra_id_97>"
191
+ ],
192
+ "urdu": [
193
+ 38,
194
+ "\u2581<extra_id_61>"
195
+ ],
196
+ "uzbek": [
197
+ 8,
198
+ "\u2581<extra_id_91>"
199
+ ],
200
+ "vietnamese": [
201
+ 12,
202
+ "\u2581<extra_id_87>"
203
+ ],
204
+ "welsh": [
205
+ 26,
206
+ "\u2581<extra_id_73>"
207
+ ],
208
+ "yoruba": [
209
+ 25,
210
+ "\u2581<extra_id_74>"
211
+ ]
212
+ }
213
+ },
214
+ "tie_word_embeddings": false,
215
+ "tokenizer_class": "T5Tokenizer",
216
+ "transformers_version": "4.33.2",
217
+ "use_cache": true,
218
+ "vocab_size": 250112
219
+ }
generation_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "decoder_start_token_id": 0,
4
+ "eos_token_id": 1,
5
+ "length_penalty": 0.6,
6
+ "max_length": 84,
7
+ "num_beams": 4,
8
+ "pad_token_id": 0,
9
+ "transformers_version": "4.33.2"
10
+ }
onnx/decoder_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3b98853530f6475374429f0bf29d6c49368e962703970f2fd3a51e063f2bfaa1
3
+ size 1990142662
onnx/decoder_model_merged.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5cc708c6b5878f7a274dbb4feb4a484c796a37bf7deae3defe18e80eeee215dc
3
+ size 1990476587
onnx/decoder_model_merged_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d320cbcf3840a7629fcf45e0750d0649b451a01345d2131c9cec19608f03eaed
3
+ size 500652977
onnx/decoder_model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7ae4401116d2fb14067839fa5e14627945b2df9c3591fbe000ae0aff32c96f39
3
+ size 500076464
onnx/decoder_with_past_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e4908352dcd271c00fae22d03df454b4487fe9791f2391163df07389a2a12d8
3
+ size 1933491010
onnx/decoder_with_past_model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:96be0b979b5d41f1079b44af0799c0bb01470d7f5185671555962b83076ba6f1
3
+ size 485767498
onnx/encoder_model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ce603211c93a2537a605963aa49c926a955315186342b8de00904f5390af792c
3
+ size 1108344239
onnx/encoder_model_quantized.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:989de67618174fbc1c45aaf55d8316f1903a23575e0fdb30dcdb515d81b5ecb8
3
+ size 277914072
quantize_config.json ADDED
@@ -0,0 +1,139 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "per_channel": true,
3
+ "reduce_range": true,
4
+ "per_model_config": {
5
+ "encoder_model": {
6
+ "op_types": [
7
+ "Abs",
8
+ "Range",
9
+ "ReduceMean",
10
+ "ConstantOfShape",
11
+ "Concat",
12
+ "Sub",
13
+ "Greater",
14
+ "Shape",
15
+ "Tanh",
16
+ "Cast",
17
+ "Reshape",
18
+ "Div",
19
+ "Pow",
20
+ "Where",
21
+ "Min",
22
+ "Gather",
23
+ "Constant",
24
+ "Transpose",
25
+ "Less",
26
+ "Log",
27
+ "Mul",
28
+ "Sqrt",
29
+ "MatMul",
30
+ "Unsqueeze",
31
+ "Softmax",
32
+ "Add"
33
+ ],
34
+ "weight_type": "QInt8"
35
+ },
36
+ "decoder_with_past_model": {
37
+ "op_types": [
38
+ "Range",
39
+ "ReduceMean",
40
+ "LessOrEqual",
41
+ "ConstantOfShape",
42
+ "Concat",
43
+ "Sub",
44
+ "Slice",
45
+ "Shape",
46
+ "Expand",
47
+ "Tanh",
48
+ "Cast",
49
+ "Reshape",
50
+ "Div",
51
+ "Pow",
52
+ "Where",
53
+ "Min",
54
+ "Gather",
55
+ "Constant",
56
+ "Tile",
57
+ "Transpose",
58
+ "Less",
59
+ "Log",
60
+ "Mul",
61
+ "Sqrt",
62
+ "MatMul",
63
+ "Unsqueeze",
64
+ "Neg",
65
+ "Softmax",
66
+ "Add"
67
+ ],
68
+ "weight_type": "QInt8"
69
+ },
70
+ "decoder_model": {
71
+ "op_types": [
72
+ "Range",
73
+ "ReduceMean",
74
+ "LessOrEqual",
75
+ "ConstantOfShape",
76
+ "Concat",
77
+ "Sub",
78
+ "Shape",
79
+ "Expand",
80
+ "Tanh",
81
+ "Cast",
82
+ "Reshape",
83
+ "Div",
84
+ "Pow",
85
+ "Where",
86
+ "Min",
87
+ "Gather",
88
+ "Constant",
89
+ "Tile",
90
+ "Transpose",
91
+ "Less",
92
+ "Log",
93
+ "Mul",
94
+ "Sqrt",
95
+ "MatMul",
96
+ "Unsqueeze",
97
+ "Neg",
98
+ "Softmax",
99
+ "Add"
100
+ ],
101
+ "weight_type": "QInt8"
102
+ },
103
+ "decoder_model_merged": {
104
+ "op_types": [
105
+ "Range",
106
+ "ReduceMean",
107
+ "LessOrEqual",
108
+ "ConstantOfShape",
109
+ "Concat",
110
+ "Sub",
111
+ "Slice",
112
+ "Shape",
113
+ "Expand",
114
+ "Tanh",
115
+ "If",
116
+ "Cast",
117
+ "Reshape",
118
+ "Pow",
119
+ "Div",
120
+ "Where",
121
+ "Min",
122
+ "Gather",
123
+ "Constant",
124
+ "Tile",
125
+ "Transpose",
126
+ "Less",
127
+ "Log",
128
+ "Mul",
129
+ "Sqrt",
130
+ "MatMul",
131
+ "Unsqueeze",
132
+ "Neg",
133
+ "Softmax",
134
+ "Add"
135
+ ],
136
+ "weight_type": "QInt8"
137
+ }
138
+ }
139
+ }
special_tokens_map.json ADDED
@@ -0,0 +1,5 @@
 
 
 
 
 
 
1
+ {
2
+ "eos_token": "</s>",
3
+ "pad_token": "<pad>",
4
+ "unk_token": "<unk>"
5
+ }
spiece.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef78f86560d809067d12bac6c09f19a462cb3af3f54d2b8acbba26e1433125d6
3
+ size 4309802
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:93c3578052e1605d8332eb961bc08d72e246071974e4cc54aa6991826b802aa5
3
+ size 16330369
tokenizer_config.json ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": null,
3
+ "clean_up_tokenization_spaces": true,
4
+ "eos_token": "</s>",
5
+ "extra_ids": 0,
6
+ "legacy": true,
7
+ "model_max_length": 1000000000000000019884624838656,
8
+ "pad_token": "<pad>",
9
+ "sp_model_kwargs": {},
10
+ "tokenizer_class": "T5Tokenizer",
11
+ "unk_token": "<unk>"
12
+ }