Upload folder using huggingface_hub

#1
by hajekad - opened
config.json ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_name_or_path": "../checkpoints/pretrain_clean/sandy-star-569_exp7_custom_rassp_neims/checkpoint-112000",
3
+ "activation_dropout": 0.0,
4
+ "activation_function": "gelu",
5
+ "architectures": [
6
+ "BartSpektroForConditionalGeneration"
7
+ ],
8
+ "attention_dropout": 0.0,
9
+ "bos_token_id": 3,
10
+ "classifier_dropout": 0.0,
11
+ "d_model": 1024,
12
+ "decoder_attention_heads": 16,
13
+ "decoder_ffn_dim": 4096,
14
+ "decoder_layerdrop": 0.0,
15
+ "decoder_layers": 12,
16
+ "decoder_max_position_embeddings": 200,
17
+ "decoder_start_token_id": 3,
18
+ "dropout": 0.2,
19
+ "encoder_attention_heads": 16,
20
+ "encoder_ffn_dim": 4096,
21
+ "encoder_layerdrop": 0.0,
22
+ "encoder_layers": 12,
23
+ "encoder_max_position_embeddings": null,
24
+ "eos_token_id": 0,
25
+ "forced_eos_token_id": 0,
26
+ "id2label": {
27
+ "0": "LABEL_0",
28
+ "1": "LABEL_1",
29
+ "2": "LABEL_2"
30
+ },
31
+ "init_std": 0.02,
32
+ "is_encoder_decoder": true,
33
+ "label2id": {
34
+ "LABEL_0": 0,
35
+ "LABEL_1": 1,
36
+ "LABEL_2": 2
37
+ },
38
+ "max_length": 200,
39
+ "max_log_id": 29,
40
+ "max_mz": 500,
41
+ "max_position_embeddings": 1024,
42
+ "min_len": 0,
43
+ "model_type": "bart",
44
+ "num_hidden_layers": 12,
45
+ "pad_token_id": 2,
46
+ "scale_embedding": false,
47
+ "separate_encoder_decoder_embeds": true,
48
+ "tie_word_embeddings": false,
49
+ "torch_dtype": "float32",
50
+ "transformers_version": "4.31.0",
51
+ "use_cache": true,
52
+ "vocab_size": 267
53
+ }
generation_config.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token_id": 3,
3
+ "decoder_start_token_id": 3,
4
+ "eos_token_id": 0,
5
+ "forced_eos_token_id": 0,
6
+ "max_length": 200,
7
+ "pad_token_id": 2,
8
+ "transformers_version": "4.31.0"
9
+ }
optimizer.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d708a80c6e9bc833d37374e4c476ca274d43c4ef47619458c6e23b45952327ca
3
+ size 2832576032
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4886159dc0fc54ffc8930ac55346b5ec6470b66b3d5c0c39d6595e72bfeb68f
3
+ size 1416249109
rng_state.pth ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:463baa2858a6f606add9d142f35585783c3c4ce2ec017d33f906d2bce920e853
3
+ size 14575
scheduler.pt ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36f9de24f91d11c05bad4a3cfc309e086d5313b0b474e31b8aa5756cdf8c1369
3
+ size 627
special_tokens_map.json ADDED
@@ -0,0 +1,6 @@
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<bos>",
3
+ "eos_token": "<eos>",
4
+ "pad_token": "<pad>",
5
+ "unk_token": "<unk>"
6
+ }
tokenizer.json ADDED
@@ -0,0 +1,406 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "1.0",
3
+ "truncation": null,
4
+ "padding": null,
5
+ "added_tokens": [
6
+ {
7
+ "id": 0,
8
+ "content": "<eos>",
9
+ "single_word": false,
10
+ "lstrip": false,
11
+ "rstrip": false,
12
+ "normalized": false,
13
+ "special": true
14
+ },
15
+ {
16
+ "id": 1,
17
+ "content": "<unk>",
18
+ "single_word": false,
19
+ "lstrip": false,
20
+ "rstrip": false,
21
+ "normalized": false,
22
+ "special": true
23
+ },
24
+ {
25
+ "id": 2,
26
+ "content": "<pad>",
27
+ "single_word": false,
28
+ "lstrip": false,
29
+ "rstrip": false,
30
+ "normalized": false,
31
+ "special": true
32
+ },
33
+ {
34
+ "id": 3,
35
+ "content": "<bos>",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
+ },
42
+ {
43
+ "id": 4,
44
+ "content": "<neims>",
45
+ "single_word": false,
46
+ "lstrip": false,
47
+ "rstrip": false,
48
+ "normalized": false,
49
+ "special": true
50
+ },
51
+ {
52
+ "id": 5,
53
+ "content": "<nist>",
54
+ "single_word": false,
55
+ "lstrip": false,
56
+ "rstrip": false,
57
+ "normalized": false,
58
+ "special": true
59
+ },
60
+ {
61
+ "id": 6,
62
+ "content": "<rassp>",
63
+ "single_word": false,
64
+ "lstrip": false,
65
+ "rstrip": false,
66
+ "normalized": false,
67
+ "special": true
68
+ },
69
+ {
70
+ "id": 7,
71
+ "content": "<trafo>",
72
+ "single_word": false,
73
+ "lstrip": false,
74
+ "rstrip": false,
75
+ "normalized": false,
76
+ "special": true
77
+ },
78
+ {
79
+ "id": 8,
80
+ "content": "<source1>",
81
+ "single_word": false,
82
+ "lstrip": false,
83
+ "rstrip": false,
84
+ "normalized": false,
85
+ "special": true
86
+ },
87
+ {
88
+ "id": 9,
89
+ "content": "<source2>",
90
+ "single_word": false,
91
+ "lstrip": false,
92
+ "rstrip": false,
93
+ "normalized": false,
94
+ "special": true
95
+ },
96
+ {
97
+ "id": 10,
98
+ "content": "<source3>",
99
+ "single_word": false,
100
+ "lstrip": false,
101
+ "rstrip": false,
102
+ "normalized": false,
103
+ "special": true
104
+ }
105
+ ],
106
+ "normalizer": {
107
+ "type": "Sequence",
108
+ "normalizers": [
109
+ {
110
+ "type": "NFKC"
111
+ }
112
+ ]
113
+ },
114
+ "pre_tokenizer": {
115
+ "type": "ByteLevel",
116
+ "add_prefix_space": true,
117
+ "trim_offsets": true,
118
+ "use_regex": true
119
+ },
120
+ "post_processor": null,
121
+ "decoder": {
122
+ "type": "ByteLevel",
123
+ "add_prefix_space": true,
124
+ "trim_offsets": true,
125
+ "use_regex": true
126
+ },
127
+ "model": {
128
+ "type": "BPE",
129
+ "dropout": null,
130
+ "unk_token": null,
131
+ "continuing_subword_prefix": null,
132
+ "end_of_word_suffix": null,
133
+ "fuse_unk": false,
134
+ "byte_fallback": false,
135
+ "vocab": {
136
+ "<eos>": 0,
137
+ "<unk>": 1,
138
+ "<pad>": 2,
139
+ "<bos>": 3,
140
+ "<neims>": 4,
141
+ "<nist>": 5,
142
+ "<rassp>": 6,
143
+ "<trafo>": 7,
144
+ "<source1>": 8,
145
+ "<source2>": 9,
146
+ "<source3>": 10,
147
+ "!": 11,
148
+ "\"": 12,
149
+ "#": 13,
150
+ "$": 14,
151
+ "%": 15,
152
+ "&": 16,
153
+ "'": 17,
154
+ "(": 18,
155
+ ")": 19,
156
+ "*": 20,
157
+ "+": 21,
158
+ ",": 22,
159
+ "-": 23,
160
+ ".": 24,
161
+ "/": 25,
162
+ "0": 26,
163
+ "1": 27,
164
+ "2": 28,
165
+ "3": 29,
166
+ "4": 30,
167
+ "5": 31,
168
+ "6": 32,
169
+ "7": 33,
170
+ "8": 34,
171
+ "9": 35,
172
+ ":": 36,
173
+ ";": 37,
174
+ "<": 38,
175
+ "=": 39,
176
+ ">": 40,
177
+ "?": 41,
178
+ "@": 42,
179
+ "A": 43,
180
+ "B": 44,
181
+ "C": 45,
182
+ "D": 46,
183
+ "E": 47,
184
+ "F": 48,
185
+ "G": 49,
186
+ "H": 50,
187
+ "I": 51,
188
+ "J": 52,
189
+ "K": 53,
190
+ "L": 54,
191
+ "M": 55,
192
+ "N": 56,
193
+ "O": 57,
194
+ "P": 58,
195
+ "Q": 59,
196
+ "R": 60,
197
+ "S": 61,
198
+ "T": 62,
199
+ "U": 63,
200
+ "V": 64,
201
+ "W": 65,
202
+ "X": 66,
203
+ "Y": 67,
204
+ "Z": 68,
205
+ "[": 69,
206
+ "\\": 70,
207
+ "]": 71,
208
+ "^": 72,
209
+ "_": 73,
210
+ "`": 74,
211
+ "a": 75,
212
+ "b": 76,
213
+ "c": 77,
214
+ "d": 78,
215
+ "e": 79,
216
+ "f": 80,
217
+ "g": 81,
218
+ "h": 82,
219
+ "i": 83,
220
+ "j": 84,
221
+ "k": 85,
222
+ "l": 86,
223
+ "m": 87,
224
+ "n": 88,
225
+ "o": 89,
226
+ "p": 90,
227
+ "q": 91,
228
+ "r": 92,
229
+ "s": 93,
230
+ "t": 94,
231
+ "u": 95,
232
+ "v": 96,
233
+ "w": 97,
234
+ "x": 98,
235
+ "y": 99,
236
+ "z": 100,
237
+ "{": 101,
238
+ "|": 102,
239
+ "}": 103,
240
+ "~": 104,
241
+ "¡": 105,
242
+ "¢": 106,
243
+ "£": 107,
244
+ "¤": 108,
245
+ "¥": 109,
246
+ "¦": 110,
247
+ "§": 111,
248
+ "¨": 112,
249
+ "©": 113,
250
+ "ª": 114,
251
+ "«": 115,
252
+ "¬": 116,
253
+ "®": 117,
254
+ "¯": 118,
255
+ "°": 119,
256
+ "±": 120,
257
+ "²": 121,
258
+ "³": 122,
259
+ "´": 123,
260
+ "µ": 124,
261
+ "¶": 125,
262
+ "·": 126,
263
+ "¸": 127,
264
+ "¹": 128,
265
+ "º": 129,
266
+ "»": 130,
267
+ "¼": 131,
268
+ "½": 132,
269
+ "¾": 133,
270
+ "¿": 134,
271
+ "À": 135,
272
+ "Á": 136,
273
+ "Â": 137,
274
+ "Ã": 138,
275
+ "Ä": 139,
276
+ "Å": 140,
277
+ "Æ": 141,
278
+ "Ç": 142,
279
+ "È": 143,
280
+ "É": 144,
281
+ "Ê": 145,
282
+ "Ë": 146,
283
+ "Ì": 147,
284
+ "Í": 148,
285
+ "Î": 149,
286
+ "Ï": 150,
287
+ "Ð": 151,
288
+ "Ñ": 152,
289
+ "Ò": 153,
290
+ "Ó": 154,
291
+ "Ô": 155,
292
+ "Õ": 156,
293
+ "Ö": 157,
294
+ "×": 158,
295
+ "Ø": 159,
296
+ "Ù": 160,
297
+ "Ú": 161,
298
+ "Û": 162,
299
+ "Ü": 163,
300
+ "Ý": 164,
301
+ "Þ": 165,
302
+ "ß": 166,
303
+ "à": 167,
304
+ "á": 168,
305
+ "â": 169,
306
+ "ã": 170,
307
+ "ä": 171,
308
+ "å": 172,
309
+ "æ": 173,
310
+ "ç": 174,
311
+ "è": 175,
312
+ "é": 176,
313
+ "ê": 177,
314
+ "ë": 178,
315
+ "ì": 179,
316
+ "í": 180,
317
+ "î": 181,
318
+ "ï": 182,
319
+ "ð": 183,
320
+ "ñ": 184,
321
+ "ò": 185,
322
+ "ó": 186,
323
+ "ô": 187,
324
+ "õ": 188,
325
+ "ö": 189,
326
+ "÷": 190,
327
+ "ø": 191,
328
+ "ù": 192,
329
+ "ú": 193,
330
+ "û": 194,
331
+ "ü": 195,
332
+ "ý": 196,
333
+ "þ": 197,
334
+ "ÿ": 198,
335
+ "Ā": 199,
336
+ "ā": 200,
337
+ "Ă": 201,
338
+ "ă": 202,
339
+ "Ą": 203,
340
+ "ą": 204,
341
+ "Ć": 205,
342
+ "ć": 206,
343
+ "Ĉ": 207,
344
+ "ĉ": 208,
345
+ "Ċ": 209,
346
+ "ċ": 210,
347
+ "Č": 211,
348
+ "č": 212,
349
+ "Ď": 213,
350
+ "ď": 214,
351
+ "Đ": 215,
352
+ "đ": 216,
353
+ "Ē": 217,
354
+ "ē": 218,
355
+ "Ĕ": 219,
356
+ "ĕ": 220,
357
+ "Ė": 221,
358
+ "ė": 222,
359
+ "Ę": 223,
360
+ "ę": 224,
361
+ "Ě": 225,
362
+ "ě": 226,
363
+ "Ĝ": 227,
364
+ "ĝ": 228,
365
+ "Ğ": 229,
366
+ "ğ": 230,
367
+ "Ġ": 231,
368
+ "ġ": 232,
369
+ "Ģ": 233,
370
+ "ģ": 234,
371
+ "Ĥ": 235,
372
+ "ĥ": 236,
373
+ "Ħ": 237,
374
+ "ħ": 238,
375
+ "Ĩ": 239,
376
+ "ĩ": 240,
377
+ "Ī": 241,
378
+ "ī": 242,
379
+ "Ĭ": 243,
380
+ "ĭ": 244,
381
+ "Į": 245,
382
+ "į": 246,
383
+ "İ": 247,
384
+ "ı": 248,
385
+ "IJ": 249,
386
+ "ij": 250,
387
+ "Ĵ": 251,
388
+ "ĵ": 252,
389
+ "Ķ": 253,
390
+ "ķ": 254,
391
+ "ĸ": 255,
392
+ "Ĺ": 256,
393
+ "ĺ": 257,
394
+ "Ļ": 258,
395
+ "ļ": 259,
396
+ "Ľ": 260,
397
+ "ľ": 261,
398
+ "Ŀ": 262,
399
+ "ŀ": 263,
400
+ "Ł": 264,
401
+ "ł": 265,
402
+ "Ń": 266
403
+ },
404
+ "merges": []
405
+ }
406
+ }
tokenizer_config.json ADDED
@@ -0,0 +1,10 @@
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bos_token": "<bos>",
3
+ "clean_up_tokenization_spaces": true,
4
+ "eos_token": "<eos>",
5
+ "is_split_into_words": true,
6
+ "model_max_length": 1000000000000000019884624838656,
7
+ "pad_token": "<pad>",
8
+ "tokenizer_class": "PreTrainedTokenizerFast",
9
+ "unk_token": "<unk>"
10
+ }
trainer_state.json ADDED
The diff for this file is too large to render. See raw diff
 
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f44cabb93c86813d270d999171981798d8e3c0e70a676d193590100f9c45fde
3
+ size 4283