CompactAI commited on
Commit
31397f0
·
verified ·
1 Parent(s): d3ab0fb

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - pruned
5
+ - python
6
+ - optimized
7
+ - wanda
8
+ base_model: tiiuae/Falcon-H1-Tiny-R-90M
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+ # Falcon-H1-Tiny-R-90M-python-safe
13
+
14
+ > 🎯 **PYTHON-optimized** | 📦 **Safe** pruning | ⚡ **1% weights pruned**
15
+
16
+ This model is a **conservatively pruned** version of [tiiuae/Falcon-H1-Tiny-R-90M](https://huggingface.co/tiiuae/Falcon-H1-Tiny-R-90M).
17
+
18
+ ## Performance Comparison
19
+
20
+ | Category | Original | Pruned | Change |
21
+ |----------|----------|--------|--------|
22
+ | **Python** | 0.0% | 0.0% ⭐ | → |
23
+ | Html | 0.0% | 0.0% | → |
24
+ | Trivia | 0.0% | 0.0% | → |
25
+ | Math | 0.0% | 0.0% | → |
26
+ | Reasoning | 0.0% | 0.0% | → |
27
+ | Medical | 0.0% | 16.7% | ↑ 16.7% |
28
+ | Linux | 0.0% | 0.0% | → |
29
+ | Writing | 0.0% | 0.0% | → |
30
+
31
+ **Average**: 0.0% → 2.1% (+2.1%)
32
+
33
+
34
+
35
+ ![Comparison Graph](comparison_graph.png)
36
+
37
+ ## Quick Start
38
+
39
+ ```python
40
+ from transformers import AutoModelForCausalLM, AutoTokenizer
41
+
42
+ model = AutoModelForCausalLM.from_pretrained("CompactAI/Falcon-H1-Tiny-R-90M-python-safe")
43
+ tokenizer = AutoTokenizer.from_pretrained("CompactAI/Falcon-H1-Tiny-R-90M-python-safe")
44
+
45
+ inputs = tokenizer("Your prompt here", return_tensors="pt")
46
+ outputs = model.generate(**inputs, max_new_tokens=100)
47
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
48
+ ```
49
+
50
+ ## Technical Details
51
+
52
+ | Property | Value |
53
+ |----------|-------|
54
+ | Base Model | [tiiuae/Falcon-H1-Tiny-R-90M](https://huggingface.co/tiiuae/Falcon-H1-Tiny-R-90M) |
55
+ | Specialization | Python |
56
+ | Prune Mode | Safe |
57
+ | Weight Reduction | 1% weights pruned |
58
+
59
+ ## License
60
+
61
+ This model inherits the license from the base model.
chat_template.jinja ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %} {{bos_token}}<|system|>
2
+ {%- if messages[0]['role'] == 'system' %}
3
+ {{ messages[0]['content'] }}
4
+ {%- set remaining_messages = messages[1:] %}
5
+ {%- else %}
6
+ {%- set remaining_messages = messages %}
7
+ {%- endif %}
8
+ {{ 'You are a Falcon assistant skilled in function calling. You are helpful, respectful, and concise.
9
+
10
+ # Tools
11
+
12
+ You have access to the following functions. You MUST use them to answer questions when needed. For each function call, you MUST return a JSON object inside <tool_call></tool_call> tags.
13
+
14
+ <tools>' + tools|tojson(indent=2) + '</tools>
15
+
16
+ # Output Format
17
+
18
+ Your response MUST follow this format when making function calls:
19
+ <tool_call>
20
+ [
21
+ {"name": "function_name", "arguments": {"arg1": "value1", "arg2": "value2"}},
22
+ {"name": "another_function", "arguments": {"arg": "value"}}
23
+ ]
24
+ </tool_call>
25
+ If no function calls are needed, respond normally without the tool_call tags.' }}
26
+ {%- for message in remaining_messages %}
27
+ {%- if message['role'] == 'user' %}
28
+ <|im_start|>user
29
+ {{ message['content'] }}<|im_end|>
30
+ {%- elif message['role'] == 'assistant' %}
31
+ {%- if message.content %}
32
+ <|im_start|>assistant
33
+ {{ message['content'] }}
34
+ <|im_end|>
35
+ {%- endif %}
36
+ {%- if message.tool_calls %}
37
+ <tool_call>
38
+ {{ message.tool_calls|tojson(indent=2) }}
39
+ </tool_call>
40
+ {%- endif %}
41
+ {%- elif message['role'] == 'tool' %}
42
+ <|im_start|>assistant
43
+ <tool_response>
44
+ {{ message['content'] }}
45
+ </tool_response><|im_end|>
46
+ {%- endif %}
47
+ {%- endfor %}
48
+ {{ '<|im_start|>assistant
49
+ ' if add_generation_prompt }}
50
+ {%- else %} {{bos_token}}{% for message in messages %} {{ '<|im_start|>' + message['role'] + '
51
+ ' + message['content'] + '<|im_end|>
52
+ ' }} {% endfor %} {% if add_generation_prompt %}{{ '<|im_start|>assistant
53
+ ' }}{% endif %} {%- endif %}
comparison_graph.png ADDED
config.json ADDED
@@ -0,0 +1,71 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "architectures": [
4
+ "FalconH1ForCausalLM"
5
+ ],
6
+ "attention_bias": false,
7
+ "attention_dropout": 0.0,
8
+ "attention_in_multiplier": 1.0,
9
+ "attention_out_multiplier": 1.0,
10
+ "bos_token_id": 1,
11
+ "dtype": "float16",
12
+ "embedding_multiplier": 0.034423828125,
13
+ "eos_token_id": 11,
14
+ "expansion_factor": 1.5,
15
+ "head_dim": 64,
16
+ "hidden_act": "silu",
17
+ "hidden_size": 512,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 768,
20
+ "key_multiplier": 1.0,
21
+ "lm_head_multiplier": 0.078125,
22
+ "mamba_chunk_size": 128,
23
+ "mamba_conv_bias": true,
24
+ "mamba_d_conv": 4,
25
+ "mamba_d_head": 32,
26
+ "mamba_d_ssm": 768,
27
+ "mamba_d_state": 64,
28
+ "mamba_expand": 2,
29
+ "mamba_n_groups": 1,
30
+ "mamba_n_heads": 24,
31
+ "mamba_norm_before_gate": false,
32
+ "mamba_proj_bias": false,
33
+ "mamba_rms_norm": false,
34
+ "mamba_use_mlp": true,
35
+ "max_position_embeddings": 262144,
36
+ "mlp_bias": false,
37
+ "mlp_multipliers": [
38
+ 1.0,
39
+ 1.0
40
+ ],
41
+ "model_type": "falcon_h1",
42
+ "num_attention_heads": 8,
43
+ "num_hidden_layers": 24,
44
+ "num_key_value_heads": 2,
45
+ "num_logits_to_keep": 1,
46
+ "pad_token_id": 0,
47
+ "projectors_bias": false,
48
+ "rms_norm_eps": 1e-05,
49
+ "rope_parameters": {
50
+ "rope_theta": 100000000000.0,
51
+ "rope_type": "default"
52
+ },
53
+ "sliding_window": null,
54
+ "ssm_in_multiplier": 1.0,
55
+ "ssm_multipliers": [
56
+ 1.0,
57
+ 1.0,
58
+ 1.0,
59
+ 1.0,
60
+ 1.0
61
+ ],
62
+ "ssm_out_multiplier": 1.0,
63
+ "tie_word_embeddings": true,
64
+ "time_step_floor": 0.0001,
65
+ "time_step_max": 0.1,
66
+ "time_step_min": 0.001,
67
+ "time_step_rank": "auto",
68
+ "transformers_version": "5.0.0",
69
+ "use_cache": true,
70
+ "vocab_size": 32768
71
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": false,
5
+ "eos_token_id": [
6
+ 11,
7
+ 228
8
+ ],
9
+ "pad_token_id": 0,
10
+ "transformers_version": "5.0.0"
11
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b45daac5015cdb4cd5c2037a0e966434c5f7be269a0e53bdcad312b7dd8296df
3
+ size 182303728
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,343 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|end_of_text|>",
6
+ "extra_special_tokens": [
7
+ "<|pad|>",
8
+ ">>ABSTRACT<<",
9
+ ">>INTRODUCTION<<",
10
+ ">>SUMMARY<<",
11
+ ">>COMMENT<<",
12
+ ">>ANSWER<<",
13
+ ">>QUESTION<<",
14
+ ">>DOMAIN<<",
15
+ ">>PREFIX<<",
16
+ ">>SUFFIX<<",
17
+ ">>MIDDLE<<",
18
+ "<|finetune_right_pad_id|>",
19
+ "<|start_header_id|>",
20
+ "<|end_header_id|>",
21
+ "<|eom_id|>",
22
+ "<|eot_id|>",
23
+ "<|begin_of_text|>",
24
+ ">>TITLE<<",
25
+ "<tool_response>",
26
+ "</tool_response>",
27
+ "<tool_call>",
28
+ "</tool_call>",
29
+ "<schema>",
30
+ "</schema>",
31
+ "<scratch_pad>",
32
+ "</scratch_pad>",
33
+ "<thinking>",
34
+ "</thinking>",
35
+ "<explanation>",
36
+ "</explanation>",
37
+ "<file_sep>",
38
+ "<repo_name>",
39
+ "<|im_end|>",
40
+ "<|im_start|>",
41
+ "<|system|>",
42
+ "<|file_sep|>",
43
+ "<|repo_name|>",
44
+ "<|repo_tree|>",
45
+ ">>UNUSED_221<<",
46
+ ">>UNUSED_222<<",
47
+ ">>UNUSED_223<<",
48
+ ">>UNUSED_224<<",
49
+ ">>UNUSED_225<<",
50
+ ">>UNUSED_226<<",
51
+ ">>UNUSED_227<<",
52
+ ">>UNUSED_228<<",
53
+ ">>UNUSED_229<<",
54
+ ">>UNUSED_230<<",
55
+ ">>UNUSED_231<<",
56
+ ">>UNUSED_232<<",
57
+ ">>UNUSED_233<<",
58
+ ">>UNUSED_234<<",
59
+ ">>UNUSED_235<<",
60
+ ">>UNUSED_236<<",
61
+ ">>UNUSED_237<<",
62
+ ">>UNUSED_238<<",
63
+ ">>UNUSED_239<<",
64
+ ">>UNUSED_240<<",
65
+ ">>UNUSED_241<<",
66
+ ">>UNUSED_242<<",
67
+ ">>UNUSED_243<<",
68
+ ">>UNUSED_244<<",
69
+ ">>UNUSED_245<<",
70
+ ">>UNUSED_246<<",
71
+ ">>UNUSED_247<<",
72
+ ">>UNUSED_248<<",
73
+ ">>UNUSED_249<<",
74
+ ">>UNUSED_250<<",
75
+ ">>UNUSED_251<<",
76
+ ">>UNUSED_252<<",
77
+ ">>UNUSED_253<<",
78
+ ">>UNUSED_254<<",
79
+ ">>UNUSED_255<<",
80
+ ">>UNUSED_256<<",
81
+ ">>UNUSED_257<<",
82
+ ">>UNUSED_258<<",
83
+ ">>UNUSED_259<<",
84
+ ">>UNUSED_260<<",
85
+ ">>UNUSED_261<<",
86
+ ">>UNUSED_262<<",
87
+ ">>UNUSED_263<<",
88
+ ">>UNUSED_264<<",
89
+ ">>UNUSED_265<<",
90
+ ">>UNUSED_266<<",
91
+ ">>UNUSED_267<<",
92
+ ">>UNUSED_268<<",
93
+ ">>UNUSED_269<<",
94
+ ">>UNUSED_270<<",
95
+ ">>UNUSED_271<<",
96
+ ">>UNUSED_272<<",
97
+ ">>UNUSED_273<<",
98
+ ">>UNUSED_274<<",
99
+ ">>UNUSED_275<<",
100
+ ">>UNUSED_276<<",
101
+ ">>UNUSED_277<<",
102
+ ">>UNUSED_278<<",
103
+ ">>UNUSED_279<<",
104
+ ">>UNUSED_280<<",
105
+ ">>UNUSED_281<<",
106
+ ">>UNUSED_282<<",
107
+ ">>UNUSED_283<<",
108
+ ">>UNUSED_284<<",
109
+ ">>UNUSED_285<<",
110
+ ">>UNUSED_286<<",
111
+ ">>UNUSED_287<<",
112
+ ">>UNUSED_288<<",
113
+ ">>UNUSED_289<<",
114
+ ">>UNUSED_290<<",
115
+ ">>UNUSED_291<<",
116
+ ">>UNUSED_292<<",
117
+ ">>UNUSED_293<<",
118
+ ">>UNUSED_294<<",
119
+ ">>UNUSED_295<<",
120
+ ">>UNUSED_296<<",
121
+ ">>UNUSED_297<<",
122
+ ">>UNUSED_298<<",
123
+ ">>UNUSED_299<<",
124
+ ">>UNUSED_300<<",
125
+ ">>UNUSED_301<<",
126
+ ">>UNUSED_302<<",
127
+ ">>UNUSED_303<<",
128
+ ">>UNUSED_304<<",
129
+ ">>UNUSED_305<<",
130
+ ">>UNUSED_306<<",
131
+ ">>UNUSED_307<<",
132
+ ">>UNUSED_308<<",
133
+ ">>UNUSED_309<<",
134
+ ">>UNUSED_310<<",
135
+ ">>UNUSED_311<<",
136
+ ">>UNUSED_312<<",
137
+ ">>UNUSED_313<<",
138
+ ">>UNUSED_314<<",
139
+ ">>UNUSED_315<<",
140
+ ">>UNUSED_316<<",
141
+ ">>UNUSED_317<<",
142
+ ">>UNUSED_318<<",
143
+ ">>UNUSED_319<<",
144
+ ">>UNUSED_320<<",
145
+ ">>UNUSED_321<<",
146
+ ">>UNUSED_322<<",
147
+ ">>UNUSED_323<<",
148
+ ">>UNUSED_324<<",
149
+ ">>UNUSED_325<<",
150
+ ">>UNUSED_326<<",
151
+ ">>UNUSED_327<<",
152
+ ">>UNUSED_328<<",
153
+ ">>UNUSED_329<<",
154
+ ">>UNUSED_330<<",
155
+ ">>UNUSED_331<<",
156
+ ">>UNUSED_332<<",
157
+ ">>UNUSED_333<<",
158
+ ">>UNUSED_334<<",
159
+ ">>UNUSED_335<<",
160
+ ">>UNUSED_336<<",
161
+ ">>UNUSED_337<<",
162
+ ">>UNUSED_338<<",
163
+ ">>UNUSED_339<<",
164
+ ">>UNUSED_340<<",
165
+ ">>UNUSED_341<<",
166
+ ">>UNUSED_342<<",
167
+ ">>UNUSED_343<<",
168
+ ">>UNUSED_344<<",
169
+ ">>UNUSED_345<<",
170
+ ">>UNUSED_346<<",
171
+ ">>UNUSED_347<<",
172
+ ">>UNUSED_348<<",
173
+ ">>UNUSED_349<<",
174
+ ">>UNUSED_350<<",
175
+ ">>UNUSED_351<<",
176
+ ">>UNUSED_352<<",
177
+ ">>UNUSED_353<<",
178
+ ">>UNUSED_354<<",
179
+ ">>UNUSED_355<<",
180
+ ">>UNUSED_356<<",
181
+ ">>UNUSED_357<<",
182
+ ">>UNUSED_358<<",
183
+ ">>UNUSED_359<<",
184
+ ">>UNUSED_360<<",
185
+ ">>UNUSED_361<<",
186
+ ">>UNUSED_362<<",
187
+ ">>UNUSED_363<<",
188
+ ">>UNUSED_364<<",
189
+ ">>UNUSED_365<<",
190
+ ">>UNUSED_366<<",
191
+ ">>UNUSED_367<<",
192
+ ">>UNUSED_368<<",
193
+ ">>UNUSED_369<<",
194
+ ">>UNUSED_370<<",
195
+ ">>UNUSED_371<<",
196
+ ">>UNUSED_372<<",
197
+ ">>UNUSED_373<<",
198
+ ">>UNUSED_374<<",
199
+ ">>UNUSED_375<<",
200
+ ">>UNUSED_376<<",
201
+ ">>UNUSED_377<<",
202
+ ">>UNUSED_378<<",
203
+ ">>UNUSED_379<<",
204
+ ">>UNUSED_380<<",
205
+ ">>UNUSED_381<<",
206
+ ">>UNUSED_382<<",
207
+ ">>UNUSED_383<<",
208
+ ">>UNUSED_384<<",
209
+ ">>UNUSED_385<<",
210
+ ">>UNUSED_386<<",
211
+ ">>UNUSED_387<<",
212
+ ">>UNUSED_388<<",
213
+ ">>UNUSED_389<<",
214
+ ">>UNUSED_390<<",
215
+ ">>UNUSED_391<<",
216
+ ">>UNUSED_392<<",
217
+ ">>UNUSED_393<<",
218
+ ">>UNUSED_394<<",
219
+ ">>UNUSED_395<<",
220
+ ">>UNUSED_396<<",
221
+ ">>UNUSED_397<<",
222
+ ">>UNUSED_398<<",
223
+ ">>UNUSED_399<<",
224
+ ">>UNUSED_400<<",
225
+ ">>UNUSED_401<<",
226
+ ">>UNUSED_402<<",
227
+ ">>UNUSED_403<<",
228
+ ">>UNUSED_404<<",
229
+ ">>UNUSED_405<<",
230
+ ">>UNUSED_406<<",
231
+ ">>UNUSED_407<<",
232
+ ">>UNUSED_408<<",
233
+ ">>UNUSED_409<<",
234
+ ">>UNUSED_410<<",
235
+ ">>UNUSED_411<<",
236
+ ">>UNUSED_412<<",
237
+ ">>UNUSED_413<<",
238
+ ">>UNUSED_414<<",
239
+ ">>UNUSED_415<<",
240
+ ">>UNUSED_416<<",
241
+ ">>UNUSED_417<<",
242
+ ">>UNUSED_418<<",
243
+ ">>UNUSED_419<<",
244
+ ">>UNUSED_420<<",
245
+ ">>UNUSED_421<<",
246
+ ">>UNUSED_422<<",
247
+ ">>UNUSED_423<<",
248
+ ">>UNUSED_424<<",
249
+ ">>UNUSED_425<<",
250
+ ">>UNUSED_426<<",
251
+ ">>UNUSED_427<<",
252
+ ">>UNUSED_428<<",
253
+ ">>UNUSED_429<<",
254
+ ">>UNUSED_430<<",
255
+ ">>UNUSED_431<<",
256
+ ">>UNUSED_432<<",
257
+ ">>UNUSED_433<<",
258
+ ">>UNUSED_434<<",
259
+ ">>UNUSED_435<<",
260
+ ">>UNUSED_436<<",
261
+ ">>UNUSED_437<<",
262
+ ">>UNUSED_438<<",
263
+ ">>UNUSED_439<<",
264
+ ">>UNUSED_440<<",
265
+ ">>UNUSED_441<<",
266
+ ">>UNUSED_442<<",
267
+ ">>UNUSED_443<<",
268
+ ">>UNUSED_444<<",
269
+ ">>UNUSED_445<<",
270
+ ">>UNUSED_446<<",
271
+ ">>UNUSED_447<<",
272
+ ">>UNUSED_448<<",
273
+ ">>UNUSED_449<<",
274
+ ">>UNUSED_450<<",
275
+ ">>UNUSED_451<<",
276
+ ">>UNUSED_452<<",
277
+ ">>UNUSED_453<<",
278
+ ">>UNUSED_454<<",
279
+ ">>UNUSED_455<<",
280
+ ">>UNUSED_456<<",
281
+ ">>UNUSED_457<<",
282
+ ">>UNUSED_458<<",
283
+ ">>UNUSED_459<<",
284
+ ">>UNUSED_460<<",
285
+ ">>UNUSED_461<<",
286
+ ">>UNUSED_462<<",
287
+ ">>UNUSED_463<<",
288
+ ">>UNUSED_464<<",
289
+ ">>UNUSED_465<<",
290
+ ">>UNUSED_466<<",
291
+ ">>UNUSED_467<<",
292
+ ">>UNUSED_468<<",
293
+ ">>UNUSED_469<<",
294
+ ">>UNUSED_470<<",
295
+ ">>UNUSED_471<<",
296
+ ">>UNUSED_472<<",
297
+ ">>UNUSED_473<<",
298
+ ">>UNUSED_474<<",
299
+ ">>UNUSED_475<<",
300
+ ">>UNUSED_476<<",
301
+ ">>UNUSED_477<<",
302
+ ">>UNUSED_478<<",
303
+ ">>UNUSED_479<<",
304
+ ">>UNUSED_480<<",
305
+ ">>UNUSED_481<<",
306
+ ">>UNUSED_482<<",
307
+ ">>UNUSED_483<<",
308
+ ">>UNUSED_484<<",
309
+ ">>UNUSED_485<<",
310
+ ">>UNUSED_486<<",
311
+ ">>UNUSED_487<<",
312
+ ">>UNUSED_488<<",
313
+ ">>UNUSED_489<<",
314
+ ">>UNUSED_490<<",
315
+ ">>UNUSED_491<<",
316
+ ">>UNUSED_492<<",
317
+ ">>UNUSED_493<<",
318
+ ">>UNUSED_494<<",
319
+ ">>UNUSED_495<<",
320
+ ">>UNUSED_496<<",
321
+ ">>UNUSED_497<<",
322
+ ">>UNUSED_498<<",
323
+ ">>UNUSED_499<<",
324
+ ">>UNUSED_500<<",
325
+ ">>UNUSED_501<<",
326
+ ">>UNUSED_502<<",
327
+ ">>UNUSED_503<<",
328
+ ">>UNUSED_504<<",
329
+ ">>UNUSED_505<<",
330
+ ">>UNUSED_506<<",
331
+ ">>UNUSED_507<<",
332
+ ">>UNUSED_508<<",
333
+ ">>UNUSED_509<<",
334
+ ">>UNUSED_510<<",
335
+ ">>UNUSED_511<<"
336
+ ],
337
+ "is_local": false,
338
+ "model_max_length": 1000000000000000019884624838656,
339
+ "model_specific_special_tokens": {},
340
+ "pad_token": "<|pad|>",
341
+ "pad_token_id": 0,
342
+ "tokenizer_class": "TokenizersBackend"
343
+ }