CompactAI commited on
Commit
b12b4c3
·
verified ·
1 Parent(s): 91b439d

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - pruned
5
+ - python
6
+ - optimized
7
+ - wanda
8
+ base_model: tiiuae/Falcon-H1-Tiny-90M-Instruct-Curriculum
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+ # Falcon-H1-Tiny-90M-Instruct-Curriculum-python-aggressive
13
+
14
+ > 🎯 **PYTHON-optimized** | 📦 **Aggressive** pruning | ⚡ **1% weights pruned**
15
+
16
+ This model is a **aggressively pruned** version of [tiiuae/Falcon-H1-Tiny-90M-Instruct-Curriculum](https://huggingface.co/tiiuae/Falcon-H1-Tiny-90M-Instruct-Curriculum).
17
+
18
+ ## Performance Comparison
19
+
20
+ | Category | Original | Pruned | Change |
21
+ |----------|----------|--------|--------|
22
+ | **Python** | 0.0% | 0.0% ⭐ | → |
23
+ | Html | 16.7% | 16.7% | → |
24
+ | Trivia | 16.7% | 16.7% | → |
25
+ | Math | 0.0% | 0.0% | → |
26
+ | Reasoning | 0.0% | 0.0% | → |
27
+ | Medical | 0.0% | 0.0% | → |
28
+ | Linux | 0.0% | 0.0% | → |
29
+ | Writing | 0.0% | 0.0% | → |
30
+
31
+ **Average**: 4.2% → 4.2% (+0.0%)
32
+
33
+
34
+
35
+ ![Comparison Graph](comparison_graph.png)
36
+
37
+ ## Quick Start
38
+
39
+ ```python
40
+ from transformers import AutoModelForCausalLM, AutoTokenizer
41
+
42
+ model = AutoModelForCausalLM.from_pretrained("CompactAI/Falcon-H1-Tiny-90M-Instruct-Curriculum-python-aggressive")
43
+ tokenizer = AutoTokenizer.from_pretrained("CompactAI/Falcon-H1-Tiny-90M-Instruct-Curriculum-python-aggressive")
44
+
45
+ inputs = tokenizer("Your prompt here", return_tensors="pt")
46
+ outputs = model.generate(**inputs, max_new_tokens=100)
47
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
48
+ ```
49
+
50
+ ## Technical Details
51
+
52
+ | Property | Value |
53
+ |----------|-------|
54
+ | Base Model | [tiiuae/Falcon-H1-Tiny-90M-Instruct-Curriculum](https://huggingface.co/tiiuae/Falcon-H1-Tiny-90M-Instruct-Curriculum) |
55
+ | Specialization | Python |
56
+ | Prune Mode | Aggressive |
57
+ | Weight Reduction | 1% weights pruned |
58
+
59
+ ## License
60
+
61
+ This model inherits the license from the base model.
chat_template.jinja ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {# --- System Prompt Handling --- #}
2
+ {%- if messages and messages[0]['role'] == 'system' %}
3
+ {%- set remaining_messages = messages[1:] %}
4
+ {%- else %}
5
+ {%- set remaining_messages = messages %}
6
+ {%- endif %}
7
+
8
+ {%- if tools %}
9
+ {{- '<|im_start|>system\n' }}
10
+ {%- if messages[0].role == 'system' %}
11
+ {{- messages[0].content + '\n' }}
12
+ {%- endif %}
13
+ # Tools
14
+ You may call one or more functions to assist with the user query. You are provided with function signatures within <tools></tools> XML tags.
15
+ <tools>
16
+ {%- for tool in tools %}
17
+ {{- "" }}
18
+ {{ tool | tojson }}
19
+ {%- endfor %}
20
+ {{- "" }}
21
+ </tools>
22
+ For each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:
23
+ <tool_call>
24
+ {"name": <function-name>, "arguments": <args-json-object>}
25
+ </tool_call>
26
+ {{- '<|im_end|>\n' }}
27
+ {%- else %}
28
+ {%- if messages[0].role == 'system' %}
29
+ {{- '<|im_start|>system\n' + messages[0].content + '<|im_end|>\n' }}
30
+ {%- endif %}
31
+ {%- endif %}
32
+
33
+ {# --- Render remaining messages --- #}
34
+ {%- for message in remaining_messages %}
35
+ {%- if message['role'] == 'user' %}
36
+ {{- '<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>\n' }}
37
+ {%- elif message['role'] == 'assistant' %}
38
+ {{- '<|im_start|>' + message['role'] +'\n' }}
39
+ {%- if message.get('content','') %}
40
+ {{- message['content'] + '\n' }}
41
+ {%- endif %}
42
+ {%- if tools and message.tool_calls %}
43
+ {%- for tool_call in message.tool_calls %}
44
+ {%- if tool_call.function is defined %}
45
+ {%- set tool_call = tool_call.function %}
46
+ {%- endif %}
47
+ {{-'<tool_call>\n' }}
48
+ {{- '{"name": "'+ tool_call.name + '", "arguments":' }}
49
+ {%- if tool_call.arguments is string -%}
50
+ {{ tool_call.arguments }}
51
+ {%- else -%}
52
+ {{ tool_call.arguments | tojson }}
53
+ {%- endif -%}
54
+ {{- '}' }}
55
+ {{- '\n</tool_call>\n' }}
56
+ {%- endfor %}
57
+ {%- endif %}
58
+ {%- if not loop.last %}
59
+ {{- '<|im_end|>' + '\n' }}
60
+ {%- else %}
61
+ {{- '<|im_end|>' }}
62
+ {%- endif %}
63
+ {%- elif message['role'] == 'tool' %}
64
+ {# Tool responses treated as user messages #}
65
+ {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != "tool") %}
66
+ {{- '<|im_start|>user' }}
67
+ {%- endif %}
68
+ {{- '\n<tool_response>\n' + message['content'] + '\n</tool_response>' }}
69
+ {%- if loop.last or (messages[loop.index0 + 1].role != "tool") %}
70
+ {{- '<|im_end|>\n' }}
71
+ {%- endif %}
72
+ {%- endif %}
73
+ {# --- Add generation prompt after last message if requested --- #}
74
+ {%- if loop.last and add_generation_prompt %}
75
+ {{- '<|im_start|>assistant\n' }}
76
+ {%- endif %}
77
+ {%- endfor %}
comparison_graph.png ADDED
config.json ADDED
@@ -0,0 +1,72 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "FalconH1ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_in_multiplier": 1.0,
8
+ "attention_out_multiplier": 1.0,
9
+ "attn_layer_indices": null,
10
+ "bos_token_id": 1,
11
+ "dtype": "float16",
12
+ "embedding_multiplier": 0.0751953125,
13
+ "eos_token_id": 11,
14
+ "expansion_factor": 1.5,
15
+ "head_dim": 64,
16
+ "hidden_act": "silu",
17
+ "hidden_size": 512,
18
+ "initializer_range": 0.02,
19
+ "intermediate_size": 768,
20
+ "key_multiplier": 1.0,
21
+ "lm_head_multiplier": 0.078125,
22
+ "mamba_chunk_size": 128,
23
+ "mamba_conv_bias": true,
24
+ "mamba_d_conv": 4,
25
+ "mamba_d_head": 32,
26
+ "mamba_d_ssm": 768,
27
+ "mamba_d_state": 64,
28
+ "mamba_expand": 2,
29
+ "mamba_n_groups": 1,
30
+ "mamba_n_heads": 24,
31
+ "mamba_norm_before_gate": false,
32
+ "mamba_proj_bias": false,
33
+ "mamba_rms_norm": false,
34
+ "mamba_use_mlp": true,
35
+ "max_position_embeddings": 262144,
36
+ "mlp_bias": false,
37
+ "mlp_expansion_factor": 8,
38
+ "mlp_multipliers": [
39
+ 1.0,
40
+ 1.0
41
+ ],
42
+ "model_type": "falcon_h1",
43
+ "num_attention_heads": 8,
44
+ "num_hidden_layers": 24,
45
+ "num_key_value_heads": 2,
46
+ "num_logits_to_keep": 1,
47
+ "pad_token_id": 0,
48
+ "projectors_bias": false,
49
+ "rms_norm_eps": 1e-05,
50
+ "rope_parameters": {
51
+ "rope_theta": 100000000000.0,
52
+ "rope_type": "default"
53
+ },
54
+ "sliding_window": null,
55
+ "ssm_in_multiplier": 1.0,
56
+ "ssm_multipliers": [
57
+ 1.0,
58
+ 1.0,
59
+ 1.0,
60
+ 1.0,
61
+ 1.0
62
+ ],
63
+ "ssm_out_multiplier": 1.0,
64
+ "tie_word_embeddings": true,
65
+ "time_step_floor": 0.0001,
66
+ "time_step_max": 0.1,
67
+ "time_step_min": 0.001,
68
+ "time_step_rank": "auto",
69
+ "transformers_version": "5.0.0",
70
+ "use_cache": true,
71
+ "vocab_size": 32768
72
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": false,
5
+ "eos_token_id": [
6
+ 228,
7
+ 11
8
+ ],
9
+ "pad_token_id": 0,
10
+ "transformers_version": "5.0.0"
11
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:65954c739c5ebc37fd73d8c97322efbb780f0783eb832689e2305d164f95e3fc
3
+ size 182303728
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,348 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|end_of_text|>",
6
+ "extra_special_tokens": [
7
+ "<|pad|>",
8
+ ">>ABSTRACT<<",
9
+ ">>INTRODUCTION<<",
10
+ ">>SUMMARY<<",
11
+ ">>COMMENT<<",
12
+ ">>ANSWER<<",
13
+ ">>QUESTION<<",
14
+ ">>DOMAIN<<",
15
+ ">>PREFIX<<",
16
+ ">>SUFFIX<<",
17
+ ">>MIDDLE<<",
18
+ "<|finetune_right_pad_id|>",
19
+ "<|start_header_id|>",
20
+ "<|end_header_id|>",
21
+ "<|eom_id|>",
22
+ "<|eot_id|>",
23
+ "<|begin_of_text|>",
24
+ ">>TITLE<<",
25
+ "<tool_response>",
26
+ "</tool_response>",
27
+ "<tool_call>",
28
+ "</tool_call>",
29
+ "<schema>",
30
+ "</schema>",
31
+ "<scratch_pad>",
32
+ "</scratch_pad>",
33
+ "<think>",
34
+ "</think>",
35
+ "<explanation>",
36
+ "</explanation>",
37
+ "<file_sep>",
38
+ "<repo_name>",
39
+ "<|im_end|>",
40
+ "<|im_start|>",
41
+ ">>UNUSED_119<<",
42
+ ">>UNUSED_120<<",
43
+ "<|system|>",
44
+ ">>UNUSED_218<<",
45
+ ">>UNUSED_219<<",
46
+ ">>UNUSED_220<<",
47
+ ">>UNUSED_221<<",
48
+ ">>UNUSED_222<<",
49
+ ">>UNUSED_223<<",
50
+ ">>UNUSED_224<<",
51
+ ">>UNUSED_225<<",
52
+ ">>UNUSED_226<<",
53
+ ">>UNUSED_227<<",
54
+ ">>UNUSED_228<<",
55
+ ">>UNUSED_229<<",
56
+ ">>UNUSED_230<<",
57
+ ">>UNUSED_231<<",
58
+ ">>UNUSED_232<<",
59
+ ">>UNUSED_233<<",
60
+ ">>UNUSED_234<<",
61
+ ">>UNUSED_235<<",
62
+ ">>UNUSED_236<<",
63
+ ">>UNUSED_237<<",
64
+ ">>UNUSED_238<<",
65
+ ">>UNUSED_239<<",
66
+ ">>UNUSED_240<<",
67
+ ">>UNUSED_241<<",
68
+ ">>UNUSED_242<<",
69
+ ">>UNUSED_243<<",
70
+ ">>UNUSED_244<<",
71
+ ">>UNUSED_245<<",
72
+ ">>UNUSED_246<<",
73
+ ">>UNUSED_247<<",
74
+ ">>UNUSED_248<<",
75
+ ">>UNUSED_249<<",
76
+ ">>UNUSED_250<<",
77
+ ">>UNUSED_251<<",
78
+ ">>UNUSED_252<<",
79
+ ">>UNUSED_253<<",
80
+ ">>UNUSED_254<<",
81
+ ">>UNUSED_255<<",
82
+ ">>UNUSED_256<<",
83
+ ">>UNUSED_257<<",
84
+ ">>UNUSED_258<<",
85
+ ">>UNUSED_259<<",
86
+ ">>UNUSED_260<<",
87
+ ">>UNUSED_261<<",
88
+ ">>UNUSED_262<<",
89
+ ">>UNUSED_263<<",
90
+ ">>UNUSED_264<<",
91
+ ">>UNUSED_265<<",
92
+ ">>UNUSED_266<<",
93
+ ">>UNUSED_267<<",
94
+ ">>UNUSED_268<<",
95
+ ">>UNUSED_269<<",
96
+ ">>UNUSED_270<<",
97
+ ">>UNUSED_271<<",
98
+ ">>UNUSED_272<<",
99
+ ">>UNUSED_273<<",
100
+ ">>UNUSED_274<<",
101
+ ">>UNUSED_275<<",
102
+ ">>UNUSED_276<<",
103
+ ">>UNUSED_277<<",
104
+ ">>UNUSED_278<<",
105
+ ">>UNUSED_279<<",
106
+ ">>UNUSED_280<<",
107
+ ">>UNUSED_281<<",
108
+ ">>UNUSED_282<<",
109
+ ">>UNUSED_283<<",
110
+ ">>UNUSED_284<<",
111
+ ">>UNUSED_285<<",
112
+ ">>UNUSED_286<<",
113
+ ">>UNUSED_287<<",
114
+ ">>UNUSED_288<<",
115
+ ">>UNUSED_289<<",
116
+ ">>UNUSED_290<<",
117
+ ">>UNUSED_291<<",
118
+ ">>UNUSED_292<<",
119
+ ">>UNUSED_293<<",
120
+ ">>UNUSED_294<<",
121
+ ">>UNUSED_295<<",
122
+ ">>UNUSED_296<<",
123
+ ">>UNUSED_297<<",
124
+ ">>UNUSED_298<<",
125
+ ">>UNUSED_299<<",
126
+ ">>UNUSED_300<<",
127
+ ">>UNUSED_301<<",
128
+ ">>UNUSED_302<<",
129
+ ">>UNUSED_303<<",
130
+ ">>UNUSED_304<<",
131
+ ">>UNUSED_305<<",
132
+ ">>UNUSED_306<<",
133
+ ">>UNUSED_307<<",
134
+ ">>UNUSED_308<<",
135
+ ">>UNUSED_309<<",
136
+ ">>UNUSED_310<<",
137
+ ">>UNUSED_311<<",
138
+ ">>UNUSED_312<<",
139
+ ">>UNUSED_313<<",
140
+ ">>UNUSED_314<<",
141
+ ">>UNUSED_315<<",
142
+ ">>UNUSED_316<<",
143
+ ">>UNUSED_317<<",
144
+ ">>UNUSED_318<<",
145
+ ">>UNUSED_319<<",
146
+ ">>UNUSED_320<<",
147
+ ">>UNUSED_321<<",
148
+ ">>UNUSED_322<<",
149
+ ">>UNUSED_323<<",
150
+ ">>UNUSED_324<<",
151
+ ">>UNUSED_325<<",
152
+ ">>UNUSED_326<<",
153
+ ">>UNUSED_327<<",
154
+ ">>UNUSED_328<<",
155
+ ">>UNUSED_329<<",
156
+ ">>UNUSED_330<<",
157
+ ">>UNUSED_331<<",
158
+ ">>UNUSED_332<<",
159
+ ">>UNUSED_333<<",
160
+ ">>UNUSED_334<<",
161
+ ">>UNUSED_335<<",
162
+ ">>UNUSED_336<<",
163
+ ">>UNUSED_337<<",
164
+ ">>UNUSED_338<<",
165
+ ">>UNUSED_339<<",
166
+ ">>UNUSED_340<<",
167
+ ">>UNUSED_341<<",
168
+ ">>UNUSED_342<<",
169
+ ">>UNUSED_343<<",
170
+ ">>UNUSED_344<<",
171
+ ">>UNUSED_345<<",
172
+ ">>UNUSED_346<<",
173
+ ">>UNUSED_347<<",
174
+ ">>UNUSED_348<<",
175
+ ">>UNUSED_349<<",
176
+ ">>UNUSED_350<<",
177
+ ">>UNUSED_351<<",
178
+ ">>UNUSED_352<<",
179
+ ">>UNUSED_353<<",
180
+ ">>UNUSED_354<<",
181
+ ">>UNUSED_355<<",
182
+ ">>UNUSED_356<<",
183
+ ">>UNUSED_357<<",
184
+ ">>UNUSED_358<<",
185
+ ">>UNUSED_359<<",
186
+ ">>UNUSED_360<<",
187
+ ">>UNUSED_361<<",
188
+ ">>UNUSED_362<<",
189
+ ">>UNUSED_363<<",
190
+ ">>UNUSED_364<<",
191
+ ">>UNUSED_365<<",
192
+ ">>UNUSED_366<<",
193
+ ">>UNUSED_367<<",
194
+ ">>UNUSED_368<<",
195
+ ">>UNUSED_369<<",
196
+ ">>UNUSED_370<<",
197
+ ">>UNUSED_371<<",
198
+ ">>UNUSED_372<<",
199
+ ">>UNUSED_373<<",
200
+ ">>UNUSED_374<<",
201
+ ">>UNUSED_375<<",
202
+ ">>UNUSED_376<<",
203
+ ">>UNUSED_377<<",
204
+ ">>UNUSED_378<<",
205
+ ">>UNUSED_379<<",
206
+ ">>UNUSED_380<<",
207
+ ">>UNUSED_381<<",
208
+ ">>UNUSED_382<<",
209
+ ">>UNUSED_383<<",
210
+ ">>UNUSED_384<<",
211
+ ">>UNUSED_385<<",
212
+ ">>UNUSED_386<<",
213
+ ">>UNUSED_387<<",
214
+ ">>UNUSED_388<<",
215
+ ">>UNUSED_389<<",
216
+ ">>UNUSED_390<<",
217
+ ">>UNUSED_391<<",
218
+ ">>UNUSED_392<<",
219
+ ">>UNUSED_393<<",
220
+ ">>UNUSED_394<<",
221
+ ">>UNUSED_395<<",
222
+ ">>UNUSED_396<<",
223
+ ">>UNUSED_397<<",
224
+ ">>UNUSED_398<<",
225
+ ">>UNUSED_399<<",
226
+ ">>UNUSED_400<<",
227
+ ">>UNUSED_401<<",
228
+ ">>UNUSED_402<<",
229
+ ">>UNUSED_403<<",
230
+ ">>UNUSED_404<<",
231
+ ">>UNUSED_405<<",
232
+ ">>UNUSED_406<<",
233
+ ">>UNUSED_407<<",
234
+ ">>UNUSED_408<<",
235
+ ">>UNUSED_409<<",
236
+ ">>UNUSED_410<<",
237
+ ">>UNUSED_411<<",
238
+ ">>UNUSED_412<<",
239
+ ">>UNUSED_413<<",
240
+ ">>UNUSED_414<<",
241
+ ">>UNUSED_415<<",
242
+ ">>UNUSED_416<<",
243
+ ">>UNUSED_417<<",
244
+ ">>UNUSED_418<<",
245
+ ">>UNUSED_419<<",
246
+ ">>UNUSED_420<<",
247
+ ">>UNUSED_421<<",
248
+ ">>UNUSED_422<<",
249
+ ">>UNUSED_423<<",
250
+ ">>UNUSED_424<<",
251
+ ">>UNUSED_425<<",
252
+ ">>UNUSED_426<<",
253
+ ">>UNUSED_427<<",
254
+ ">>UNUSED_428<<",
255
+ ">>UNUSED_429<<",
256
+ ">>UNUSED_430<<",
257
+ ">>UNUSED_431<<",
258
+ ">>UNUSED_432<<",
259
+ ">>UNUSED_433<<",
260
+ ">>UNUSED_434<<",
261
+ ">>UNUSED_435<<",
262
+ ">>UNUSED_436<<",
263
+ ">>UNUSED_437<<",
264
+ ">>UNUSED_438<<",
265
+ ">>UNUSED_439<<",
266
+ ">>UNUSED_440<<",
267
+ ">>UNUSED_441<<",
268
+ ">>UNUSED_442<<",
269
+ ">>UNUSED_443<<",
270
+ ">>UNUSED_444<<",
271
+ ">>UNUSED_445<<",
272
+ ">>UNUSED_446<<",
273
+ ">>UNUSED_447<<",
274
+ ">>UNUSED_448<<",
275
+ ">>UNUSED_449<<",
276
+ ">>UNUSED_450<<",
277
+ ">>UNUSED_451<<",
278
+ ">>UNUSED_452<<",
279
+ ">>UNUSED_453<<",
280
+ ">>UNUSED_454<<",
281
+ ">>UNUSED_455<<",
282
+ ">>UNUSED_456<<",
283
+ ">>UNUSED_457<<",
284
+ ">>UNUSED_458<<",
285
+ ">>UNUSED_459<<",
286
+ ">>UNUSED_460<<",
287
+ ">>UNUSED_461<<",
288
+ ">>UNUSED_462<<",
289
+ ">>UNUSED_463<<",
290
+ ">>UNUSED_464<<",
291
+ ">>UNUSED_465<<",
292
+ ">>UNUSED_466<<",
293
+ ">>UNUSED_467<<",
294
+ ">>UNUSED_468<<",
295
+ ">>UNUSED_469<<",
296
+ ">>UNUSED_470<<",
297
+ ">>UNUSED_471<<",
298
+ ">>UNUSED_472<<",
299
+ ">>UNUSED_473<<",
300
+ ">>UNUSED_474<<",
301
+ ">>UNUSED_475<<",
302
+ ">>UNUSED_476<<",
303
+ ">>UNUSED_477<<",
304
+ ">>UNUSED_478<<",
305
+ ">>UNUSED_479<<",
306
+ ">>UNUSED_480<<",
307
+ ">>UNUSED_481<<",
308
+ ">>UNUSED_482<<",
309
+ ">>UNUSED_483<<",
310
+ ">>UNUSED_484<<",
311
+ ">>UNUSED_485<<",
312
+ ">>UNUSED_486<<",
313
+ ">>UNUSED_487<<",
314
+ ">>UNUSED_488<<",
315
+ ">>UNUSED_489<<",
316
+ ">>UNUSED_490<<",
317
+ ">>UNUSED_491<<",
318
+ ">>UNUSED_492<<",
319
+ ">>UNUSED_493<<",
320
+ ">>UNUSED_494<<",
321
+ ">>UNUSED_495<<",
322
+ ">>UNUSED_496<<",
323
+ ">>UNUSED_497<<",
324
+ ">>UNUSED_498<<",
325
+ ">>UNUSED_499<<",
326
+ ">>UNUSED_500<<",
327
+ ">>UNUSED_501<<",
328
+ ">>UNUSED_502<<",
329
+ ">>UNUSED_503<<",
330
+ ">>UNUSED_504<<",
331
+ ">>UNUSED_505<<",
332
+ ">>UNUSED_506<<",
333
+ ">>UNUSED_507<<",
334
+ ">>UNUSED_508<<",
335
+ ">>UNUSED_509<<",
336
+ ">>UNUSED_510<<",
337
+ ">>UNUSED_511<<"
338
+ ],
339
+ "is_local": false,
340
+ "model_input_names": [
341
+ "input_ids",
342
+ "attention_mask"
343
+ ],
344
+ "model_max_length": 1000000000000000019884624838656,
345
+ "model_specific_special_tokens": {},
346
+ "pad_token": "<|pad|>",
347
+ "tokenizer_class": "TokenizersBackend"
348
+ }