CompactAI commited on
Commit
11a9df8
·
verified ·
1 Parent(s): 2f32f92

Upload folder using huggingface_hub

Browse files
README.md ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - pruned
5
+ - python
6
+ - optimized
7
+ - wanda
8
+ base_model: tiiuae/Falcon-H1-Tiny-Multilingual-100M-Instruct
9
+ pipeline_tag: text-generation
10
+ ---
11
+
12
+ # Falcon-H1-Tiny-Multilingual-100M-Instruct-python-safe
13
+
14
+ > 🎯 **PYTHON-optimized** | 📦 **Safe** pruning | ⚡ **1% weights pruned**
15
+
16
+ This model is a **conservatively pruned** version of [tiiuae/Falcon-H1-Tiny-Multilingual-100M-Instruct](https://huggingface.co/tiiuae/Falcon-H1-Tiny-Multilingual-100M-Instruct).
17
+
18
+ ## Performance Comparison
19
+
20
+ | Category | Original | Pruned | Change |
21
+ |----------|----------|--------|--------|
22
+ | **Python** | 0.0% | 0.0% ⭐ | → |
23
+ | Html | 0.0% | 0.0% | → |
24
+ | Trivia | 16.7% | 16.7% | → |
25
+ | Math | 0.0% | 0.0% | → |
26
+ | Reasoning | 0.0% | 0.0% | → |
27
+ | Medical | 33.3% | 33.3% | → |
28
+ | Linux | 0.0% | 0.0% | → |
29
+ | Writing | 0.0% | 0.0% | → |
30
+
31
+ **Average**: 6.2% → 6.2% (+0.0%)
32
+
33
+
34
+
35
+ ![Comparison Graph](comparison_graph.png)
36
+
37
+ ## Quick Start
38
+
39
+ ```python
40
+ from transformers import AutoModelForCausalLM, AutoTokenizer
41
+
42
+ model = AutoModelForCausalLM.from_pretrained("CompactAI/Falcon-H1-Tiny-Multilingual-100M-Instruct-python-safe")
43
+ tokenizer = AutoTokenizer.from_pretrained("CompactAI/Falcon-H1-Tiny-Multilingual-100M-Instruct-python-safe")
44
+
45
+ inputs = tokenizer("Your prompt here", return_tensors="pt")
46
+ outputs = model.generate(**inputs, max_new_tokens=100)
47
+ print(tokenizer.decode(outputs[0], skip_special_tokens=True))
48
+ ```
49
+
50
+ ## Technical Details
51
+
52
+ | Property | Value |
53
+ |----------|-------|
54
+ | Base Model | [tiiuae/Falcon-H1-Tiny-Multilingual-100M-Instruct](https://huggingface.co/tiiuae/Falcon-H1-Tiny-Multilingual-100M-Instruct) |
55
+ | Specialization | Python |
56
+ | Prune Mode | Safe |
57
+ | Weight Reduction | 1% weights pruned |
58
+
59
+ ## License
60
+
61
+ This model inherits the license from the base model.
chat_template.jinja ADDED
@@ -0,0 +1,53 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {%- if tools %} {{bos_token}}<|system|>
2
+ {%- if messages[0]['role'] == 'system' %}
3
+ {{ messages[0]['content'] }}
4
+ {%- set remaining_messages = messages[1:] %}
5
+ {%- else %}
6
+ {%- set remaining_messages = messages %}
7
+ {%- endif %}
8
+ {{ 'You are a Falcon assistant skilled in function calling. You are helpful, respectful, and concise.
9
+
10
+ # Tools
11
+
12
+ You have access to the following functions. You MUST use them to answer questions when needed. For each function call, you MUST return a JSON object inside <tool_call></tool_call> tags.
13
+
14
+ <tools>' + tools|tojson(indent=2) + '</tools>
15
+
16
+ # Output Format
17
+
18
+ Your response MUST follow this format when making function calls:
19
+ <tool_call>
20
+ [
21
+ {"name": "function_name", "arguments": {"arg1": "value1", "arg2": "value2"}},
22
+ {"name": "another_function", "arguments": {"arg": "value"}}
23
+ ]
24
+ </tool_call>
25
+ If no function calls are needed, respond normally without the tool_call tags.' }}
26
+ {%- for message in remaining_messages %}
27
+ {%- if message['role'] == 'user' %}
28
+ <|im_start|>user
29
+ {{ message['content'] }}<|im_end|>
30
+ {%- elif message['role'] == 'assistant' %}
31
+ {%- if message.content %}
32
+ <|im_start|>assistant
33
+ {{ message['content'] }}
34
+ <|im_end|>
35
+ {%- endif %}
36
+ {%- if message.tool_calls %}
37
+ <tool_call>
38
+ {{ message.tool_calls|tojson(indent=2) }}
39
+ </tool_call>
40
+ {%- endif %}
41
+ {%- elif message['role'] == 'tool' %}
42
+ <|im_start|>assistant
43
+ <tool_response>
44
+ {{ message['content'] }}
45
+ </tool_response><|im_end|>
46
+ {%- endif %}
47
+ {%- endfor %}
48
+ {{ '<|im_start|>assistant
49
+ ' if add_generation_prompt }}
50
+ {%- else %} {{bos_token}}{% for message in messages %} {{ '<|im_start|>' + message['role'] + '
51
+ ' + message['content'] + '<|im_end|>
52
+ ' }} {% endfor %} {% if add_generation_prompt %}{{ '<|im_start|>assistant
53
+ ' }}{% endif %} {%- endif %}
comparison_graph.png ADDED
config.json ADDED
@@ -0,0 +1,70 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "architectures": [
3
+ "FalconH1ForCausalLM"
4
+ ],
5
+ "attention_bias": false,
6
+ "attention_dropout": 0.0,
7
+ "attention_in_multiplier": 1.0,
8
+ "attention_out_multiplier": 1.0,
9
+ "bos_token_id": 1,
10
+ "dtype": "float16",
11
+ "embedding_multiplier": 0.10888671875,
12
+ "eos_token_id": 11,
13
+ "expansion_factor": 1.5,
14
+ "head_dim": 64,
15
+ "hidden_act": "silu",
16
+ "hidden_size": 512,
17
+ "initializer_range": 0.02,
18
+ "intermediate_size": 768,
19
+ "key_multiplier": 1.0,
20
+ "lm_head_multiplier": 0.078125,
21
+ "mamba_chunk_size": 128,
22
+ "mamba_conv_bias": true,
23
+ "mamba_d_conv": 4,
24
+ "mamba_d_head": 32,
25
+ "mamba_d_ssm": 768,
26
+ "mamba_d_state": 64,
27
+ "mamba_expand": 2,
28
+ "mamba_n_groups": 1,
29
+ "mamba_n_heads": 24,
30
+ "mamba_norm_before_gate": false,
31
+ "mamba_proj_bias": false,
32
+ "mamba_rms_norm": false,
33
+ "mamba_use_mlp": true,
34
+ "max_position_embeddings": 262144,
35
+ "mlp_bias": false,
36
+ "mlp_multipliers": [
37
+ 1.0,
38
+ 1.0
39
+ ],
40
+ "model_type": "falcon_h1",
41
+ "num_attention_heads": 8,
42
+ "num_hidden_layers": 24,
43
+ "num_key_value_heads": 2,
44
+ "num_logits_to_keep": 1,
45
+ "pad_token_id": 0,
46
+ "projectors_bias": false,
47
+ "rms_norm_eps": 1e-05,
48
+ "rope_parameters": {
49
+ "rope_theta": 100000000000.0,
50
+ "rope_type": "default"
51
+ },
52
+ "sliding_window": null,
53
+ "ssm_in_multiplier": 1.0,
54
+ "ssm_multipliers": [
55
+ 1.0,
56
+ 1.0,
57
+ 1.0,
58
+ 1.0,
59
+ 1.0
60
+ ],
61
+ "ssm_out_multiplier": 1.0,
62
+ "tie_word_embeddings": true,
63
+ "time_step_floor": 0.0001,
64
+ "time_step_max": 0.1,
65
+ "time_step_min": 0.001,
66
+ "time_step_rank": "auto",
67
+ "transformers_version": "5.0.0",
68
+ "use_cache": true,
69
+ "vocab_size": 65536
70
+ }
generation_config.json ADDED
@@ -0,0 +1,11 @@
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "_from_model_config": true,
3
+ "bos_token_id": 1,
4
+ "do_sample": false,
5
+ "eos_token_id": [
6
+ 11,
7
+ 228
8
+ ],
9
+ "pad_token_id": 0,
10
+ "transformers_version": "5.0.0"
11
+ }
model.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3e1935aaee88e35ca489003adf8c1c8931194b97023def429443ac09412f987b
3
+ size 215858336
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json ADDED
@@ -0,0 +1,344 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "backend": "tokenizers",
3
+ "bos_token": "<|begin_of_text|>",
4
+ "clean_up_tokenization_spaces": true,
5
+ "eos_token": "<|end_of_text|>",
6
+ "extra_special_tokens": [
7
+ "<|pad|>",
8
+ ">>ABSTRACT<<",
9
+ ">>INTRODUCTION<<",
10
+ ">>SUMMARY<<",
11
+ ">>COMMENT<<",
12
+ ">>ANSWER<<",
13
+ ">>QUESTION<<",
14
+ ">>DOMAIN<<",
15
+ ">>PREFIX<<",
16
+ ">>SUFFIX<<",
17
+ ">>MIDDLE<<",
18
+ "<|finetune_right_pad_id|>",
19
+ "<|start_header_id|>",
20
+ "<|end_header_id|>",
21
+ "<|eom_id|>",
22
+ "<|eot_id|>",
23
+ "<|begin_of_text|>",
24
+ ">>TITLE<<",
25
+ "<tool_response>",
26
+ "</tool_response>",
27
+ "<tool_call>",
28
+ "</tool_call>",
29
+ "<schema>",
30
+ "</schema>",
31
+ "<scratch_pad>",
32
+ "</scratch_pad>",
33
+ "<thinking>",
34
+ "</thinking>",
35
+ "<explanation>",
36
+ "</explanation>",
37
+ "<file_sep>",
38
+ "<repo_name>",
39
+ "<|im_end|>",
40
+ "<|im_start|>",
41
+ ">>UNUSED_119<<",
42
+ ">>UNUSED_120<<",
43
+ "<|system|>",
44
+ "<|file_sep|>",
45
+ "<|repo_name|>",
46
+ "<|repo_tree|>",
47
+ ">>UNUSED_221<<",
48
+ ">>UNUSED_222<<",
49
+ ">>UNUSED_223<<",
50
+ ">>UNUSED_224<<",
51
+ ">>UNUSED_225<<",
52
+ ">>UNUSED_226<<",
53
+ ">>UNUSED_227<<",
54
+ ">>UNUSED_228<<",
55
+ ">>UNUSED_229<<",
56
+ ">>UNUSED_230<<",
57
+ ">>UNUSED_231<<",
58
+ ">>UNUSED_232<<",
59
+ ">>UNUSED_233<<",
60
+ ">>UNUSED_234<<",
61
+ ">>UNUSED_235<<",
62
+ ">>UNUSED_236<<",
63
+ ">>UNUSED_237<<",
64
+ ">>UNUSED_238<<",
65
+ ">>UNUSED_239<<",
66
+ ">>UNUSED_240<<",
67
+ ">>UNUSED_241<<",
68
+ ">>UNUSED_242<<",
69
+ ">>UNUSED_243<<",
70
+ ">>UNUSED_244<<",
71
+ ">>UNUSED_245<<",
72
+ ">>UNUSED_246<<",
73
+ ">>UNUSED_247<<",
74
+ ">>UNUSED_248<<",
75
+ ">>UNUSED_249<<",
76
+ ">>UNUSED_250<<",
77
+ ">>UNUSED_251<<",
78
+ ">>UNUSED_252<<",
79
+ ">>UNUSED_253<<",
80
+ ">>UNUSED_254<<",
81
+ ">>UNUSED_255<<",
82
+ ">>UNUSED_256<<",
83
+ ">>UNUSED_257<<",
84
+ ">>UNUSED_258<<",
85
+ ">>UNUSED_259<<",
86
+ ">>UNUSED_260<<",
87
+ ">>UNUSED_261<<",
88
+ ">>UNUSED_262<<",
89
+ ">>UNUSED_263<<",
90
+ ">>UNUSED_264<<",
91
+ ">>UNUSED_265<<",
92
+ ">>UNUSED_266<<",
93
+ ">>UNUSED_267<<",
94
+ ">>UNUSED_268<<",
95
+ ">>UNUSED_269<<",
96
+ ">>UNUSED_270<<",
97
+ ">>UNUSED_271<<",
98
+ ">>UNUSED_272<<",
99
+ ">>UNUSED_273<<",
100
+ ">>UNUSED_274<<",
101
+ ">>UNUSED_275<<",
102
+ ">>UNUSED_276<<",
103
+ ">>UNUSED_277<<",
104
+ ">>UNUSED_278<<",
105
+ ">>UNUSED_279<<",
106
+ ">>UNUSED_280<<",
107
+ ">>UNUSED_281<<",
108
+ ">>UNUSED_282<<",
109
+ ">>UNUSED_283<<",
110
+ ">>UNUSED_284<<",
111
+ ">>UNUSED_285<<",
112
+ ">>UNUSED_286<<",
113
+ ">>UNUSED_287<<",
114
+ ">>UNUSED_288<<",
115
+ ">>UNUSED_289<<",
116
+ ">>UNUSED_290<<",
117
+ ">>UNUSED_291<<",
118
+ ">>UNUSED_292<<",
119
+ ">>UNUSED_293<<",
120
+ ">>UNUSED_294<<",
121
+ ">>UNUSED_295<<",
122
+ ">>UNUSED_296<<",
123
+ ">>UNUSED_297<<",
124
+ ">>UNUSED_298<<",
125
+ ">>UNUSED_299<<",
126
+ ">>UNUSED_300<<",
127
+ ">>UNUSED_301<<",
128
+ ">>UNUSED_302<<",
129
+ ">>UNUSED_303<<",
130
+ ">>UNUSED_304<<",
131
+ ">>UNUSED_305<<",
132
+ ">>UNUSED_306<<",
133
+ ">>UNUSED_307<<",
134
+ ">>UNUSED_308<<",
135
+ ">>UNUSED_309<<",
136
+ ">>UNUSED_310<<",
137
+ ">>UNUSED_311<<",
138
+ ">>UNUSED_312<<",
139
+ ">>UNUSED_313<<",
140
+ ">>UNUSED_314<<",
141
+ ">>UNUSED_315<<",
142
+ ">>UNUSED_316<<",
143
+ ">>UNUSED_317<<",
144
+ ">>UNUSED_318<<",
145
+ ">>UNUSED_319<<",
146
+ ">>UNUSED_320<<",
147
+ ">>UNUSED_321<<",
148
+ ">>UNUSED_322<<",
149
+ ">>UNUSED_323<<",
150
+ ">>UNUSED_324<<",
151
+ ">>UNUSED_325<<",
152
+ ">>UNUSED_326<<",
153
+ ">>UNUSED_327<<",
154
+ ">>UNUSED_328<<",
155
+ ">>UNUSED_329<<",
156
+ ">>UNUSED_330<<",
157
+ ">>UNUSED_331<<",
158
+ ">>UNUSED_332<<",
159
+ ">>UNUSED_333<<",
160
+ ">>UNUSED_334<<",
161
+ ">>UNUSED_335<<",
162
+ ">>UNUSED_336<<",
163
+ ">>UNUSED_337<<",
164
+ ">>UNUSED_338<<",
165
+ ">>UNUSED_339<<",
166
+ ">>UNUSED_340<<",
167
+ ">>UNUSED_341<<",
168
+ ">>UNUSED_342<<",
169
+ ">>UNUSED_343<<",
170
+ ">>UNUSED_344<<",
171
+ ">>UNUSED_345<<",
172
+ ">>UNUSED_346<<",
173
+ ">>UNUSED_347<<",
174
+ ">>UNUSED_348<<",
175
+ ">>UNUSED_349<<",
176
+ ">>UNUSED_350<<",
177
+ ">>UNUSED_351<<",
178
+ ">>UNUSED_352<<",
179
+ ">>UNUSED_353<<",
180
+ ">>UNUSED_354<<",
181
+ ">>UNUSED_355<<",
182
+ ">>UNUSED_356<<",
183
+ ">>UNUSED_357<<",
184
+ ">>UNUSED_358<<",
185
+ ">>UNUSED_359<<",
186
+ ">>UNUSED_360<<",
187
+ ">>UNUSED_361<<",
188
+ ">>UNUSED_362<<",
189
+ ">>UNUSED_363<<",
190
+ ">>UNUSED_364<<",
191
+ ">>UNUSED_365<<",
192
+ ">>UNUSED_366<<",
193
+ ">>UNUSED_367<<",
194
+ ">>UNUSED_368<<",
195
+ ">>UNUSED_369<<",
196
+ ">>UNUSED_370<<",
197
+ ">>UNUSED_371<<",
198
+ ">>UNUSED_372<<",
199
+ ">>UNUSED_373<<",
200
+ ">>UNUSED_374<<",
201
+ ">>UNUSED_375<<",
202
+ ">>UNUSED_376<<",
203
+ ">>UNUSED_377<<",
204
+ ">>UNUSED_378<<",
205
+ ">>UNUSED_379<<",
206
+ ">>UNUSED_380<<",
207
+ ">>UNUSED_381<<",
208
+ ">>UNUSED_382<<",
209
+ ">>UNUSED_383<<",
210
+ ">>UNUSED_384<<",
211
+ ">>UNUSED_385<<",
212
+ ">>UNUSED_386<<",
213
+ ">>UNUSED_387<<",
214
+ ">>UNUSED_388<<",
215
+ ">>UNUSED_389<<",
216
+ ">>UNUSED_390<<",
217
+ ">>UNUSED_391<<",
218
+ ">>UNUSED_392<<",
219
+ ">>UNUSED_393<<",
220
+ ">>UNUSED_394<<",
221
+ ">>UNUSED_395<<",
222
+ ">>UNUSED_396<<",
223
+ ">>UNUSED_397<<",
224
+ ">>UNUSED_398<<",
225
+ ">>UNUSED_399<<",
226
+ ">>UNUSED_400<<",
227
+ ">>UNUSED_401<<",
228
+ ">>UNUSED_402<<",
229
+ ">>UNUSED_403<<",
230
+ ">>UNUSED_404<<",
231
+ ">>UNUSED_405<<",
232
+ ">>UNUSED_406<<",
233
+ ">>UNUSED_407<<",
234
+ ">>UNUSED_408<<",
235
+ ">>UNUSED_409<<",
236
+ ">>UNUSED_410<<",
237
+ ">>UNUSED_411<<",
238
+ ">>UNUSED_412<<",
239
+ ">>UNUSED_413<<",
240
+ ">>UNUSED_414<<",
241
+ ">>UNUSED_415<<",
242
+ ">>UNUSED_416<<",
243
+ ">>UNUSED_417<<",
244
+ ">>UNUSED_418<<",
245
+ ">>UNUSED_419<<",
246
+ ">>UNUSED_420<<",
247
+ ">>UNUSED_421<<",
248
+ ">>UNUSED_422<<",
249
+ ">>UNUSED_423<<",
250
+ ">>UNUSED_424<<",
251
+ ">>UNUSED_425<<",
252
+ ">>UNUSED_426<<",
253
+ ">>UNUSED_427<<",
254
+ ">>UNUSED_428<<",
255
+ ">>UNUSED_429<<",
256
+ ">>UNUSED_430<<",
257
+ ">>UNUSED_431<<",
258
+ ">>UNUSED_432<<",
259
+ ">>UNUSED_433<<",
260
+ ">>UNUSED_434<<",
261
+ ">>UNUSED_435<<",
262
+ ">>UNUSED_436<<",
263
+ ">>UNUSED_437<<",
264
+ ">>UNUSED_438<<",
265
+ ">>UNUSED_439<<",
266
+ ">>UNUSED_440<<",
267
+ ">>UNUSED_441<<",
268
+ ">>UNUSED_442<<",
269
+ ">>UNUSED_443<<",
270
+ ">>UNUSED_444<<",
271
+ ">>UNUSED_445<<",
272
+ ">>UNUSED_446<<",
273
+ ">>UNUSED_447<<",
274
+ ">>UNUSED_448<<",
275
+ ">>UNUSED_449<<",
276
+ ">>UNUSED_450<<",
277
+ ">>UNUSED_451<<",
278
+ ">>UNUSED_452<<",
279
+ ">>UNUSED_453<<",
280
+ ">>UNUSED_454<<",
281
+ ">>UNUSED_455<<",
282
+ ">>UNUSED_456<<",
283
+ ">>UNUSED_457<<",
284
+ ">>UNUSED_458<<",
285
+ ">>UNUSED_459<<",
286
+ ">>UNUSED_460<<",
287
+ ">>UNUSED_461<<",
288
+ ">>UNUSED_462<<",
289
+ ">>UNUSED_463<<",
290
+ ">>UNUSED_464<<",
291
+ ">>UNUSED_465<<",
292
+ ">>UNUSED_466<<",
293
+ ">>UNUSED_467<<",
294
+ ">>UNUSED_468<<",
295
+ ">>UNUSED_469<<",
296
+ ">>UNUSED_470<<",
297
+ ">>UNUSED_471<<",
298
+ ">>UNUSED_472<<",
299
+ ">>UNUSED_473<<",
300
+ ">>UNUSED_474<<",
301
+ ">>UNUSED_475<<",
302
+ ">>UNUSED_476<<",
303
+ ">>UNUSED_477<<",
304
+ ">>UNUSED_478<<",
305
+ ">>UNUSED_479<<",
306
+ ">>UNUSED_480<<",
307
+ ">>UNUSED_481<<",
308
+ ">>UNUSED_482<<",
309
+ ">>UNUSED_483<<",
310
+ ">>UNUSED_484<<",
311
+ ">>UNUSED_485<<",
312
+ ">>UNUSED_486<<",
313
+ ">>UNUSED_487<<",
314
+ ">>UNUSED_488<<",
315
+ ">>UNUSED_489<<",
316
+ ">>UNUSED_490<<",
317
+ ">>UNUSED_491<<",
318
+ ">>UNUSED_492<<",
319
+ ">>UNUSED_493<<",
320
+ ">>UNUSED_494<<",
321
+ ">>UNUSED_495<<",
322
+ ">>UNUSED_496<<",
323
+ ">>UNUSED_497<<",
324
+ ">>UNUSED_498<<",
325
+ ">>UNUSED_499<<",
326
+ ">>UNUSED_500<<",
327
+ ">>UNUSED_501<<",
328
+ ">>UNUSED_502<<",
329
+ ">>UNUSED_503<<",
330
+ ">>UNUSED_504<<",
331
+ ">>UNUSED_505<<",
332
+ ">>UNUSED_506<<",
333
+ ">>UNUSED_507<<",
334
+ ">>UNUSED_508<<",
335
+ ">>UNUSED_509<<",
336
+ ">>UNUSED_510<<",
337
+ ">>UNUSED_511<<"
338
+ ],
339
+ "is_local": false,
340
+ "model_max_length": 1000000000000000019884624838656,
341
+ "model_specific_special_tokens": {},
342
+ "pad_token": "<|end_of_text|>",
343
+ "tokenizer_class": "TokenizersBackend"
344
+ }