flash-beta1 / tokenizer_config.json
dphnAI's picture
Add files using upload-large-folder tool
4ace562 verified
raw
history blame contribute delete
382 Bytes
{
"backend": "tokenizers",
"clean_up_tokenization_spaces": false,
"do_lower_case": false,
"eos_token": "<|endoftext|>",
"is_local": true,
"max_length": null,
"model_max_length": 128000,
"pad_to_multiple_of": null,
"pad_token": "<|endoftext|>",
"pad_token_type_id": 0,
"padding_side": "left",
"remove_space": false,
"tokenizer_class": "TokenizersBackend"
}