{ "backend": "tokenizers", "clean_up_tokenization_spaces": false, "do_lower_case": false, "eos_token": "<|endoftext|>", "is_local": true, "max_length": null, "model_max_length": 128000, "pad_to_multiple_of": null, "pad_token": "<|endoftext|>", "pad_token_type_id": 0, "padding_side": "left", "remove_space": false, "tokenizer_class": "TokenizersBackend" }