{ "add_bos_token": false, "add_eos_token": false, "bos_token": "<|startoftext|>", "clean_up_tokenization_spaces": false, "cls_token": "[CLS]", "eos_token": "<|endoftext|>", "gmask_token": "[gMASK]", "merges_file": null, "model_max_length": 1000000000000000019884624838656, "tokenizer_class": "RodimusTokenizer", "vocab_file": null, "auto_map": { "AutoTokenizer": [ null, "tokenization_rodimus_fast.RodimusTokenizer" ] } }