kobert-base-v1 / tokenizer_config.json
digit82's picture
update special tokens
a9f5849
raw
history blame contribute delete
432 Bytes
{"do_lower_case": false, "remove_space": true, "keep_accents": false, "bos_token": "[CLS]", "eos_token": "[SEP]", "unk_token": "[UNK]", "sep_token": "[SEP]", "pad_token": "[PAD]", "cls_token": "[CLS]", "mask_token": {"content": "[MASK]", "single_word": false, "lstrip": true, "rstrip": false, "normalized": true, "__type": "AddedToken"}, "additional_special_tokens": null, "sp_model_kwargs": {}, "tokenizer_class": "XLNetTokenizer"}