ERNIE-Layout-Pytorch / tokenizer_config.json
Norm's picture
Update tokenizer_config.json
df33674
raw
history blame
302 Bytes
{
"do_tokenize_postprocess": false,
"sep_token": "[SEP]",
"cls_token": "[CLS]",
"unk_token": "[UNK]",
"pad_token": "[PAD]",
"mask_token": "[MASK]",
"do_lower_case": true,
"model_max_length": 512,
"vocab_file": "./vocab.txt",
"sentencepiece_model_file": "./sentencepiece.bpe.model"
}