ERNIE-Layout-Pytorch / tokenizer_config.json
Norm's picture
Update tokenizer_config.json
3a26195
raw
history blame
320 Bytes
{
"do_tokenize_postprocess": false,
"sep_token": "[SEP]",
"cls_token": "[CLS]",
"unk_token": "[UNK]",
"pad_token": "[PAD]",
"mask_token": "[MASK]",
"do_lower_case": true,
"model_max_length": 512,
"vocab_file": "my/path/to/vocab.txt",
"sentencepiece_model_file": "my/path/to/sentencepiece.bpe.model"
}