chatglm3-alpaca-finetune-base-model / tokenizer_config.json
seamoon2333's picture
Upload 7 files
7b8b68f
raw
history blame contribute delete
474 Bytes
{
"added_tokens_decoder": {},
"additional_special_tokens": [],
"auto_map": {
"AutoTokenizer": [
"THUDM/chatglm3-6b-base--tokenization_chatglm.ChatGLMTokenizer",
null
]
},
"clean_up_tokenization_spaces": false,
"do_lower_case": false,
"encode_special_tokens": false,
"model_max_length": 1000000000000000019884624838656,
"padding_side": "left",
"remove_space": false,
"tokenizer_class": "ChatGLMTokenizer",
"tokenizer_file": null
}