ekrombouts
commited on
Finetuned BramVanroy/fietje-2-instruct on ekrombouts/Gardenia_instruct_dataset
Browse files- tokenizer_config.json +0 -7
tokenizer_config.json
CHANGED
@@ -338,16 +338,9 @@
|
|
338 |
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
339 |
"clean_up_tokenization_spaces": true,
|
340 |
"eos_token": "<|im_end|>",
|
341 |
-
"max_length": 2048,
|
342 |
"model_max_length": 2048,
|
343 |
-
"pad_to_multiple_of": null,
|
344 |
"pad_token": "<|im_end|>",
|
345 |
-
"pad_token_type_id": 0,
|
346 |
-
"padding_side": "right",
|
347 |
"return_token_type_ids": false,
|
348 |
-
"stride": 0,
|
349 |
"tokenizer_class": "CodeGenTokenizer",
|
350 |
-
"truncation_side": "right",
|
351 |
-
"truncation_strategy": "longest_first",
|
352 |
"unk_token": "<|endoftext|>"
|
353 |
}
|
|
|
338 |
"chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
|
339 |
"clean_up_tokenization_spaces": true,
|
340 |
"eos_token": "<|im_end|>",
|
|
|
341 |
"model_max_length": 2048,
|
|
|
342 |
"pad_token": "<|im_end|>",
|
|
|
|
|
343 |
"return_token_type_ids": false,
|
|
|
344 |
"tokenizer_class": "CodeGenTokenizer",
|
|
|
|
|
345 |
"unk_token": "<|endoftext|>"
|
346 |
}
|