cryptgpt-large / tokenizer_utils.py
diwank's picture
Upload tokenizer
0280a5a verified
raw
history blame contribute delete
264 Bytes
from transformers import PreTrainedTokenizerFast
class CryptGPTTokenizer(PreTrainedTokenizerFast):
@staticmethod
def clean_up_tokenization(out_string):
return out_string.replace(' ', "")
CryptGPTTokenizer.register_for_auto_class("AutoTokenizer")