from transformers import PreTrainedTokenizerFast | |
tokenizer = PreTrainedTokenizerFast( | |
tokenizer_file="tokenizer.json", | |
unk_token="<unk>", | |
pad_token="<pad>", | |
cls_token="<cls>", | |
sep_token="<sep>", | |
mask_token="<mask>" | |
) |
from transformers import PreTrainedTokenizerFast | |
tokenizer = PreTrainedTokenizerFast( | |
tokenizer_file="tokenizer.json", | |
unk_token="<unk>", | |
pad_token="<pad>", | |
cls_token="<cls>", | |
sep_token="<sep>", | |
mask_token="<mask>" | |
) |