modern_french_normalisation / tokenizer_config.json
Rachel Bawden
redone tokeniser
d9498bd
raw
history blame
131 Bytes
{"unk_token": "<unk>", "eos_token": "</s>", "bos_token": "<s>", "pad_token": "<pad>", "tokenizer_class": "PreTrainedTokenizerFast"}