Text Generation
Transformers
Safetensors
llama
go
text-generation-inference
Inference Endpoints
goformer-v0.1 / tokenizer_config.json
kenhktsui's picture
Update tokenizer_config.json
af4e0cb verified
raw
history blame contribute delete
265 Bytes
{
"added_tokens_decoder": {},
"auto_map": {
"AutoTokenizer": [
"tokenizer.AlphabetTokenizer",
null
]
},
"clean_up_tokenization_spaces": true,
"model_max_length": 2048,
"pad_token": "[PAD]",
"tokenizer_class": "AlphabetTokenizer"
}