{ "model_max_length": 512, "name_or_path": "/global/scratch/users/aniketh/PromoGen/HepG2_SentencePieceUnigramTokenizer_4096_log_bins_fast", "special_tokens": [ "", "", "", "", "", "", "" ], "special_tokens_map_file": "/global/scratch/users/aniketh/PromoGen/HepG2_SentencePieceUnigramTokenizer_4096_log_bins_fast/special_tokens_map.json", "tokenizer_class": "PreTrainedTokenizerFast" }