File size: 338 Bytes
62f8266 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
{
"bos_token": "<|begin_of_text|>",
"clean_up_tokenization_spaces": true,
"eos_token": "<|end_of_text|>",
"model_input_names": [
"input_ids",
"attention_mask"
],
"model_max_length": 131072,
"vocab_file": null,
"name_or_path": "meta-llama/Llama-3.2-1B",
"padding_side": "right",
"pad_token": "<|end_of_text|>"
} |