Update tokenizer_config.json
Browse files- tokenizer_config.json +1 -2
tokenizer_config.json
CHANGED
@@ -140,8 +140,7 @@
|
|
140 |
}
|
141 |
},
|
142 |
"bos_token": "<|begin▁of▁sentence|>",
|
143 |
-
"chat_template": "{% for message in messages %}\n{% if message['
|
144 |
-
"model_max_length": 4096,
|
145 |
"pad_token": "!",
|
146 |
"sp_model_kwargs": {},
|
147 |
"tokenizer_class": "LlamaTokenizer",
|
|
|
140 |
}
|
141 |
},
|
142 |
"bos_token": "<|begin▁of▁sentence|>",
|
143 |
+
"chat_template": "{% for message in messages %}\n{% if message['from'] == 'human' %}\n{{ '<|user|>\n' + message['value'] + eos_token }}\n{% elif message['from'] == 'system' %}\n{{ '<|system|>\n' + message['value'] + eos_token }}\n{% elif message['from'] == 'gpt' %}\n{{ '<|assistant|>\n' + message['value'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "clean_up_tokenization_spaces": false, "model_max_length": 4096,
|
|
|
144 |
"pad_token": "!",
|
145 |
"sp_model_kwargs": {},
|
146 |
"tokenizer_class": "LlamaTokenizer",
|