Update tokenizer_script.py
Browse files- tokenizer_script.py +3 -3
tokenizer_script.py
CHANGED
@@ -82,13 +82,13 @@ class CharacterTokenizer(PreTrainedTokenizer):
|
|
82 |
if max_length is not None:
|
83 |
encoded_texts = [ids[:max_length] for ids in encoded_texts]
|
84 |
if add_special_tokens:
|
85 |
-
bos_token_id = self.convert_tokens_to_ids(
|
86 |
-
eos_token_id = self.convert_tokens_to_ids(
|
87 |
encoded_texts = [[bos_token_id] + ids + [eos_token_id] for ids in encoded_texts]
|
88 |
# Handle padding
|
89 |
if padding:
|
90 |
# properly handle padding side
|
91 |
-
pad_id = self.vocab.get(
|
92 |
max_len = max(len(ids) for ids in encoded_texts) if max_length is None else max_length
|
93 |
if tokenizer.padding_side == "right":
|
94 |
encoded_texts = [ids + [pad_id] * (max_len - len(ids)) for ids in encoded_texts]
|
|
|
82 |
if max_length is not None:
|
83 |
encoded_texts = [ids[:max_length] for ids in encoded_texts]
|
84 |
if add_special_tokens:
|
85 |
+
bos_token_id = self.convert_tokens_to_ids(self.bos_token)
|
86 |
+
eos_token_id = self.convert_tokens_to_ids(self.eos_token)
|
87 |
encoded_texts = [[bos_token_id] + ids + [eos_token_id] for ids in encoded_texts]
|
88 |
# Handle padding
|
89 |
if padding:
|
90 |
# properly handle padding side
|
91 |
+
pad_id = self.vocab.get(self.pad_token, 0)
|
92 |
max_len = max(len(ids) for ids in encoded_texts) if max_length is None else max_length
|
93 |
if tokenizer.padding_side == "right":
|
94 |
encoded_texts = [ids + [pad_id] * (max_len - len(ids)) for ids in encoded_texts]
|