Tokenization updates (#8)
Browse files- Tokenization updates (49fb434703c73a1a8b8e30f7d3aac46daa0b6016)
- Update tokenizer_config.json (f5839f9de43de6cc57bf3d80e46f0cd82d5c694b)
- Update tokenization_xgen.py (35686226b2c0fc4f9a5b19f9ca36686e5ef187f3)
- Update tokenization_xgen.py (db5edb9790bf97938c381f2208b149c2ff2ee01f)
- Update tokenization_xgen.py (5017b1efccce33443a00f8f9b7a34c1db53116ce)
Co-authored-by: Singer <[email protected]>
- tokenization_xgen.py +12 -6
- tokenizer_config.json +1 -0
tokenization_xgen.py
CHANGED
@@ -149,20 +149,22 @@ class XgenTokenizer(PreTrainedTokenizer):
|
|
149 |
def _convert_token_to_id(self, token):
|
150 |
"""Converts a token (str) in an id using the vocab."""
|
151 |
if isinstance(token, str):
|
152 |
-
|
153 |
-
|
154 |
-
|
155 |
|
156 |
def _convert_id_to_token(self, index):
|
157 |
"""Converts an index (integer) in a token (str) using the vocab."""
|
158 |
return self.encoder.decode_single_token_bytes(index)
|
159 |
|
160 |
def _decode(self, token_ids: List[int], skip_special_tokens: bool = False, **kwargs):
|
|
|
|
|
161 |
return self.encoder.decode(token_ids)
|
162 |
|
163 |
def build_inputs_with_special_tokens(self, token_ids_0, token_ids_1=None) -> List[int]:
|
164 |
"""Build model inputs from a sequence by appending eos_token_id."""
|
165 |
-
eos_token_id = [
|
166 |
|
167 |
output = token_ids_0 + eos_token_id
|
168 |
|
@@ -218,11 +220,15 @@ class XgenTokenizer(PreTrainedTokenizer):
|
|
218 |
Returns:
|
219 |
`List[int]`: List of [token type IDs](../glossary#token-type-ids) according to the given sequence(s).
|
220 |
"""
|
221 |
-
eos_token_id = [
|
222 |
|
223 |
output = [0] * len(token_ids_0 + eos_token_id)
|
224 |
|
225 |
if token_ids_1 is not None:
|
226 |
output += [1] * len(token_ids_1 + eos_token_id)
|
227 |
|
228 |
-
return output
|
|
|
|
|
|
|
|
|
|
149 |
def _convert_token_to_id(self, token):
|
150 |
"""Converts a token (str) in an id using the vocab."""
|
151 |
if isinstance(token, str):
|
152 |
+
return self.encoder.encode_single_token(token)
|
153 |
+
else:
|
154 |
+
return token
|
155 |
|
156 |
def _convert_id_to_token(self, index):
|
157 |
"""Converts an index (integer) in a token (str) using the vocab."""
|
158 |
return self.encoder.decode_single_token_bytes(index)
|
159 |
|
160 |
def _decode(self, token_ids: List[int], skip_special_tokens: bool = False, **kwargs):
|
161 |
+
if skip_special_tokens:
|
162 |
+
token_ids = [t for t in token_ids if t not in self.all_special_ids]
|
163 |
return self.encoder.decode(token_ids)
|
164 |
|
165 |
def build_inputs_with_special_tokens(self, token_ids_0, token_ids_1=None) -> List[int]:
|
166 |
"""Build model inputs from a sequence by appending eos_token_id."""
|
167 |
+
eos_token_id = [self.eos_token_id] if self.add_eos_token else []
|
168 |
|
169 |
output = token_ids_0 + eos_token_id
|
170 |
|
|
|
220 |
Returns:
|
221 |
`List[int]`: List of [token type IDs](../glossary#token-type-ids) according to the given sequence(s).
|
222 |
"""
|
223 |
+
eos_token_id = [self.eos_token_id] if self.add_eos_token else []
|
224 |
|
225 |
output = [0] * len(token_ids_0 + eos_token_id)
|
226 |
|
227 |
if token_ids_1 is not None:
|
228 |
output += [1] * len(token_ids_1 + eos_token_id)
|
229 |
|
230 |
+
return output
|
231 |
+
|
232 |
+
# has no vocab file
|
233 |
+
def save_vocabulary(self, save_directory: str, filename_prefix: Optional[str] = None):
|
234 |
+
return ()
|
tokenizer_config.json
CHANGED
@@ -2,6 +2,7 @@
|
|
2 |
"add_eos_token": false,
|
3 |
"add_special_tokens": true,
|
4 |
"clean_up_tokenization_spaces": true,
|
|
|
5 |
"model_max_length": 1000000000000000019884624838656,
|
6 |
"pad_token": null,
|
7 |
"tokenizer_class": "XgenTokenizer",
|
|
|
2 |
"add_eos_token": false,
|
3 |
"add_special_tokens": true,
|
4 |
"clean_up_tokenization_spaces": true,
|
5 |
+
"eos_token": "<|endoftext|>",
|
6 |
"model_max_length": 1000000000000000019884624838656,
|
7 |
"pad_token": null,
|
8 |
"tokenizer_class": "XgenTokenizer",
|