Datasets:

Modalities:
Text
Formats:
json
Size:
< 1K
Libraries:
Datasets
pandas
License:
tokenization_test_data / tokenization_data_cases.json
itazap's picture
itazap HF Staff
Upload tokenization_data_cases.json
0157101 verified
[
{
"TestClass": "AlbertTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁i",
"▁was",
"▁born",
"▁in",
"▁9",
"2000",
",",
"▁and",
"▁this",
"▁is",
"▁false",
"."
],
"encoded": [
31,
23,
386,
19,
561,
3050,
15,
17,
48,
25,
4997,
9
],
"encoded_special": [
2,
31,
23,
386,
19,
561,
3050,
15,
17,
48,
25,
4997,
9,
3
],
"params": {},
"params_encode": {}
},
{
"TestClass": "BertTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "BertTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {
"do_lower_case": true
},
"params_encode": {}
},
{
"TestClass": "BigBirdPegasusTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁",
"t",
"est"
],
"encoded": [
288,
46,
9,
3,
12,
390
],
"encoded_special": [
288,
46,
9,
3,
12,
390,
1
],
"params": {},
"params_encode": {}
},
{
"TestClass": "BigBirdTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4
],
"encoded_special": [
1002,
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4,
1000
],
"params": {},
"params_encode": {}
},
{
"TestClass": "CLIPTokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"lo",
"w",
"er</w>",
"n",
"e",
"w",
"er</w>"
],
"encoded": [
10,
2,
16,
9,
3,
2,
16
],
"encoded_special": [
21,
10,
2,
16,
9,
3,
2,
16,
22
],
"params": {},
"params_encode": {}
},
{
"TestClass": "CamembertTokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
12,
25,
88,
59,
28,
23,
11,
3,
606,
351,
351,
351,
7,
16,
70,
50,
76,
84,
10,
3,
8
],
"encoded_special": [
5,
12,
25,
88,
59,
28,
23,
11,
3,
606,
351,
351,
351,
7,
16,
70,
50,
76,
84,
10,
3,
8,
6
],
"params": {},
"params_encode": {}
},
{
"TestClass": "CodeGenTokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"Ġlow",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
14,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
14,
15,
10,
9,
3,
2,
15
],
"params": {
"add_prefix_space": true
},
"params_encode": {}
},
{
"TestClass": "DPRContextEncoderTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "DPRQuestionEncoderTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "DPRReaderTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "DebertaTokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"l",
"o",
"w",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
0,
1,
2,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
0,
1,
2,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {}
},
{
"TestClass": "DebertaV2TokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé!",
"tokens": [
"▁",
"I",
"▁was",
"▁born",
"▁in",
"▁9",
"2000",
",",
"▁and",
"▁this",
"▁is",
"▁fal",
"s",
"é",
"!"
],
"encoded": [
13,
1,
23,
386,
19,
561,
3050,
15,
17,
48,
25,
8256,
18,
1,
187
],
"encoded_special": [
2,
13,
1,
23,
386,
19,
561,
3050,
15,
17,
48,
25,
8256,
18,
1,
187,
3
],
"params": {},
"params_encode": {}
},
{
"TestClass": "DistilBertTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "ElectraTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "ElectraTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {
"do_lower_case": true
},
"params_encode": {}
},
{
"TestClass": "FNetTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁",
"I",
"▁was",
"▁born",
"▁in",
"▁9",
"2000",
",",
"▁and",
"▁this",
"▁is",
"▁fal",
"s",
"é",
"."
],
"encoded": [
13,
1,
23,
386,
19,
561,
3050,
15,
17,
48,
25,
8256,
18,
1,
9
],
"encoded_special": [
2,
13,
1,
23,
386,
19,
561,
3050,
15,
17,
48,
25,
8256,
18,
1,
9,
3
],
"params": {},
"params_encode": {}
},
{
"TestClass": "FunnelTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
7,
4,
5,
10,
8,
9
],
"encoded_special": [
1,
7,
4,
5,
10,
8,
9,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "GPT2TokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"Ġlow",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
14,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
14,
15,
10,
9,
3,
2,
15
],
"params": {
"add_prefix_space": true
},
"params_encode": {}
},
{
"TestClass": "HerbertTokenizationTest",
"sp_base": false,
"sequence": "lower,newer",
"tokens": [
"low",
"er</w>",
",</w>",
"n",
"e",
"w",
"er</w>"
],
"encoded": [
16,
17,
22,
11,
5,
4,
17
],
"encoded_special": [
0,
16,
17,
22,
11,
5,
4,
17,
1
],
"params": {},
"params_encode": {}
},
{
"TestClass": "LayoutLMTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
7,
4,
5,
10,
8,
9
],
"encoded_special": [
1,
7,
4,
5,
10,
8,
9,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "LayoutLMv2TokenizationTest",
"sp_base": false,
"sequence": null,
"tokens": null,
"encoded": [
10,
11,
12,
13
],
"encoded_special": [
1,
10,
11,
12,
13,
2
],
"params": {},
"params_encode": {
"text": [
"a",
"weirdly",
"test"
],
"boxes": [
[
423,
237,
440,
251
],
[
427,
272,
441,
287
],
[
419,
115,
437,
129
]
]
}
},
{
"TestClass": "LayoutLMv3TokenizationTest",
"sp_base": false,
"sequence": null,
"tokens": null,
"encoded": [
14,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
14,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {
"text": [
"lower",
"newer"
],
"boxes": [
[
423,
237,
440,
251
],
[
427,
272,
441,
287
]
]
}
},
{
"TestClass": "LayoutXLMTokenizationTest",
"sp_base": true,
"sequence": null,
"tokens": null,
"encoded": [
11,
113,
159,
17,
39,
171,
383
],
"encoded_special": [
0,
11,
113,
159,
17,
39,
171,
383,
2
],
"params": {},
"params_encode": {
"text": [
"a",
"weirdly",
"test"
],
"boxes": [
[
423,
237,
440,
251
],
[
427,
272,
441,
287
],
[
419,
115,
437,
129
]
]
}
},
{
"TestClass": "LongformerTokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"l",
"o",
"w",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
0,
1,
2,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
0,
1,
2,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {}
},
{
"TestClass": "LxmertTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"[UNK]",
"[UNK]",
"[UNK]",
"[UNK]",
"[UNK]",
",",
"[UNK]",
"[UNK]",
"[UNK]",
"[UNK]",
"[UNK]"
],
"encoded": [
0,
0,
0,
0,
0,
10,
0,
0,
0,
0,
0
],
"encoded_special": [
1,
0,
0,
0,
0,
0,
10,
0,
0,
0,
0,
0,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MBart50TokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49
],
"encoded_special": [
1004,
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MBartTokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49
],
"encoded_special": [
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49,
2,
1004
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MPNetTokenizerTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
15,
9,
6,
7,
12,
10,
11,
16
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MarkupLMTokenizationTest",
"sp_base": false,
"sequence": null,
"tokens": null,
"encoded": [
21,
3,
0,
0,
1,
2,
1,
4,
0,
8
],
"encoded_special": [
22,
21,
3,
0,
0,
1,
2,
1,
4,
0,
8,
23
],
"params": {},
"params_encode": {
"text": [
"hello",
"world"
],
"xpaths": [
",/html/body/div/li[1]/div/span",
",/html/body/div/li[1]/div/span"
]
}
},
{
"TestClass": "MobileBERTTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "NllbTokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49
],
"encoded_special": [
1048,
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "OpenAIGPTTokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"low",
"er</w>",
"n",
"e",
"w",
"er</w>"
],
"encoded": [
14,
15,
9,
3,
2,
15
],
"encoded_special": [
14,
15,
9,
3,
2,
15
],
"params": {},
"params_encode": {}
},
{
"TestClass": "PegasusTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁",
"t",
"est"
],
"encoded": [
391,
149,
112,
106,
115,
493
],
"encoded_special": [
391,
149,
112,
106,
115,
493,
1
],
"params": {},
"params_encode": {}
},
{
"TestClass": "Qwen2TokenizationTest",
"sp_base": false,
"sequence": "lower lower newer 010;}\r\n<|endoftext|>ϓ",
"tokens": [
"l",
"o",
"w",
"er",
"Ġlow",
"er",
"Ġ",
"n",
"e",
"w",
"er",
"Ġ",
"0",
"1",
"0",
";}",
"č",
"Ċ",
"<|endoftext|>",
"Ïĵ"
],
"encoded": [
75,
78,
86,
260,
259,
260,
220,
77,
68,
86,
260,
220,
15,
16,
15,
265,
201,
198,
270,
267
],
"encoded_special": [
75,
78,
86,
260,
259,
260,
220,
77,
68,
86,
260,
220,
15,
16,
15,
265,
201,
198,
270,
267
],
"params": {},
"params_encode": {}
},
{
"TestClass": "ReformerTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4
],
"encoded_special": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4
],
"params": {},
"params_encode": {}
},
{
"TestClass": "RemBertTokenizationTest",
"sp_base": true,
"sequence": "this is a test",
"tokens": [
"▁this",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
66,
46,
10,
170,
382
],
"encoded_special": [
1000,
66,
46,
10,
170,
382,
1001
],
"params": {},
"params_encode": {}
},
{
"TestClass": "RobertaTokenizationTest",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"l",
"o",
"w",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
0,
1,
2,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
0,
1,
2,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {}
},
{
"TestClass": "SeamlessM4TTokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49
],
"encoded_special": [
3,
1,
6,
9,
10,
11,
13,
14,
20,
22,
26,
32,
35,
37,
38,
41,
42,
44,
45,
47,
48,
49,
3
],
"params": {},
"params_encode": {}
},
{
"TestClass": "SplinterTokenizationTest",
"sp_base": false,
"sequence": "I need to test this rigor",
"tokens": [
"[UNK]",
"[UNK]",
"[UNK]",
"test",
"this",
"rigor"
],
"encoded": [
3,
10,
10,
10,
16,
13,
21,
1
],
"encoded_special": [
3,
10,
10,
10,
16,
13,
21,
1
],
"params": {},
"params_encode": {}
},
{
"TestClass": "SqueezeBertTokenizationTest",
"sp_base": false,
"sequence": "UNwantéd,running",
"tokens": [
"un",
"##want",
"##ed",
",",
"runn",
"##ing"
],
"encoded": [
9,
6,
7,
12,
10,
11
],
"encoded_special": [
1,
9,
6,
7,
12,
10,
11,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "T5TokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4
],
"encoded_special": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "TestTokenizationBart",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"l",
"o",
"w",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
0,
1,
2,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
0,
1,
2,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {}
},
{
"TestClass": "TestTokenizationLED",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"l",
"o",
"w",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
0,
1,
2,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
0,
1,
2,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {}
},
{
"TestClass": "TestTokenizationMvp",
"sp_base": false,
"sequence": "lower newer",
"tokens": [
"l",
"o",
"w",
"er",
"Ġ",
"n",
"e",
"w",
"er"
],
"encoded": [
0,
1,
2,
15,
10,
9,
3,
2,
15
],
"encoded_special": [
20,
0,
1,
2,
15,
10,
9,
3,
2,
15,
21
],
"params": {},
"params_encode": {}
},
{
"TestClass": "UdopTokenizationTest",
"sp_base": true,
"sequence": null,
"tokens": null,
"encoded": [
10,
112,
158,
16,
38,
170,
382,
37,
86,
20
],
"encoded_special": [
10,
112,
158,
16,
38,
170,
382,
37,
86,
20,
2
],
"params": {},
"params_encode": {
"text": [
"a",
"weirdly",
"test",
"hello"
],
"boxes": [
[
423,
237,
440,
251
],
[
427,
272,
441,
287
],
[
419,
115,
437,
129
],
[
961,
885,
992,
912
]
]
}
},
{
"TestClass": "WhisperTokenizerTest",
"sp_base": false,
"sequence": "A BCDEFGHIJKLMNOPQRST",
"tokens": [
"A",
"ĠBC",
"DE",
"F",
"GH",
"I",
"J",
"K",
"L",
"M",
"N",
"OP",
"Q",
"R",
"ST"
],
"encoded": [
32,
14359,
22296,
37,
4269,
40,
41,
42,
43,
44,
45,
12059,
48,
49,
6840
],
"encoded_special": [
50258,
50363,
32,
14359,
22296,
37,
4269,
40,
41,
42,
43,
44,
45,
12059,
48,
49,
6840,
50257
],
"params": {},
"params_encode": {}
},
{
"TestClass": "XGLMTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5
],
"encoded_special": [
2,
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5
],
"params": {},
"params_encode": {}
},
{
"TestClass": "XLMRobertaTokenizationTest",
"sp_base": false,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5
],
"encoded_special": [
0,
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "XLNetTokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
5,
8,
9,
10,
12,
13,
19,
21,
25,
31,
34,
36,
37,
40,
41,
43,
44,
46,
47,
48
],
"encoded_special": [
5,
8,
9,
10,
12,
13,
19,
21,
25,
31,
34,
36,
37,
40,
41,
43,
44,
46,
47,
48,
1000,
1002
],
"params": {},
"params_encode": {}
},
{
"TestClass": "CamembertTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
289,
50,
14,
174,
386
],
"encoded_special": [
5,
289,
50,
14,
174,
386,
6
],
"params": {},
"params_encode": {}
},
{
"TestClass": "CamembertTokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
9,
12,
13,
14,
16,
17,
23,
25,
29,
35,
38,
40,
41,
44,
45,
47,
48,
50,
51,
52
],
"encoded_special": [
5,
9,
12,
13,
14,
16,
17,
23,
25,
29,
35,
38,
40,
41,
44,
45,
47,
48,
50,
51,
52,
6
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MBartTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
286,
47,
11,
171,
383
],
"encoded_special": [
286,
47,
11,
171,
383,
2,
1004
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MBartTokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5
],
"encoded_special": [
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5,
2,
1004
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MBart50TokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
286,
47,
11,
171,
383
],
"encoded_special": [
1004,
286,
47,
11,
171,
383,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "MBart50TokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5
],
"encoded_special": [
1004,
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "NllbTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
286,
47,
11,
171,
383
],
"encoded_special": [
1048,
286,
47,
11,
171,
383,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "NllbTokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5
],
"encoded_special": [
1048,
9,
22,
85,
56,
25,
20,
8,
3,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
3,
5,
2
],
"params": {},
"params_encode": {}
},
{
"TestClass": "RemBertTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
285,
46,
10,
170,
382
],
"encoded_special": [
1000,
285,
46,
10,
170,
382,
1001
],
"params": {},
"params_encode": {}
},
{
"TestClass": "RemBertTokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4
],
"encoded_special": [
1000,
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4,
1001
],
"params": {},
"params_encode": {}
},
{
"TestClass": "RemBertTokenizationTest",
"sp_base": true,
"sequence": "the I to a and of in was it me that be he for with my not is s you",
"tokens": [
"▁the",
"▁I",
"▁to",
"▁a",
"▁and",
"▁of",
"▁in",
"▁was",
"▁it",
"▁me",
"▁that",
"▁be",
"▁he",
"▁for",
"▁with",
"▁my",
"▁not",
"▁is",
"▁s",
"▁you"
],
"encoded": [
5,
8,
9,
10,
12,
13,
19,
21,
25,
31,
34,
36,
37,
40,
41,
43,
44,
46,
47,
48
],
"encoded_special": [
1000,
5,
8,
9,
10,
12,
13,
19,
21,
25,
31,
34,
36,
37,
40,
41,
43,
44,
46,
47,
48,
1001
],
"params": {},
"params_encode": {}
},
{
"TestClass": "SeamlessM4TTokenizationTest",
"sp_base": true,
"sequence": "This is a test",
"tokens": [
"▁This",
"▁is",
"▁a",
"▁t",
"est"
],
"encoded": [
286,
47,
11,
171,
383
],
"encoded_special": [
3,
1,
286,
47,
11,
171,
383,
3
],
"params": {},
"params_encode": {}
},
{
"TestClass": "SeamlessM4TTokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
9,
22,
85,
56,
25,
20,
8,
1,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
1,
5
],
"encoded_special": [
3,
1,
9,
22,
85,
56,
25,
20,
8,
1,
603,
348,
348,
348,
4,
13,
67,
47,
73,
81,
7,
1,
5,
3
],
"params": {},
"params_encode": {}
},
{
"TestClass": "XLNetTokenizationTest",
"sp_base": true,
"sequence": "I was born in 92000, and this is falsé.",
"tokens": [
"▁I",
"▁was",
"▁b",
"or",
"n",
"▁in",
"▁",
"9",
"2",
"0",
"0",
"0",
",",
"▁and",
"▁this",
"▁is",
"▁f",
"al",
"s",
"é",
"."
],
"encoded": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4
],
"encoded_special": [
8,
21,
84,
55,
24,
19,
7,
0,
602,
347,
347,
347,
3,
12,
66,
46,
72,
80,
6,
0,
4,
1000,
1002
],
"params": {},
"params_encode": {}
}
]