adibm commited on
Commit
a50cd23
·
1 Parent(s): e8baad1

Upload tokenizer

Browse files
.gitattributes CHANGED
@@ -33,3 +33,4 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
 
 
33
  *.zip filter=lfs diff=lfs merge=lfs -text
34
  *.zst filter=lfs diff=lfs merge=lfs -text
35
  *tfevents* filter=lfs diff=lfs merge=lfs -text
36
+ tokenizer.json filter=lfs diff=lfs merge=lfs -text
special_tokens_map.json ADDED
@@ -0,0 +1,61 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "additional_special_tokens": [
3
+ "<|begintarget|>",
4
+ "<|endtarget|>",
5
+ "<|begincontext|>",
6
+ "<|endcontext|>",
7
+ "<|system|>",
8
+ "<|user|>",
9
+ "<|beginlastuserutterance|>",
10
+ "<|endlastuserutterance|>",
11
+ "<|begindsts|>",
12
+ "<|enddsts|>",
13
+ "<|begindst|>",
14
+ "<|enddst|>",
15
+ "<|beginbelief|>",
16
+ "<|endbelief|>",
17
+ "<|beginresponse|>",
18
+ "<|endresponse|>",
19
+ "<|beginaction|>",
20
+ "<|endaction|>",
21
+ "<|beginuseraction|>",
22
+ "<|enduseraction|>",
23
+ "<|sysactions|>",
24
+ "<|beginintent|>",
25
+ "<|endintent|>",
26
+ "<|beginrequestedslots|>",
27
+ "<|endrequestedslots|>",
28
+ "<|promptdst|>",
29
+ "<|promptaction|>",
30
+ "<|promptresponse|>",
31
+ "<|pad|>",
32
+ "<|endoftext|>",
33
+ "<|startoftext|>",
34
+ "<|beginschema|>",
35
+ "<|endschema|>",
36
+ "<|schemaname|>",
37
+ "<|schemadescription|>",
38
+ "<|beginschemaintent|>",
39
+ "<|endschemaintent|>",
40
+ "<|intentrequiredslots|>",
41
+ "<|intentresultslots|>",
42
+ "<|intentoptionalslots|>",
43
+ "<|possiblevalues|>",
44
+ "<|endpossiblevalues|>",
45
+ "<|beginschemaslot|>",
46
+ "<|endschemaslot|>",
47
+ "<|schemaslotvalues|>",
48
+ "<|beginserviceresults|>",
49
+ "<|endserviceresults|>"
50
+ ],
51
+ "bos_token": "<|startoftext|>",
52
+ "eos_token": "<|endtarget|>",
53
+ "pad_token": "<|pad|>",
54
+ "unk_token": {
55
+ "content": "<unk>",
56
+ "lstrip": false,
57
+ "normalized": true,
58
+ "rstrip": false,
59
+ "single_word": false
60
+ }
61
+ }
tokenizer.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6d28aeee1343a1deccb4b21cf8d1fe43556cf2da412c5416a357c3b1fd0ce753
3
+ size 25189832
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
3
+ size 499723
tokenizer_config.json ADDED
@@ -0,0 +1,67 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_prefix_space": false,
3
+ "additional_special_tokens": [
4
+ "<|begintarget|>",
5
+ "<|endtarget|>",
6
+ "<|begincontext|>",
7
+ "<|endcontext|>",
8
+ "<|system|>",
9
+ "<|user|>",
10
+ "<|beginlastuserutterance|>",
11
+ "<|endlastuserutterance|>",
12
+ "<|begindsts|>",
13
+ "<|enddsts|>",
14
+ "<|begindst|>",
15
+ "<|enddst|>",
16
+ "<|beginbelief|>",
17
+ "<|endbelief|>",
18
+ "<|beginresponse|>",
19
+ "<|endresponse|>",
20
+ "<|beginaction|>",
21
+ "<|endaction|>",
22
+ "<|beginuseraction|>",
23
+ "<|enduseraction|>",
24
+ "<|sysactions|>",
25
+ "<|beginintent|>",
26
+ "<|endintent|>",
27
+ "<|beginrequestedslots|>",
28
+ "<|endrequestedslots|>",
29
+ "<|promptdst|>",
30
+ "<|promptaction|>",
31
+ "<|promptresponse|>",
32
+ "<|pad|>",
33
+ "<|endoftext|>",
34
+ "<|startoftext|>",
35
+ "<|beginschema|>",
36
+ "<|endschema|>",
37
+ "<|schemaname|>",
38
+ "<|schemadescription|>",
39
+ "<|beginschemaintent|>",
40
+ "<|endschemaintent|>",
41
+ "<|intentrequiredslots|>",
42
+ "<|intentresultslots|>",
43
+ "<|intentoptionalslots|>",
44
+ "<|possiblevalues|>",
45
+ "<|endpossiblevalues|>",
46
+ "<|beginschemaslot|>",
47
+ "<|endschemaslot|>",
48
+ "<|schemaslotvalues|>",
49
+ "<|beginserviceresults|>",
50
+ "<|endserviceresults|>"
51
+ ],
52
+ "bos_token": "<|startoftext|>",
53
+ "clean_up_tokenization_spaces": false,
54
+ "eos_token": "<|endtarget|>",
55
+ "model_max_length": 2048,
56
+ "pad_token": "<|pad|>",
57
+ "sp_model_kwargs": {},
58
+ "tokenizer_class": "LlamaTokenizer",
59
+ "unk_token": {
60
+ "__type": "AddedToken",
61
+ "content": "<unk>",
62
+ "lstrip": false,
63
+ "normalized": true,
64
+ "rstrip": false,
65
+ "single_word": false
66
+ }
67
+ }