File size: 856 Bytes
7ab99c9
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
# Ref: https://github.com/openai/openai-cookbook/blob/main/examples/How_to_count_tokens_with_tiktoken.ipynb

import sys
import json
from collections import OrderedDict

import tiktoken

if len(sys.argv) <= 1:
    sys.stderr.write("Expect a file name")
    sys.exit(-1)

vocab_json = sys.argv[1]

# cl100k_base: gpt-4, gpt-3.5-turbo, text-embedding-ada-002
# p50k_base: Codex models, text-davinci-002, text-davinci-003
# r50k_base (or gpt2): GPT-3 models like davinci
encoding = tiktoken.get_encoding("r50k_base")

vocabulary = OrderedDict()

for i in range(encoding.n_vocab):
    try:
        bytes = encoding.decode_single_token_bytes(i).decode('utf-8')
        vocabulary[bytes] = i
    except:
        sys.stderr.write("no token for %d\n" %i)

with open(vocab_json, "w", encoding='utf8') as f:
    json.dump(vocabulary, f, indent=4, ensure_ascii=False)