Siyuan commited on
Commit
bae8f97
·
verified ·
1 Parent(s): 9eca361

Upload folder using huggingface_hub

Browse files
mlc-chat-config.json ADDED
@@ -0,0 +1,83 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": "0.1.0",
3
+ "model_type": "internlm2",
4
+ "quantization": "q4f16_1",
5
+ "model_config": {
6
+ "vocab_size": 92544,
7
+ "hidden_size": 2048,
8
+ "num_hidden_layers": 24,
9
+ "num_attention_heads": 16,
10
+ "num_key_value_heads": 8,
11
+ "rms_norm_eps": 1e-05,
12
+ "intermediate_size": 8192,
13
+ "bias": false,
14
+ "use_cache": true,
15
+ "rope_theta": 1000000,
16
+ "pad_token_id": 2,
17
+ "bos_token_id": 1,
18
+ "eos_token_id": 2,
19
+ "context_window_size": 32768,
20
+ "prefill_chunk_size": 8192,
21
+ "tensor_parallel_shards": 1,
22
+ "max_batch_size": 128,
23
+ "head_dim": 128
24
+ },
25
+ "vocab_size": 92544,
26
+ "context_window_size": 32768,
27
+ "sliding_window_size": -1,
28
+ "prefill_chunk_size": 8192,
29
+ "attention_sink_size": -1,
30
+ "tensor_parallel_shards": 1,
31
+ "pipeline_parallel_stages": 1,
32
+ "temperature": 1.0,
33
+ "presence_penalty": 0.0,
34
+ "frequency_penalty": 0.0,
35
+ "repetition_penalty": 1.0,
36
+ "top_p": 1.0,
37
+ "tokenizer_files": [
38
+ "tokenizer.model",
39
+ "tokenizer_config.json",
40
+ "tokenizer.json"
41
+ ],
42
+ "tokenizer_info": {
43
+ "token_postproc_method": "byte_fallback",
44
+ "prepend_space_in_encode": false,
45
+ "strip_space_in_decode": false
46
+ },
47
+ "conv_template": {
48
+ "name": "chatml",
49
+ "system_template": "<|im_start|>system\n{system_message}<|im_end|>\n",
50
+ "system_message": "A conversation between a user and an LLM-based AI assistant. The assistant gives helpful and honest answers.",
51
+ "system_prefix_token_ids": null,
52
+ "add_role_after_system_message": true,
53
+ "roles": {
54
+ "user": "<|im_start|>user",
55
+ "assistant": "<|im_start|>assistant"
56
+ },
57
+ "role_templates": {
58
+ "user": "{user_message}",
59
+ "assistant": "{assistant_message}",
60
+ "tool": "{tool_message}"
61
+ },
62
+ "messages": [],
63
+ "seps": [
64
+ "<|im_end|>\n"
65
+ ],
66
+ "role_content_sep": "\n",
67
+ "role_empty_sep": "\n",
68
+ "stop_str": [
69
+ "<|im_end|>"
70
+ ],
71
+ "stop_token_ids": [
72
+ 2
73
+ ],
74
+ "function_string": "",
75
+ "use_function_calling": false
76
+ },
77
+ "pad_token_id": 2,
78
+ "bos_token_id": 1,
79
+ "eos_token_id": [
80
+ 2,
81
+ 92542
82
+ ]
83
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d90af0b101c05c60cc9c0a9213ae16c9b6d512e46d03433343a40d19584df8db
3
+ size 25956352
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5d1e30c033d4a992b6b2ab044177fd8b5d4db65acf7112cf627fa17b307ea076
3
+ size 33300480
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f6f7a83866d495da52c301624ac4b4c664bc0fb6458ccb0d0d56d931ec8ddabf
3
+ size 25956352
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:866ffd250f8034dff7c936129a8e71c071b02854a545e467c00c4a2249facdd4
3
+ size 33300480
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69d445d1373ededb5f2e31f987c3a3ce1557cedc05263759468c01071fba9cc2
3
+ size 16777216
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:468fb8820cdda933c398af6e1a508ef3a7a0d49854b0c4053fd224255822aa01
3
+ size 32518144
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d8d116111d2c9731ba14a80619330101dbb87d285ceb7c97d63798878ce1027
3
+ size 33038336
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7d59ddd4fc390594656f1e2ac25bdb930b7c455545e05f40f74cfb1cc8e04c9a
3
+ size 25956352
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:491a39a3d8590f78313e8f93f20028637fcead24d488a46fac52257f1368e2f8
3
+ size 94765056
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd12e0e6b9ab99b6c1ade731783e1918ac5c2835e884edb2e6c2f327a5344332
3
+ size 33091584
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:250eea53342207e00e3f2fb87ddb559907684074875060fe1b00b3b27527dd71
3
+ size 33038336
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6ce589e4cbd56b667dd208848dc3f1bcd99a66a25f5ef48490259fa203ebddf2
3
+ size 25956352
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5647a83ff57459ec3679d9bd64976f4ae1a8a03c6558c748d67bfdc9a21b0a4
3
+ size 16777216
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:36bc4eecce4703304a0ac315367de9620cfd98575070564183cf7a0a8e3c0b9f
3
+ size 33300480
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e753308a47c2a6ddcca2e4b33fd8ce890174a584a02699f384d5e80202e0d553
3
+ size 16777216
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:68dd111f5f7e545d38fca4cd58908ad071ede6436613e30a505e4a8cd9e45e67
3
+ size 32518144
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:90bf632fc00ab38cb649cec92e523d4cacafb52830cc1c16cb952251af8d2bfc
3
+ size 33038336
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a222e2759ad59516312667a475d202898c076b93f1c1f33849bfd61a4114198d
3
+ size 25956352
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:81fba662c86f745ed5c91eb926c6d3b05af6480a94b13db816932523268cab0d
3
+ size 33300480
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3268e5d2fb531210b63811b000c7ce82cfdf3acc0c4281835726af31cf7b7782
3
+ size 16777216
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aabc935998b88eafbb308601d5c21bc30aa9dd128ac2d082b35dc580e1d955e7
3
+ size 32518144
params_shard_28.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f5801cb95aaefa5345f229c934894cc4ba946ccdc4bca455feba6d392a3e4d2
3
+ size 33038336
params_shard_29.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b7d9bb5bd1398f3a64df663789deac8ff9afeb314ba8fa941d453a5ed3007eea
3
+ size 25956352
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b5cd8d5a33094187b0536a6b7432dd1fabb361e0bcaf1c17c11e0b5e849f98e8
3
+ size 32518144
params_shard_30.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d253f469683d08626ec22f5deb25084007acd4a55752249e4720f625b63abfcb
3
+ size 33300480
params_shard_31.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9b92df9381566ea0133d94a7595371c7dbb4f59eca4067e47caebcb3c40d162d
3
+ size 94765056
params_shard_32.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:744984706905ec457afc7e1d17dd53e42994ca7e97381062284ee0166802de81
3
+ size 23388160
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4e3da8178c22d97b8f0f2bb9e1f8f721594623bd20bf05f76eaa277dca8123a4
3
+ size 33038336
params_shard_5.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ddd97398c214887c01a04c625c6b7a027d94bb4209f75649be6f375110866970
3
+ size 16777216
params_shard_6.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ecb7314f4751ce97fbdaee82ad5f6a07a95ba2170caba2ccda50b6fc72d37fc1
3
+ size 33034240
params_shard_7.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4301ffad70496c0d693c34c1663f09888c274dee6187486095c0895106965df2
3
+ size 16777216
params_shard_8.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:48bd9f46ab2a196849c020bfb741dd0b8f234d349d5eb26ff8ed1857c0199150
3
+ size 32518144
params_shard_9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e3283ebaf4019423d889c34cb9f72cd9e193b71bd089c0d75481e62b772537f5
3
+ size 33038336
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f868398fc4e05ee1e8aeba95ddf18ddcc45b8bce55d5093bead5bbf80429b48b
3
+ size 1477754
tokenizer_config.json ADDED
@@ -0,0 +1,102 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": true,
3
+ "add_eos_token": false,
4
+ "added_tokens_decoder": {
5
+ "0": {
6
+ "content": "<unk>",
7
+ "lstrip": false,
8
+ "normalized": false,
9
+ "rstrip": false,
10
+ "single_word": false,
11
+ "special": true
12
+ },
13
+ "1": {
14
+ "content": "<s>",
15
+ "lstrip": false,
16
+ "normalized": false,
17
+ "rstrip": false,
18
+ "single_word": false,
19
+ "special": true
20
+ },
21
+ "2": {
22
+ "content": "</s>",
23
+ "lstrip": false,
24
+ "normalized": false,
25
+ "rstrip": false,
26
+ "single_word": false,
27
+ "special": true
28
+ },
29
+ "92538": {
30
+ "content": "<|plugin|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false,
35
+ "special": true
36
+ },
37
+ "92539": {
38
+ "content": "<|interpreter|>",
39
+ "lstrip": false,
40
+ "normalized": false,
41
+ "rstrip": false,
42
+ "single_word": false,
43
+ "special": true
44
+ },
45
+ "92540": {
46
+ "content": "<|action_end|>",
47
+ "lstrip": false,
48
+ "normalized": false,
49
+ "rstrip": false,
50
+ "single_word": false,
51
+ "special": true
52
+ },
53
+ "92541": {
54
+ "content": "<|action_start|>",
55
+ "lstrip": false,
56
+ "normalized": false,
57
+ "rstrip": false,
58
+ "single_word": false,
59
+ "special": true
60
+ },
61
+ "92542": {
62
+ "content": "<|im_end|>",
63
+ "lstrip": false,
64
+ "normalized": false,
65
+ "rstrip": false,
66
+ "single_word": false,
67
+ "special": true
68
+ },
69
+ "92543": {
70
+ "content": "<|im_start|>",
71
+ "lstrip": false,
72
+ "normalized": false,
73
+ "rstrip": false,
74
+ "single_word": false,
75
+ "special": true
76
+ }
77
+ },
78
+ "additional_special_tokens": [
79
+ "<|im_start|>",
80
+ "<|im_end|>",
81
+ "<|action_start|>",
82
+ "<|action_end|>",
83
+ "<|interpreter|>",
84
+ "<|plugin|>"
85
+ ],
86
+ "auto_map": {
87
+ "AutoTokenizer": [
88
+ "tokenization_internlm2.InternLM2Tokenizer",
89
+ "tokenization_internlm2_fast.InternLM2TokenizerFast"
90
+ ]
91
+ },
92
+ "bos_token": "<s>",
93
+ "chat_template": "{{ bos_token }}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}",
94
+ "clean_up_tokenization_spaces": false,
95
+ "decode_with_prefix_space": false,
96
+ "eos_token": "</s>",
97
+ "model_max_length": 1000000000000000019884624838656,
98
+ "pad_token": "</s>",
99
+ "sp_model_kwargs": null,
100
+ "tokenizer_class": "InternLM2Tokenizer",
101
+ "unk_token": "<unk>"
102
+ }