TitanML Co commited on
Commit
017c3c3
·
verified ·
1 Parent(s): ba3daef

Upload folder using huggingface_hub

Browse files
Files changed (5) hide show
  1. README.md +3 -0
  2. awq_cache.json +87 -0
  3. ct2_cache.json +170 -0
  4. ggml_cache.json +176 -0
  5. name_funnel.json +45 -0
README.md ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ # AWQ Cache
2
+
3
+ Mapping HF model names to the names as saved in the TitanML directory.
awq_cache.json ADDED
@@ -0,0 +1,87 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "llama-2-7b": {
3
+ "repo": "TitanML/llama2-7B-base-4bit-AWQ"
4
+ },
5
+ "llama-2-7b-hf": {
6
+ "repo": "TitanML/llama2-7B-base-4bit-AWQ"
7
+ },
8
+ "meta-llama/llama-2-7b": {
9
+ "repo": "TitanML/llama2-7B-base-4bit-AWQ"
10
+ },
11
+ "meta-llama/llama-2-7b-hf": {
12
+ "repo": "TitanML/llama2-7B-base-4bit-AWQ"
13
+ },
14
+ "llama-2-13b": {
15
+ "repo": "TitanML/llama2-13B-base-4bit-AWQ"
16
+ },
17
+ "llama-2-13b-hf": {
18
+ "repo": "TitanML/llama2-13B-base-4bit-AWQ"
19
+ },
20
+ "meta-llama/llama-2-13b": {
21
+ "repo": "TitanML/llama2-13B-base-4bit-AWQ"
22
+ },
23
+ "meta-llama/llama-2-13b-hf": {
24
+ "repo": "TitanML/llama2-13B-base-4bit-AWQ"
25
+ },
26
+ "llama-2-7b-chat": {
27
+ "repo": "TitanML/llama2-7B-chat-4bit-AWQ"
28
+ },
29
+ "llama-2-7b-chat-hf": {
30
+ "repo": "TitanML/llama2-7B-chat-4bit-AWQ"
31
+ },
32
+ "meta-llama/llama-2-7b-chat": {
33
+ "repo": "TitanML/llama2-7B-chat-4bit-AWQ"
34
+ },
35
+ "meta-llama/llama-2-7b-chat-hf": {
36
+ "repo": "TitanML/llama2-7B-chat-4bit-AWQ"
37
+ },
38
+ "llama-2-13b-chat": {
39
+ "repo": "TitanML/llama2-13B-chat-4bit-AWQ"
40
+ },
41
+ "llama-2-13b-chat-hf": {
42
+ "repo": "TitanML/llama2-13B-chat-4bit-AWQ"
43
+ },
44
+ "meta-llama/llama-2-13b-chat": {
45
+ "repo": "TitanML/llama2-13B-chat-4bit-AWQ"
46
+ },
47
+ "meta-llama/llama-2-13b-chat-hf": {
48
+ "repo": "TitanML/llama2-13B-chat-4bit-AWQ"
49
+ },
50
+ "vicuna-13b": {
51
+ "repo": "TitanML/vicuna-13B-base-4bit-AWQ"
52
+ },
53
+ "vicuna-7b": {
54
+ "repo": "TitanML/vicuna-7B-base-4bit-AWQ"
55
+ },
56
+ "mpt-7b": {
57
+ "repo": "TitanML/mpt-7b-base-8k-4bit-AWQ"
58
+ },
59
+ "mosaicml/mpt-7b": {
60
+ "repo": "TitanML/mpt-7b-base-8k-4bit-AWQ"
61
+ },
62
+ "mpt-7b-chat": {
63
+ "repo": "TitanML/mpt-7b-chat-8k-4bit-AWQ"
64
+ },
65
+ "mosaicml/mpt-7b-chat": {
66
+ "repo": "TitanML/mpt-7b-chat-8k-4bit-AWQ"
67
+ },
68
+ "mpt-7b-instruct": {
69
+ "repo": "TitanML/mpt-7b-instruct-8k-4bit-AWQ"
70
+ },
71
+ "mosaicml/mpt-7b-instruct": {
72
+ "repo": "TitanML/mpt-7b-instruct-8k-4bit-AWQ"
73
+ },
74
+ "opt-13b": {
75
+ "repo": "TitanML/opt-13b-base-4bit-AWQ"
76
+ },
77
+ "facebook/opt-13b": {
78
+ "repo": "TitanML/opt-13b-base-4bit-AWQ"
79
+ },
80
+ "opt-30b": {
81
+ "repo": "TitanML/opt-30b-base-4bit-AWQ"
82
+ },
83
+ "facebook/opt-30b": {
84
+ "repo": "TitanML/opt-30b-base-4bit-AWQ"
85
+ }
86
+ }
87
+
ct2_cache.json ADDED
@@ -0,0 +1,170 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "tiiuae/falcon-7b": {
3
+ "quant_type": "int8",
4
+ "repo": "TitanML/ct2-int8-falcon-7b"
5
+ },
6
+ "falcon-7b": {
7
+ "quant_type": "int8",
8
+ "repo": "TitanML/ct2-int8-falcon-7b"
9
+ },
10
+ "mosaicml/mpt-7b": {
11
+ "quant_type": "int8",
12
+ "repo": "TitanML/ct2-int8-mtb-7b"
13
+ },
14
+ "mpt-7b": {
15
+ "quant_type": "int8",
16
+ "repo": "TitanML/ct2-int8-mtb-7b"
17
+ },
18
+ "mosaicml/mpt-7b-storywriter": {
19
+ "quant_type": "int8",
20
+ "repo": "TitanML/ct2-int8-mtb-7b-storywriter"
21
+ },
22
+ "mpt-7b-storywriter": {
23
+ "quant_type": "int8",
24
+ "repo": "TitanML/ct2-int8-mtb-7b-storywriter"
25
+ },
26
+ "mosaicml/mpt-7b-chat": {
27
+ "quant_type": "int8",
28
+ "repo": "TitanML/ct2-int8-mtb-7b-chat"
29
+ },
30
+ "mpt-7b-chat": {
31
+ "quant_type": "int8",
32
+ "repo": "TitanML/ct2-int8-mtb-7b-chat"
33
+ },
34
+ "mosaicml/mpt-7b-8k": {
35
+ "quant_type": "int8",
36
+ "repo": "TitanML/ct2-int8-mtb-7b-8k"
37
+ },
38
+ "mpt-7b-8k": {
39
+ "quant_type": "int8",
40
+ "repo": "TitanML/ct2-int8-mtb-7b-8k"
41
+ },
42
+ "mosaicml/mpt-7b-8k-instruct": {
43
+ "quant_type": "int8",
44
+ "repo": "TitanML/ct2-int8-mtb-7b-8k-instruct"
45
+ },
46
+ "mpt-7b-8k-instruct": {
47
+ "quant_type": "int8",
48
+ "repo": "TitanML/ct2-int8-mtb-7b-8k-instruct"
49
+ },
50
+ "mosaicml/mpt-7b-8k-chat": {
51
+ "quant_type": "int8",
52
+ "repo": "TitanML/ct2-int8-mtb-7b-8k-chat"
53
+ },
54
+ "mpt-7b-8k-chat": {
55
+ "quant_type": "int8",
56
+ "repo": "TitanML/ct2-int8-mtb-7b-8k-chat"
57
+ },
58
+ "togethercomputer/redpajama-incite-7b-base": {
59
+ "quant_type": "int8",
60
+ "repo": "TitanML/ct2-int8-redpajama-7b-base"
61
+ },
62
+ "redpajama-incite-7b-base": {
63
+ "quant_type": "int8",
64
+ "repo": "TitanML/ct2-int8-redpajama-7b-base"
65
+ },
66
+ "togethercomputer/redpajama-incite-7b-instruct": {
67
+ "quant_type": "int8",
68
+ "repo": "TitanML/ct2-int8-redpajama-7b-instruct"
69
+ },
70
+ "redpajama-incite-7b-instruct": {
71
+ "quant_type": "int8",
72
+ "repo": "TitanML/ct2-int8-redpajama-7b-instruct"
73
+ },
74
+ "togethercomputer/redpajama-incite-7b-chat": {
75
+ "quant_type": "int8",
76
+ "repo": "TitanML/ct2-int8-redpajama-7b-chat"
77
+ },
78
+ "redpajama-incite-7b-chat": {
79
+ "quant_type": "int8",
80
+ "repo": "TitanML/ct2-int8-redpajama-7b-chat"
81
+ },
82
+ "meta-llama/llama-2-7b-chat-hf": {
83
+ "quant_type": "int8",
84
+ "repo": "TitanML/ct2-int8-llama-2-7b-chat"
85
+ },
86
+ "meta-llama/llama-2-7b-chat": {
87
+ "quant_type": "int8",
88
+ "repo": "TitanML/ct2-int8-llama-2-7b-chat"
89
+ },
90
+ "llama-2-7b-chat-hf": {
91
+ "quant_type": "int8",
92
+ "repo": "TitanML/ct2-int8-llama-2-7b-chat"
93
+ },
94
+ "llama-2-7b-chat": {
95
+ "quant_type": "int8",
96
+ "repo": "TitanML/ct2-int8-llama-2-7b-chat"
97
+ },
98
+ "google/flan-t5-xl": {
99
+ "quant_type": "int8",
100
+ "repo": "TitanML/ct2-int8-flan-xl"
101
+ },
102
+ "flan-t5-xl": {
103
+ "quant_type": "int8",
104
+ "repo": "TitanML/ct2-int8-flan-xl"
105
+ },
106
+ "conceptofmind/flan-open-llama-7b": {
107
+ "quant_type": "int8",
108
+ "repo": "TitanML/ct2-int8-flan-open-llama-7b"
109
+ },
110
+ "flan-open-llama-7b": {
111
+ "quant_type": "int8",
112
+ "repo": "TitanML/ct2-int8-flan-open-llama-7b"
113
+ },
114
+ "openlm-research/open_llama_7b": {
115
+ "quant_type": "int8",
116
+ "repo": "TitanML/ct2-int8-open-llama-7b"
117
+ },
118
+ "open_llama_7b": {
119
+ "quant_type": "int8",
120
+ "repo": "TitanML/ct2-int8-open-llama-7b"
121
+ },
122
+ "openlm-research/open_llama_7b_v2": {
123
+ "quant_type": "int8",
124
+ "repo": "TitanML/ct2-int8-open-llama-7b-v2"
125
+ },
126
+ "open_llama_7b_v2": {
127
+ "quant_type": "int8",
128
+ "repo": "TitanML/ct2-int8-open-llama-7b-v2"
129
+ },
130
+ "salesforce/xgen-7b-8k-base": {
131
+ "quant_type": "int8",
132
+ "repo": "TitanML/ct2-int8-xgen-7b-8k-base"
133
+ },
134
+ "xgen-7b-8k-base": {
135
+ "quant_type": "int8",
136
+ "repo": "TitanML/ct2-int8-xgen-7b-8k-base"
137
+ },
138
+ "salesforce/xgen-7b-8k-inst": {
139
+ "quant_type": "int8",
140
+ "repo": "TitanML/ct2-int8-xgen-7b-8k-inst"
141
+ },
142
+ "xgen-7b-8k-inst": {
143
+ "quant_type": "int8",
144
+ "repo": "TitanML/ct2-int8-xgen-7b-8k-inst"
145
+ },
146
+ "salesforce/codegen25-7b-instruct": {
147
+ "quant_type": "int8",
148
+ "repo": "TitanML/ct2-int8-codegen25-7b-instruct"
149
+ },
150
+ "codegen25-7b-instruct": {
151
+ "quant_type": "int8",
152
+ "repo": "TitanML/ct2-int8-codegen25-7b-instruct"
153
+ },
154
+ "stabilityai/stablelm-tuned-alpha-7b": {
155
+ "quant_type": "int8",
156
+ "repo": "TitanML/ct2-int8-stablelm-7b"
157
+ },
158
+ "stablelm-tuned-alpha-7b": {
159
+ "quant_type": "int8",
160
+ "repo": "TitanML/ct2-int8-stablelm-7b"
161
+ },
162
+ "tiiuae/falcon-7b-instruct": {
163
+ "quant_type": "int8",
164
+ "repo": "TitanML/ct2-int8-falcon-7b-instruct"
165
+ },
166
+ "falcon-7b-instruct": {
167
+ "quant_type": "int8",
168
+ "repo": "TitanML/ct2-int8-falcon-7b-instruct"
169
+ }
170
+ }
ggml_cache.json ADDED
@@ -0,0 +1,176 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "llama-2-7b": {
3
+ "repo": "TheBloke/Llama-2-7B-GGML",
4
+ "filename": "llama-2-7b.ggmlv3.q4_0.bin",
5
+ "model_type": "llama",
6
+ "tokenizer": "llama-local"
7
+ },
8
+ "meta-llama/llama-2-7b": {
9
+ "repo": "TheBloke/Llama-2-7B-GGML",
10
+ "filename": "llama-2-7b.ggmlv3.q4_0.bin",
11
+ "model_type": "llama",
12
+ "tokenizer": "llama-local"
13
+ },
14
+ "llama-2-7b-hf": {
15
+ "repo": "TheBloke/Llama-2-7B-GGML",
16
+ "filename": "llama-2-7b.ggmlv3.q4_0.bin",
17
+ "model_type": "llama",
18
+ "tokenizer": "llama-local"
19
+ },
20
+ "meta-llama/llama-2-7b-hf": {
21
+ "repo": "TheBloke/Llama-2-7B-GGML",
22
+ "filename": "llama-2-7b.ggmlv3.q4_0.bin",
23
+ "model_type": "llama",
24
+ "tokenizer": "llama-local"
25
+ },
26
+ "llama-2-7b-chat": {
27
+ "repo": "TheBloke/Llama-2-7B-Chat-GGML",
28
+ "filename": "llama-2-7b-chat.ggmlv3.q4_0.bin",
29
+ "model_type": "llama",
30
+ "tokenizer": "llama-local"
31
+ },
32
+ "meta-llama/llama-2-7b-chat": {
33
+ "repo": "TheBloke/Llama-2-7B-Chat-GGML",
34
+ "filename": "llama-2-7b-chat.ggmlv3.q4_0.bin",
35
+ "model_type": "llama",
36
+ "tokenizer": "llama-local"
37
+ },
38
+ "llama-2-7b-chat-hf": {
39
+ "repo": "TheBloke/Llama-2-7B-Chat-GGML",
40
+ "filename": "llama-2-7b-chat.ggmlv3.q4_0.bin",
41
+ "model_type": "llama",
42
+ "tokenizer": "llama-local"
43
+ },
44
+ "meta-llama/llama-2-7b-chat-hf": {
45
+ "repo": "TheBloke/Llama-2-7B-Chat-GGML",
46
+ "filename": "llama-2-7b-chat.ggmlv3.q4_0.bin",
47
+ "model_type": "llama",
48
+ "tokenizer": "llama-local"
49
+ },
50
+ "llama-2-13b": {
51
+ "repo": "TheBloke/Llama-2-13B-GGML",
52
+ "filename": "llama-2-13b.ggmlv3.q4_0.bin",
53
+ "model_type": "llama",
54
+ "tokenizer": "llama-local"
55
+ },
56
+ "meta-llama/llama-2-13b": {
57
+ "repo": "TheBloke/Llama-2-13B-GGML",
58
+ "filename": "llama-2-13b.ggmlv3.q4_0.bin",
59
+ "model_type": "llama",
60
+ "tokenizer": "llama-local"
61
+ },
62
+ "llama-2-13b-hf": {
63
+ "repo": "TheBloke/Llama-2-13B-GGML",
64
+ "filename": "llama-2-13b.ggmlv3.q4_0.bin",
65
+ "model_type": "llama",
66
+ "tokenizer": "llama-local"
67
+ },
68
+ "meta-llama/llama-2-13b-hf": {
69
+ "repo": "TheBloke/Llama-2-13B-GGML",
70
+ "filename": "llama-2-13b.ggmlv3.q4_0.bin",
71
+ "model_type": "llama",
72
+ "tokenizer": "llama-local"
73
+ },
74
+ "llama-2-13b-chat": {
75
+ "repo": "TheBloke/Llama-2-13B-Chat-GGML",
76
+ "filename": "llama-2-13b-chat.ggmlv3.q4_0.bin",
77
+ "model_type": "llama",
78
+ "tokenizer": "llama-local"
79
+ },
80
+ "meta-llama/llama-2-13b-chat": {
81
+ "repo": "TheBloke/Llama-2-13B-Chat-GGML",
82
+ "filename": "llama-2-13b-chat.ggmlv3.q4_0.bin",
83
+ "model_type": "llama",
84
+ "tokenizer": "llama-local"
85
+ },
86
+ "llama-2-13b-chat-hf": {
87
+ "repo": "TheBloke/Llama-2-13B-Chat-GGML",
88
+ "filename": "llama-2-13b-chat.ggmlv3.q4_0.bin",
89
+ "model_type": "llama",
90
+ "tokenizer": "llama-local"
91
+ },
92
+ "meta-llama/llama-2-13b-chat-hf": {
93
+ "repo": "TheBloke/Llama-2-13B-Chat-GGML",
94
+ "filename": "llama-2-13b-chat.ggmlv3.q4_0.bin",
95
+ "model_type": "llama",
96
+ "tokenizer": "llama-local"
97
+ },
98
+ "llama-2-70b": {
99
+ "repo": "TheBloke/Llama-2-70B-GGML",
100
+ "filename": "llama-2-70b.ggmlv3.q4_0.bin",
101
+ "model_type": "llama",
102
+ "tokenizer": "llama-local"
103
+ },
104
+ "meta-llama/llama-2-70b": {
105
+ "repo": "TheBloke/Llama-2-70B-GGML",
106
+ "filename": "llama-2-70b.ggmlv3.q4_0.bin",
107
+ "model_type": "llama",
108
+ "tokenizer": "llama-local"
109
+ },
110
+ "llama-2-70b-hf": {
111
+ "repo": "TheBloke/Llama-2-70B-GGML",
112
+ "filename": "llama-2-70b.ggmlv3.q4_0.bin",
113
+ "model_type": "llama",
114
+ "tokenizer": "llama-local"
115
+ },
116
+ "meta-llama/llama-2-70b-hf": {
117
+ "repo": "TheBloke/Llama-2-70B-GGML",
118
+ "filename": "llama-2-70b.ggmlv3.q4_0.bin",
119
+ "model_type": "llama",
120
+ "tokenizer": "llama-local"
121
+ },
122
+ "llama-2-70b-chat": {
123
+ "repo": "TheBloke/Llama-2-70B-Chat-GGML",
124
+ "filename": "llama-2-70b-chat.ggmlv3.q4_0.bin",
125
+ "model_type": "llama",
126
+ "tokenizer": "llama-local"
127
+ },
128
+ "meta-llama/llama-2-70b-chat": {
129
+ "repo": "TheBloke/Llama-2-70B-Chat-GGML",
130
+ "filename": "llama-2-70b-chat.ggmlv3.q4_0.bin",
131
+ "model_type": "llama",
132
+ "tokenizer": "llama-local"
133
+ },
134
+ "llama-2-70b-chat-hf": {
135
+ "repo": "TheBloke/Llama-2-70B-Chat-GGML",
136
+ "filename": "llama-2-70b-chat.ggmlv3.q4_0.bin",
137
+ "model_type": "llama",
138
+ "tokenizer": "llama-local"
139
+ },
140
+ "meta-llama/llama-2-70b-chat-hf": {
141
+ "repo": "TheBloke/Llama-2-70B-Chat-GGML",
142
+ "filename": "llama-2-70b-chat.ggmlv3.q4_0.bin",
143
+ "model_type": "llama",
144
+ "tokenizer": "llama-local"
145
+ },
146
+ "falcon-7b": {
147
+ "repo": "TheBloke/falcon-7b-instruct-GGML",
148
+ "filename": "falcon-7b-instruct.ggccv1.q4_0.bin",
149
+ "model_type": "falcon",
150
+ "tokenizer": "tiiuae/falcon-7b"
151
+ },
152
+ "tiiuae/falcon-7b": {
153
+ "repo": "TheBloke/falcon-7b-instruct-GGML",
154
+ "filename": "falcon-7b-instruct.ggccv1.q4_0.bin",
155
+ "model_type": "falcon",
156
+ "tokenizer": "tiiuae/falcon-7b"
157
+ },
158
+ "falcon-40b": {
159
+ "repo": "TheBloke/falcon-40b-instruct-GGML",
160
+ "filename": "falcon-40b-instruct.ggccv1.q4_0.bin",
161
+ "model_type": "falcon",
162
+ "tokenizer": "tiiuae/falcon-7b"
163
+ },
164
+ "ttiuae/falcon-40b": {
165
+ "repo": "TheBloke/falcon-40b-instruct-GGML",
166
+ "filename": "falcon-40b-instruct.ggccv1.q4_0.bin",
167
+ "model_type": "falcon",
168
+ "tokenizer": "tiiuae/falcon-7b"
169
+ },
170
+ "gpt2": {
171
+ "repo": "marella/gpt-2-ggml",
172
+ "filename": "ggml-model.bin",
173
+ "model_type": "gpt2",
174
+ "tokenizer": "gpt2"
175
+ }
176
+ }
name_funnel.json ADDED
@@ -0,0 +1,45 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "llama-2-7b": "llama-2-7b",
3
+ "llama-2-7b-hf": "llama-2-7b",
4
+ "meta-llama/llama-2-7b": "llama-2-7b",
5
+ "meta-llama/llama-2-7b-hf": "llama-2-7b",
6
+ "llama-2-7b-chat": "llama-2-7b-chat",
7
+ "llama-2-7b-chat-hf": "llama-2-7b-chat",
8
+ "meta-llama/llama-2-7b-chat": "llama-2-7b-chat",
9
+ "meta-llama/llama-2-7b-chat-hf": "llama-2-7b-chat",
10
+ "llama-2-13b": "llama-2-13b",
11
+ "llama-2-13b-hf": "llama-2-13b",
12
+ "meta-llama/llama-2-13b": "llama-2-13b",
13
+ "meta-llama/llama-2-13b-hf": "llama-2-13b",
14
+ "llama-2-13b-chat": "llama-2-13b-chat",
15
+ "llama-2-13b-chat-hf": "llama-2-13b-chat",
16
+ "meta-llama/llama-2-13b-chat": "llama-2-13b-chat",
17
+ "meta-llama/llama-2-13b-chat-hf": "llama-2-13b-chat",
18
+ "llama-2-70b": "llama-2-70b",
19
+ "llama-2-70b-hf": "llama-2-70b",
20
+ "meta-llama/llama-2-70b": "llama-2-70b",
21
+ "meta-llama/llama-2-70b-hf": "llama-2-70b",
22
+ "llama-2-70b-chat": "llama-2-70b-chat",
23
+ "llama-2-70b-chat-hf": "llama-2-70b-chat",
24
+ "meta-llama/llama-2-70b-chat": "llama-2-70b-chat",
25
+ "meta-llama/llama-2-70b-chat-hf": "llama-2-70b-chat",
26
+ "falcon-7b": "falcon-7b",
27
+ "tiiuae/falcon-7b": "falcon-7b",
28
+ "falcon-40b": "falcon-40b",
29
+ "tiiuae/falcon-40b": "falcon-40b",
30
+ "gpt2": "gpt2",
31
+ "vicuna-7b": "vicuna-7b",
32
+ "lmsys/vicuna-7b-v1.5": "vicuna-7b",
33
+ "vicuna-13b": "vicuna-13b",
34
+ "lmsys/vicuna-13b-v1.5": "vicuna-13b",
35
+ "facebook/opt-13b": "opt-13b",
36
+ "opt-13b": "opt-13b",
37
+ "facebook/opt-30b": "opt-30b",
38
+ "opt-30b": "opt-30b",
39
+ "mpt-7b": "mpt-7b",
40
+ "mosaicml/mpt-7b": "mpt-7b",
41
+ "mpt-7b-chat": "mpt-7b-chat",
42
+ "mosaicml/mpt-7b-chat": "mpt-7b-chat",
43
+ "mpt-7b-instruct": "mpt-7b-instruct",
44
+ "mosaicml/mpt-7b-instruct": "mpt-7b-instruct",
45
+ }