MikeRoz commited on
Commit
be15bd3
1 Parent(s): 5c84469

Delete config.yaml

Browse files
Files changed (1) hide show
  1. config.yaml +0 -208
config.yaml DELETED
@@ -1,208 +0,0 @@
1
- .*(llama|alpac|vicuna|guanaco|koala|llava|wizardlm|metharme|pygmalion-7b|pygmalion-2|mythalion|wizard-mega|openbuddy|vigogne|h2ogpt-research|manticore):
2
- model_type: 'llama'
3
- .*(opt-|opt_|opt1|opt3|optfor|galactica|galpaca|pygmalion-350m):
4
- model_type: 'opt'
5
- .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1):
6
- model_type: 'gptj'
7
- .*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm):
8
- model_type: 'gptneox'
9
- .*bloom:
10
- model_type: 'bloom'
11
- .*gpt2:
12
- model_type: 'gpt2'
13
- .*falcon:
14
- model_type: 'falcon'
15
- .*mpt:
16
- model_type: 'mpt'
17
- .*(starcoder|starchat):
18
- model_type: 'starcoder'
19
- .*dolly-v2:
20
- model_type: 'dollyv2'
21
- .*replit:
22
- model_type: 'replit'
23
- .*(oasst|openassistant-|stablelm-7b-sft-v7-epoch-3):
24
- instruction_template: 'Open Assistant'
25
- skip_special_tokens: false
26
- (?!.*galactica)(?!.*reward).*openassistant:
27
- instruction_template: 'Open Assistant'
28
- skip_special_tokens: false
29
- .*galactica:
30
- skip_special_tokens: false
31
- .*dolly-v[0-9]-[0-9]*b:
32
- instruction_template: 'Alpaca'
33
- skip_special_tokens: false
34
- .*alpaca-native-4bit:
35
- instruction_template: 'Alpaca'
36
- custom_stopping_strings: '"### End"'
37
- .*llava:
38
- instruction_template: 'LLaVA'
39
- custom_stopping_strings: '"\n###"'
40
- .*llava.*1.5:
41
- instruction_template: 'Vicuna-v1.1'
42
- .*wizard.*mega:
43
- instruction_template: 'Wizard-Mega'
44
- custom_stopping_strings: '"</s>"'
45
- .*starchat-beta:
46
- instruction_template: 'Starchat-Beta'
47
- custom_stopping_strings: '"<|end|>"'
48
- (?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable)(?!.*chinese).*vicuna:
49
- instruction_template: 'Vicuna-v0'
50
- .*vicuna.*v0:
51
- instruction_template: 'Vicuna-v0'
52
- .*vicuna.*(1.1|1_1|1.3|1_3):
53
- instruction_template: 'Vicuna-v1.1'
54
- .*vicuna.*(1.5|1_5):
55
- instruction_template: 'Vicuna-v1.1'
56
- .*stable.*vicuna:
57
- instruction_template: 'StableVicuna'
58
- (?!.*chat).*chinese-vicuna:
59
- instruction_template: 'Alpaca'
60
- .*chinese-vicuna.*chat:
61
- instruction_template: 'Chinese-Vicuna-Chat'
62
- .*alpaca:
63
- instruction_template: 'Alpaca'
64
- .*koala:
65
- instruction_template: 'Koala'
66
- .*chatglm:
67
- instruction_template: 'ChatGLM'
68
- .*(metharme|pygmalion|mythalion):
69
- instruction_template: 'Metharme'
70
- .*raven:
71
- instruction_template: 'RWKV-Raven'
72
- .*moss-moon.*sft:
73
- instruction_template: 'MOSS'
74
- .*stablelm-tuned:
75
- instruction_template: 'StableLM'
76
- .*galactica.*finetuned:
77
- instruction_template: 'Galactica Finetuned'
78
- .*galactica.*-v2:
79
- instruction_template: 'Galactica v2'
80
- (?!.*finetuned)(?!.*-v2).*galactica:
81
- instruction_template: 'Galactica'
82
- .*guanaco:
83
- instruction_template: 'Guanaco non-chat'
84
- .*baize:
85
- instruction_template: 'Baize'
86
- .*mpt-.*instruct:
87
- instruction_template: 'Alpaca'
88
- .*mpt-.*chat:
89
- instruction_template: 'ChatML'
90
- (?!.*-flan-)(?!.*-t5-).*lamini-:
91
- instruction_template: 'Alpaca'
92
- .*incite.*chat:
93
- instruction_template: 'INCITE-Chat'
94
- .*incite.*instruct:
95
- instruction_template: 'INCITE-Instruct'
96
- .*ziya-:
97
- instruction_template: 'Ziya'
98
- .*koalpaca:
99
- instruction_template: 'KoAlpaca'
100
- .*openbuddy:
101
- instruction_template: 'OpenBuddy'
102
- (?!.*chat).*vigogne:
103
- instruction_template: 'Vigogne-Instruct'
104
- .*vigogne.*chat:
105
- instruction_template: 'Vigogne-Chat'
106
- .*(llama-deus|supercot|llama-natural-instructions|open-llama-0.3t-7b-instruct-dolly-hhrlhf|open-llama-0.3t-7b-open-instruct):
107
- instruction_template: 'Alpaca'
108
- .*bactrian:
109
- instruction_template: 'Bactrian'
110
- .*(h2ogpt-oig-|h2ogpt-oasst1-|h2ogpt-research-oasst1-):
111
- instruction_template: 'INCITE-Chat'
112
- .*h2ogpt-gm-:
113
- instruction_template: 'H2O-prompt_answer'
114
- .*manticore:
115
- instruction_template: 'Manticore Chat'
116
- .*bluemoonrp-(30|13)b:
117
- instruction_template: 'Bluemoon'
118
- .*Nous-Hermes-13b:
119
- instruction_template: 'Alpaca'
120
- .*airoboros:
121
- instruction_template: 'Vicuna-v1.1'
122
- .*airoboros.*1.2:
123
- instruction_template: 'Airoboros-v1.2'
124
- .*alpa(cino|sta):
125
- instruction_template: 'Alpaca'
126
- .*hippogriff:
127
- instruction_template: 'Hippogriff'
128
- .*lazarus:
129
- instruction_template: 'Alpaca'
130
- .*guanaco-.*(7|13|33|65)b:
131
- instruction_template: 'Vicuna-v0'
132
- .*hypermantis:
133
- instruction_template: 'Alpaca'
134
- .*open-llama-.*-open-instruct:
135
- instruction_template: 'Alpaca'
136
- .*starcoder-gpteacher-code-instruct:
137
- instruction_template: 'Alpaca'
138
- .*tulu:
139
- instruction_template: 'Tulu'
140
- .*chronos:
141
- instruction_template: 'Alpaca'
142
- .*samantha:
143
- instruction_template: 'Samantha'
144
- .*wizardcoder:
145
- instruction_template: 'Alpaca'
146
- .*minotaur:
147
- instruction_template: 'Manticore Chat'
148
- .*orca_mini:
149
- instruction_template: 'Orca Mini'
150
- .*(platypus|gplatty|superplatty):
151
- instruction_template: 'Alpaca'
152
- .*(openorca-platypus2):
153
- instruction_template: 'OpenOrca-Platypus2'
154
- custom_stopping_strings: '"### Instruction:", "### Response:"'
155
- .*longchat:
156
- instruction_template: 'Vicuna-v1.1'
157
- .*vicuna-33b:
158
- instruction_template: 'Vicuna-v1.1'
159
- .*redmond-hermes-coder:
160
- instruction_template: 'Alpaca'
161
- .*wizardcoder-15b:
162
- instruction_template: 'Alpaca'
163
- .*wizardlm:
164
- instruction_template: 'Vicuna-v1.1'
165
- .*godzilla:
166
- instruction_template: 'Alpaca'
167
- .*llama(-?)(2|v2).*chat:
168
- instruction_template: 'Llama-v2'
169
- .*newhope:
170
- instruction_template: 'NewHope'
171
- .*stablebeluga2:
172
- instruction_template: 'StableBeluga2'
173
- .*openchat:
174
- instruction_template: 'OpenChat'
175
- .*codellama.*instruct:
176
- instruction_template: 'Llama-v2'
177
- .*(mistral|mixtral).*instruct:
178
- instruction_template: 'Mistral'
179
- .*mistral.*openorca:
180
- instruction_template: 'ChatML'
181
- .*(WizardCoder-Python-34B-V1.0|Phind-CodeLlama-34B-v2|CodeBooga-34B-v0.1):
182
- instruction_template: 'Alpaca'
183
- .*orca-2-(13|7)b:
184
- instruction_template: 'ChatML'
185
- .*openhermes.*mistral:
186
- instruction_template: 'ChatML'
187
- .*Yi-34B-Chat:
188
- instruction_template: 'ChatML'
189
- (dolphin).*:
190
- instruction_template: 'ChatML'
191
- .*synthia:
192
- instruction_template: 'Synthia'
193
- .*(hercules|hyperion):
194
- instruction_template: 'ChatML'
195
- .*command-r:
196
- instruction_template: 'Command-R'
197
- .*xwin-lm-70b-v0.1:
198
- instruction_template: 'Vicuna-v1.1'
199
- .*platypus-yi-34b:
200
- instruction_template: 'Vicuna-v1.1'
201
- .*CausalLM-RP-34B:
202
- instruction_template: 'ChatML'
203
- 34b-beta:
204
- instruction_template: 'ChatML'
205
- .*airoboros-3_1-yi-34b-200k:
206
- instruction_template: 'Llama-v2'
207
- .*chatqa:
208
- instruction_template: 'NVIDIA-ChatQA'