NoQuest commited on
Commit
94f5005
·
1 Parent(s): c4134ff

add orca model and change config.yaml

Browse files
Files changed (2) hide show
  1. models/config.yaml +155 -82
  2. models/orca-mini-3b.bin +3 -0
models/config.yaml CHANGED
@@ -1,206 +1,279 @@
1
- .*(llama|alpac|vicuna|guanaco|koala|llava|wizardlm|metharme|pygmalion-7b|pygmalion-2|mythalion|wizard-mega|openbuddy|vigogne|h2ogpt-research|manticore):
2
  model_type: 'llama'
3
  .*(opt-|opt_|opt1|opt3|optfor|galactica|galpaca|pygmalion-350m):
4
  model_type: 'opt'
5
  .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1):
6
  model_type: 'gptj'
7
  .*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm):
8
- model_type: 'gptneox'
 
 
9
  .*bloom:
10
  model_type: 'bloom'
11
- .*gpt2:
12
- model_type: 'gpt2'
13
- .*falcon:
14
- model_type: 'falcon'
15
- .*mpt:
16
- model_type: 'mpt'
17
- .*(starcoder|starchat):
18
- model_type: 'starcoder'
19
- .*dolly-v2:
20
- model_type: 'dollyv2'
21
- .*replit:
22
- model_type: 'replit'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
23
  .*(oasst|openassistant-|stablelm-7b-sft-v7-epoch-3):
 
24
  instruction_template: 'Open Assistant'
25
  skip_special_tokens: false
26
  (?!.*galactica)(?!.*reward).*openassistant:
 
27
  instruction_template: 'Open Assistant'
28
  skip_special_tokens: false
29
- .*galactica:
30
- skip_special_tokens: false
31
- .*dolly-v[0-9]-[0-9]*b:
32
- instruction_template: 'Alpaca'
33
- skip_special_tokens: false
34
- .*alpaca-native-4bit:
35
- instruction_template: 'Alpaca'
36
- custom_stopping_strings: '"### End"'
37
- .*llava:
38
- instruction_template: 'LLaVA'
39
- custom_stopping_strings: '"\n###"'
40
- .*llava.*1.5:
41
- instruction_template: 'Vicuna-v1.1'
42
- .*wizard.*mega:
43
- instruction_template: 'Wizard-Mega'
44
- custom_stopping_strings: '"</s>"'
45
- .*starchat-beta:
46
- instruction_template: 'Starchat-Beta'
47
- custom_stopping_strings: '"<|end|>"'
48
  (?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable)(?!.*chinese).*vicuna:
 
49
  instruction_template: 'Vicuna-v0'
50
  .*vicuna.*v0:
 
51
  instruction_template: 'Vicuna-v0'
52
  .*vicuna.*(1.1|1_1|1.3|1_3):
 
53
  instruction_template: 'Vicuna-v1.1'
54
- .*vicuna.*(1.5|1_5):
 
55
  instruction_template: 'Vicuna-v1.1'
56
  .*stable.*vicuna:
 
57
  instruction_template: 'StableVicuna'
58
  (?!.*chat).*chinese-vicuna:
 
59
  instruction_template: 'Alpaca'
60
  .*chinese-vicuna.*chat:
 
61
  instruction_template: 'Chinese-Vicuna-Chat'
62
  .*alpaca:
 
 
 
 
 
 
 
 
 
 
 
63
  instruction_template: 'Alpaca'
 
 
64
  .*koala:
 
65
  instruction_template: 'Koala'
66
  .*chatglm:
 
67
  instruction_template: 'ChatGLM'
68
- .*(metharme|pygmalion|mythalion):
 
69
  instruction_template: 'Metharme'
 
 
 
 
 
70
  .*raven:
 
71
  instruction_template: 'RWKV-Raven'
 
 
72
  .*moss-moon.*sft:
 
73
  instruction_template: 'MOSS'
74
  .*stablelm-tuned:
 
75
  instruction_template: 'StableLM'
 
 
 
 
 
 
 
76
  .*galactica.*finetuned:
 
77
  instruction_template: 'Galactica Finetuned'
78
  .*galactica.*-v2:
 
79
  instruction_template: 'Galactica v2'
80
  (?!.*finetuned)(?!.*-v2).*galactica:
 
81
  instruction_template: 'Galactica'
82
  .*guanaco:
 
83
  instruction_template: 'Guanaco non-chat'
84
  .*baize:
 
85
  instruction_template: 'Baize'
86
  .*mpt-.*instruct:
 
87
  instruction_template: 'Alpaca'
88
  .*mpt-.*chat:
89
- instruction_template: 'ChatML'
 
90
  (?!.*-flan-)(?!.*-t5-).*lamini-:
 
91
  instruction_template: 'Alpaca'
92
  .*incite.*chat:
 
93
  instruction_template: 'INCITE-Chat'
94
  .*incite.*instruct:
 
95
  instruction_template: 'INCITE-Instruct'
 
 
 
 
96
  .*ziya-:
 
97
  instruction_template: 'Ziya'
98
  .*koalpaca:
 
99
  instruction_template: 'KoAlpaca'
100
  .*openbuddy:
 
101
  instruction_template: 'OpenBuddy'
102
  (?!.*chat).*vigogne:
 
103
  instruction_template: 'Vigogne-Instruct'
104
  .*vigogne.*chat:
 
105
  instruction_template: 'Vigogne-Chat'
106
  .*(llama-deus|supercot|llama-natural-instructions|open-llama-0.3t-7b-instruct-dolly-hhrlhf|open-llama-0.3t-7b-open-instruct):
 
107
  instruction_template: 'Alpaca'
108
  .*bactrian:
 
109
  instruction_template: 'Bactrian'
110
  .*(h2ogpt-oig-|h2ogpt-oasst1-|h2ogpt-research-oasst1-):
111
- instruction_template: 'INCITE-Chat'
 
112
  .*h2ogpt-gm-:
 
113
  instruction_template: 'H2O-prompt_answer'
114
  .*manticore:
 
115
  instruction_template: 'Manticore Chat'
116
  .*bluemoonrp-(30|13)b:
 
117
  instruction_template: 'Bluemoon'
 
118
  .*Nous-Hermes-13b:
 
119
  instruction_template: 'Alpaca'
120
  .*airoboros:
 
121
  instruction_template: 'Vicuna-v1.1'
122
  .*airoboros.*1.2:
 
123
  instruction_template: 'Airoboros-v1.2'
 
 
 
 
 
 
124
  .*alpa(cino|sta):
 
125
  instruction_template: 'Alpaca'
126
  .*hippogriff:
 
127
  instruction_template: 'Hippogriff'
 
 
 
128
  .*lazarus:
 
129
  instruction_template: 'Alpaca'
130
  .*guanaco-.*(7|13|33|65)b:
131
- instruction_template: 'Vicuna-v0'
 
132
  .*hypermantis:
 
133
  instruction_template: 'Alpaca'
134
  .*open-llama-.*-open-instruct:
 
135
  instruction_template: 'Alpaca'
136
  .*starcoder-gpteacher-code-instruct:
 
137
  instruction_template: 'Alpaca'
138
  .*tulu:
 
139
  instruction_template: 'Tulu'
140
  .*chronos:
 
141
  instruction_template: 'Alpaca'
142
  .*samantha:
 
143
  instruction_template: 'Samantha'
144
  .*wizardcoder:
 
145
  instruction_template: 'Alpaca'
 
 
 
146
  .*minotaur:
147
- instruction_template: 'Manticore Chat'
 
 
 
148
  .*orca_mini:
 
149
  instruction_template: 'Orca Mini'
 
 
 
 
 
 
 
150
  .*(platypus|gplatty|superplatty):
 
151
  instruction_template: 'Alpaca'
152
- .*(openorca-platypus2):
153
- instruction_template: 'OpenOrca-Platypus2'
154
- custom_stopping_strings: '"### Instruction:", "### Response:"'
155
  .*longchat:
 
156
  instruction_template: 'Vicuna-v1.1'
157
  .*vicuna-33b:
 
158
  instruction_template: 'Vicuna-v1.1'
159
  .*redmond-hermes-coder:
 
160
  instruction_template: 'Alpaca'
 
161
  .*wizardcoder-15b:
 
162
  instruction_template: 'Alpaca'
163
- .*wizardlm:
 
 
164
  instruction_template: 'Vicuna-v1.1'
165
  .*godzilla:
 
166
  instruction_template: 'Alpaca'
167
- .*llama(-?)(2|v2).*chat:
168
- instruction_template: 'Llama-v2'
169
- .*newhope:
170
- instruction_template: 'NewHope'
171
- .*stablebeluga2:
172
- instruction_template: 'StableBeluga2'
173
- .*openchat:
174
- instruction_template: 'OpenChat'
175
- .*codellama.*instruct:
176
- instruction_template: 'Llama-v2'
177
- .*(mistral|mixtral).*instruct:
178
- instruction_template: 'Mistral'
179
- .*mistral.*openorca:
180
- instruction_template: 'ChatML'
181
- .*(WizardCoder-Python-34B-V1.0|Phind-CodeLlama-34B-v2|CodeBooga-34B-v0.1):
182
- instruction_template: 'Alpaca'
183
- .*orca-2-(13|7)b:
184
- instruction_template: 'ChatML'
185
- .*openhermes.*mistral:
186
- instruction_template: 'ChatML'
187
- .*Yi-34B-Chat:
188
- instruction_template: 'ChatML'
189
- (dolphin).*:
190
- instruction_template: 'ChatML'
191
- .*synthia:
192
- instruction_template: 'Synthia'
193
- .*(hercules|hyperion):
194
- instruction_template: 'ChatML'
195
- .*command-r:
196
- instruction_template: 'Command-R'
197
- .*xwin-lm-70b-v0.1:
198
- instruction_template: 'Vicuna-v1.1'
199
- .*platypus-yi-34b:
200
- instruction_template: 'Vicuna-v1.1'
201
- .*CausalLM-RP-34B:
202
- instruction_template: 'ChatML'
203
- 34b-beta:
204
- instruction_template: 'ChatML'
205
- .*airoboros-3_1-yi-34b-200k:
206
  instruction_template: 'Llama-v2'
 
1
+ .*(llama|alpac|vicuna|guanaco|koala|llava|wizardlm|metharme|pygmalion-7b|wizard-mega|openbuddy|vigogne|h2ogpt-research|manticore):
2
  model_type: 'llama'
3
  .*(opt-|opt_|opt1|opt3|optfor|galactica|galpaca|pygmalion-350m):
4
  model_type: 'opt'
5
  .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1):
6
  model_type: 'gptj'
7
  .*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm):
8
+ model_type: 'gpt_neox'
9
+ .*llama:
10
+ model_type: 'llama'
11
  .*bloom:
12
  model_type: 'bloom'
13
+ llama-65b-gptq-3bit:
14
+ groupsize: 'None'
15
+ .*(4bit|int4):
16
+ wbits: 4
17
+ .*(3bit|int3):
18
+ wbits: 3
19
+ .*(-2bit|_2bit|int2-):
20
+ wbits: 2
21
+ .*(-1bit|_1bit|int1-):
22
+ wbits: 1
23
+ .*(8bit|int8):
24
+ wbits: 8
25
+ .*(-7bit|_7bit|int7-):
26
+ wbits: 7
27
+ .*(-6bit|_6bit|int6-):
28
+ wbits: 6
29
+ .*(-5bit|_5bit|int5-):
30
+ wbits: 5
31
+ .*(-gr32-|-32g-|groupsize32|-32g$):
32
+ groupsize: 32
33
+ .*(-gr64-|-64g-|groupsize64|-64g$):
34
+ groupsize: 64
35
+ .*(gr128|128g|groupsize128):
36
+ groupsize: 128
37
+ .*(gr1024|1024g|groupsize1024):
38
+ groupsize: 1024
39
  .*(oasst|openassistant-|stablelm-7b-sft-v7-epoch-3):
40
+ mode: 'instruct'
41
  instruction_template: 'Open Assistant'
42
  skip_special_tokens: false
43
  (?!.*galactica)(?!.*reward).*openassistant:
44
+ mode: 'instruct'
45
  instruction_template: 'Open Assistant'
46
  skip_special_tokens: false
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
  (?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable)(?!.*chinese).*vicuna:
48
+ mode: 'instruct'
49
  instruction_template: 'Vicuna-v0'
50
  .*vicuna.*v0:
51
+ mode: 'instruct'
52
  instruction_template: 'Vicuna-v0'
53
  .*vicuna.*(1.1|1_1|1.3|1_3):
54
+ mode: 'instruct'
55
  instruction_template: 'Vicuna-v1.1'
56
+ .*wizard.*vicuna:
57
+ mode: 'instruct'
58
  instruction_template: 'Vicuna-v1.1'
59
  .*stable.*vicuna:
60
+ mode: 'instruct'
61
  instruction_template: 'StableVicuna'
62
  (?!.*chat).*chinese-vicuna:
63
+ mode: 'instruct'
64
  instruction_template: 'Alpaca'
65
  .*chinese-vicuna.*chat:
66
+ mode: 'instruct'
67
  instruction_template: 'Chinese-Vicuna-Chat'
68
  .*alpaca:
69
+ mode: 'instruct'
70
+ instruction_template: 'Alpaca'
71
+ .*alpaca-native-4bit:
72
+ mode: 'instruct'
73
+ instruction_template: 'Alpaca'
74
+ wbits: 4
75
+ groupsize: 128
76
+ .*galactica:
77
+ skip_special_tokens: false
78
+ .*dolly-v[0-9]-[0-9]*b:
79
+ mode: 'instruct'
80
  instruction_template: 'Alpaca'
81
+ skip_special_tokens: false
82
+ custom_stopping_strings: '"### End"'
83
  .*koala:
84
+ mode: 'instruct'
85
  instruction_template: 'Koala'
86
  .*chatglm:
87
+ mode: 'instruct'
88
  instruction_template: 'ChatGLM'
89
+ .*metharme:
90
+ mode: 'instruct'
91
  instruction_template: 'Metharme'
92
+ .*llava:
93
+ mode: 'instruct'
94
+ model_type: 'llama'
95
+ instruction_template: 'LLaVA'
96
+ custom_stopping_strings: '"\n###"'
97
  .*raven:
98
+ mode: 'instruct'
99
  instruction_template: 'RWKV-Raven'
100
+ .*ctx8192:
101
+ truncation_length: 8192
102
  .*moss-moon.*sft:
103
+ mode: 'instruct'
104
  instruction_template: 'MOSS'
105
  .*stablelm-tuned:
106
+ mode: 'instruct'
107
  instruction_template: 'StableLM'
108
+ truncation_length: 4096
109
+ .*stablelm-base:
110
+ truncation_length: 4096
111
+ .*wizardlm:
112
+ mode: 'instruct'
113
+ model_type: 'llama'
114
+ instruction_template: 'WizardLM'
115
  .*galactica.*finetuned:
116
+ mode: 'instruct'
117
  instruction_template: 'Galactica Finetuned'
118
  .*galactica.*-v2:
119
+ mode: 'instruct'
120
  instruction_template: 'Galactica v2'
121
  (?!.*finetuned)(?!.*-v2).*galactica:
122
+ mode: 'instruct'
123
  instruction_template: 'Galactica'
124
  .*guanaco:
125
+ mode: 'instruct'
126
  instruction_template: 'Guanaco non-chat'
127
  .*baize:
128
+ mode: 'instruct'
129
  instruction_template: 'Baize'
130
  .*mpt-.*instruct:
131
+ mode: 'instruct'
132
  instruction_template: 'Alpaca'
133
  .*mpt-.*chat:
134
+ mode: 'instruct'
135
+ instruction_template: 'MPT-Chat'
136
  (?!.*-flan-)(?!.*-t5-).*lamini-:
137
+ mode: 'instruct'
138
  instruction_template: 'Alpaca'
139
  .*incite.*chat:
140
+ mode: 'instruct'
141
  instruction_template: 'INCITE-Chat'
142
  .*incite.*instruct:
143
+ mode: 'instruct'
144
  instruction_template: 'INCITE-Instruct'
145
+ .*wizard.*mega:
146
+ mode: 'instruct'
147
+ instruction_template: 'Wizard-Mega'
148
+ custom_stopping_strings: '"</s>"'
149
  .*ziya-:
150
+ mode: 'instruct'
151
  instruction_template: 'Ziya'
152
  .*koalpaca:
153
+ mode: 'instruct'
154
  instruction_template: 'KoAlpaca'
155
  .*openbuddy:
156
+ mode: 'instruct'
157
  instruction_template: 'OpenBuddy'
158
  (?!.*chat).*vigogne:
159
+ mode: 'instruct'
160
  instruction_template: 'Vigogne-Instruct'
161
  .*vigogne.*chat:
162
+ mode: 'instruct'
163
  instruction_template: 'Vigogne-Chat'
164
  .*(llama-deus|supercot|llama-natural-instructions|open-llama-0.3t-7b-instruct-dolly-hhrlhf|open-llama-0.3t-7b-open-instruct):
165
+ mode: 'instruct'
166
  instruction_template: 'Alpaca'
167
  .*bactrian:
168
+ mode: 'instruct'
169
  instruction_template: 'Bactrian'
170
  .*(h2ogpt-oig-|h2ogpt-oasst1-|h2ogpt-research-oasst1-):
171
+ mode: 'instruct'
172
+ instruction_template: 'H2O-human_bot'
173
  .*h2ogpt-gm-:
174
+ mode: 'instruct'
175
  instruction_template: 'H2O-prompt_answer'
176
  .*manticore:
177
+ mode: 'instruct'
178
  instruction_template: 'Manticore Chat'
179
  .*bluemoonrp-(30|13)b:
180
+ mode: 'instruct'
181
  instruction_template: 'Bluemoon'
182
+ truncation_length: 4096
183
  .*Nous-Hermes-13b:
184
+ mode: 'instruct'
185
  instruction_template: 'Alpaca'
186
  .*airoboros:
187
+ mode: 'instruct'
188
  instruction_template: 'Vicuna-v1.1'
189
  .*airoboros.*1.2:
190
+ mode: 'instruct'
191
  instruction_template: 'Airoboros-v1.2'
192
+ .*WizardLM-30B-V1.0:
193
+ mode: 'instruct'
194
+ instruction_template: 'Vicuna-v1.1'
195
+ TheBloke_WizardLM-30B-GPTQ:
196
+ mode: 'instruct'
197
+ instruction_template: 'Vicuna-v1.1'
198
  .*alpa(cino|sta):
199
+ mode: 'instruct'
200
  instruction_template: 'Alpaca'
201
  .*hippogriff:
202
+ mode: 'instruct'
203
  instruction_template: 'Hippogriff'
204
+ .*gpt4all-.*-snoozy:
205
+ mode: 'instruct'
206
+ instruction_template: 'WizardLM'
207
  .*lazarus:
208
+ mode: 'instruct'
209
  instruction_template: 'Alpaca'
210
  .*guanaco-.*(7|13|33|65)b:
211
+ mode: 'instruct'
212
+ instruction_template: 'Guanaco'
213
  .*hypermantis:
214
+ mode: 'instruct'
215
  instruction_template: 'Alpaca'
216
  .*open-llama-.*-open-instruct:
217
+ mode: 'instruct'
218
  instruction_template: 'Alpaca'
219
  .*starcoder-gpteacher-code-instruct:
220
+ mode: 'instruct'
221
  instruction_template: 'Alpaca'
222
  .*tulu:
223
+ mode: 'instruct'
224
  instruction_template: 'Tulu'
225
  .*chronos:
226
+ mode: 'instruct'
227
  instruction_template: 'Alpaca'
228
  .*samantha:
229
+ mode: 'instruct'
230
  instruction_template: 'Samantha'
231
  .*wizardcoder:
232
+ mode: 'instruct'
233
  instruction_template: 'Alpaca'
234
+ .*starchat-beta:
235
+ mode: 'instruct'
236
+ instruction_template: 'Starchat-Beta'
237
  .*minotaur:
238
+ mode: 'instruct'
239
+ instruction_template: 'Minotaur'
240
+ .*minotaur-15b:
241
+ truncation_length: 8192
242
  .*orca_mini:
243
+ mode: 'instruct'
244
  instruction_template: 'Orca Mini'
245
+ .*landmark:
246
+ truncation_length: 8192
247
+ .*superhot-8k:
248
+ truncation_length: 8192
249
+ .*xgen.*-inst:
250
+ truncation_length: 8192
251
+ instruction_template: 'Vicuna-v0'
252
  .*(platypus|gplatty|superplatty):
253
+ mode: 'instruct'
254
  instruction_template: 'Alpaca'
 
 
 
255
  .*longchat:
256
+ mode: 'instruct'
257
  instruction_template: 'Vicuna-v1.1'
258
  .*vicuna-33b:
259
+ mode: 'instruct'
260
  instruction_template: 'Vicuna-v1.1'
261
  .*redmond-hermes-coder:
262
+ mode: 'instruct'
263
  instruction_template: 'Alpaca'
264
+ truncation_length: 8192
265
  .*wizardcoder-15b:
266
+ mode: 'instruct'
267
  instruction_template: 'Alpaca'
268
+ truncation_length: 8192
269
+ .*wizardlm-.*-v1.1:
270
+ mode: 'instruct'
271
  instruction_template: 'Vicuna-v1.1'
272
  .*godzilla:
273
+ mode: 'instruct'
274
  instruction_template: 'Alpaca'
275
+ .*llama-(2|v2):
276
+ truncation_length: 4096
277
+ .*llama-(2|v2).*chat:
278
+ mode: 'instruct'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
279
  instruction_template: 'Llama-v2'
models/orca-mini-3b.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e84705205f71dd55be7b24a778f248f0eda9999a125d313358c087e092d83148
3
+ size 1928446208