NoQuest commited on
Commit
eb52d2d
·
1 Parent(s): bf36b13

Suppr orca

Browse files
models/model/model_metadata.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": 0,
3
+ "modelDefinition": {
4
+ "key": "hf@mistralai/Mistral-7B-Instruct-v0.2"
5
+ },
6
+ "libraryName": "transformers",
7
+ "sizeInBytes": 15010174631,
8
+ "pipelineName": "text-generation",
9
+ "tags": [
10
+ "transformers",
11
+ "pytorch",
12
+ "safetensors",
13
+ "mistral",
14
+ "text-generation",
15
+ "finetuned",
16
+ "conversational",
17
+ "arxiv:2310.06825",
18
+ "license:apache-2.0",
19
+ "autotrain_compatible",
20
+ "text-generation-inference",
21
+ "endpoints_compatible",
22
+ "region:us"
23
+ ],
24
+ "taggedLanguages": [],
25
+ "url": "https://huggingface.co/mistralai%2FMistral-7B-Instruct-v0.2/tree/main",
26
+ "lastModified": "2024-07-03T22:10:26.000Z",
27
+ "commitHash": "250544c9a802b0396550d0fd24bc80ff98bb1f5f",
28
+ "downloadedBy": "admin",
29
+ "downloadDate": 1721504950401,
30
+ "lastDssUsage": 0,
31
+ "manuallyImported": false
32
+ }
models/model_metadata.json ADDED
@@ -0,0 +1,32 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "version": 0,
3
+ "modelDefinition": {
4
+ "key": "hf@mistralai/Mistral-7B-Instruct-v0.2"
5
+ },
6
+ "libraryName": "transformers",
7
+ "sizeInBytes": 15010174631,
8
+ "pipelineName": "text-generation",
9
+ "tags": [
10
+ "transformers",
11
+ "pytorch",
12
+ "safetensors",
13
+ "mistral",
14
+ "text-generation",
15
+ "finetuned",
16
+ "conversational",
17
+ "arxiv:2310.06825",
18
+ "license:apache-2.0",
19
+ "autotrain_compatible",
20
+ "text-generation-inference",
21
+ "endpoints_compatible",
22
+ "region:us"
23
+ ],
24
+ "taggedLanguages": [],
25
+ "url": "https://huggingface.co/mistralai%2FMistral-7B-Instruct-v0.2/tree/main",
26
+ "lastModified": "2024-07-03T22:10:26.000Z",
27
+ "commitHash": "250544c9a802b0396550d0fd24bc80ff98bb1f5f",
28
+ "downloadedBy": "admin",
29
+ "downloadDate": 1721504950401,
30
+ "lastDssUsage": 0,
31
+ "manuallyImported": false
32
+ }
models/orca/adapter_config.json DELETED
@@ -1,20 +0,0 @@
1
- {
2
- "base_model_name_or_path": "google/flan-t5-large",
3
- "bias": "none",
4
- "fan_in_fan_out": false,
5
- "inference_mode": true,
6
- "init_lora_weights": true,
7
- "layers_pattern": null,
8
- "layers_to_transform": null,
9
- "lora_alpha": 32,
10
- "lora_dropout": 0.1,
11
- "modules_to_save": null,
12
- "peft_type": "LORA",
13
- "r": 16,
14
- "revision": null,
15
- "target_modules": [
16
- "q",
17
- "v"
18
- ],
19
- "task_type": "SEQ_2_SEQ_LM"
20
- }
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/orca/adapter_model.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:3a069c6ee7ea30b036e198ece2e1491b04287941668496dd23c6120494fe8676
3
- size 18980429
 
 
 
 
models/orca/config.yaml DELETED
@@ -1,279 +0,0 @@
1
- .*(llama|alpac|vicuna|guanaco|koala|llava|wizardlm|metharme|pygmalion-7b|wizard-mega|openbuddy|vigogne|h2ogpt-research|manticore):
2
- model_type: 'llama'
3
- .*(opt-|opt_|opt1|opt3|optfor|galactica|galpaca|pygmalion-350m):
4
- model_type: 'opt'
5
- .*(gpt-j|gptj|gpt4all-j|malion-6b|pygway|pygmalion-6b|dolly-v1):
6
- model_type: 'gptj'
7
- .*(gpt-neox|koalpaca-polyglot|polyglot.*koalpaca|polyglot-ko|polyglot_ko|pythia|stablelm|incite|dolly-v2|polycoder|h2ogpt-oig|h2ogpt-oasst1|h2ogpt-gm):
8
- model_type: 'gpt_neox'
9
- .*llama:
10
- model_type: 'llama'
11
- .*bloom:
12
- model_type: 'bloom'
13
- llama-65b-gptq-3bit:
14
- groupsize: 'None'
15
- .*(4bit|int4):
16
- wbits: 4
17
- .*(3bit|int3):
18
- wbits: 3
19
- .*(-2bit|_2bit|int2-):
20
- wbits: 2
21
- .*(-1bit|_1bit|int1-):
22
- wbits: 1
23
- .*(8bit|int8):
24
- wbits: 8
25
- .*(-7bit|_7bit|int7-):
26
- wbits: 7
27
- .*(-6bit|_6bit|int6-):
28
- wbits: 6
29
- .*(-5bit|_5bit|int5-):
30
- wbits: 5
31
- .*(-gr32-|-32g-|groupsize32|-32g$):
32
- groupsize: 32
33
- .*(-gr64-|-64g-|groupsize64|-64g$):
34
- groupsize: 64
35
- .*(gr128|128g|groupsize128):
36
- groupsize: 128
37
- .*(gr1024|1024g|groupsize1024):
38
- groupsize: 1024
39
- .*(oasst|openassistant-|stablelm-7b-sft-v7-epoch-3):
40
- mode: 'instruct'
41
- instruction_template: 'Open Assistant'
42
- skip_special_tokens: false
43
- (?!.*galactica)(?!.*reward).*openassistant:
44
- mode: 'instruct'
45
- instruction_template: 'Open Assistant'
46
- skip_special_tokens: false
47
- (?!.*v0)(?!.*1.1)(?!.*1_1)(?!.*stable)(?!.*chinese).*vicuna:
48
- mode: 'instruct'
49
- instruction_template: 'Vicuna-v0'
50
- .*vicuna.*v0:
51
- mode: 'instruct'
52
- instruction_template: 'Vicuna-v0'
53
- .*vicuna.*(1.1|1_1|1.3|1_3):
54
- mode: 'instruct'
55
- instruction_template: 'Vicuna-v1.1'
56
- .*wizard.*vicuna:
57
- mode: 'instruct'
58
- instruction_template: 'Vicuna-v1.1'
59
- .*stable.*vicuna:
60
- mode: 'instruct'
61
- instruction_template: 'StableVicuna'
62
- (?!.*chat).*chinese-vicuna:
63
- mode: 'instruct'
64
- instruction_template: 'Alpaca'
65
- .*chinese-vicuna.*chat:
66
- mode: 'instruct'
67
- instruction_template: 'Chinese-Vicuna-Chat'
68
- .*alpaca:
69
- mode: 'instruct'
70
- instruction_template: 'Alpaca'
71
- .*alpaca-native-4bit:
72
- mode: 'instruct'
73
- instruction_template: 'Alpaca'
74
- wbits: 4
75
- groupsize: 128
76
- .*galactica:
77
- skip_special_tokens: false
78
- .*dolly-v[0-9]-[0-9]*b:
79
- mode: 'instruct'
80
- instruction_template: 'Alpaca'
81
- skip_special_tokens: false
82
- custom_stopping_strings: '"### End"'
83
- .*koala:
84
- mode: 'instruct'
85
- instruction_template: 'Koala'
86
- .*chatglm:
87
- mode: 'instruct'
88
- instruction_template: 'ChatGLM'
89
- .*metharme:
90
- mode: 'instruct'
91
- instruction_template: 'Metharme'
92
- .*llava:
93
- mode: 'instruct'
94
- model_type: 'llama'
95
- instruction_template: 'LLaVA'
96
- custom_stopping_strings: '"\n###"'
97
- .*raven:
98
- mode: 'instruct'
99
- instruction_template: 'RWKV-Raven'
100
- .*ctx8192:
101
- truncation_length: 8192
102
- .*moss-moon.*sft:
103
- mode: 'instruct'
104
- instruction_template: 'MOSS'
105
- .*stablelm-tuned:
106
- mode: 'instruct'
107
- instruction_template: 'StableLM'
108
- truncation_length: 4096
109
- .*stablelm-base:
110
- truncation_length: 4096
111
- .*wizardlm:
112
- mode: 'instruct'
113
- model_type: 'llama'
114
- instruction_template: 'WizardLM'
115
- .*galactica.*finetuned:
116
- mode: 'instruct'
117
- instruction_template: 'Galactica Finetuned'
118
- .*galactica.*-v2:
119
- mode: 'instruct'
120
- instruction_template: 'Galactica v2'
121
- (?!.*finetuned)(?!.*-v2).*galactica:
122
- mode: 'instruct'
123
- instruction_template: 'Galactica'
124
- .*guanaco:
125
- mode: 'instruct'
126
- instruction_template: 'Guanaco non-chat'
127
- .*baize:
128
- mode: 'instruct'
129
- instruction_template: 'Baize'
130
- .*mpt-.*instruct:
131
- mode: 'instruct'
132
- instruction_template: 'Alpaca'
133
- .*mpt-.*chat:
134
- mode: 'instruct'
135
- instruction_template: 'MPT-Chat'
136
- (?!.*-flan-)(?!.*-t5-).*lamini-:
137
- mode: 'instruct'
138
- instruction_template: 'Alpaca'
139
- .*incite.*chat:
140
- mode: 'instruct'
141
- instruction_template: 'INCITE-Chat'
142
- .*incite.*instruct:
143
- mode: 'instruct'
144
- instruction_template: 'INCITE-Instruct'
145
- .*wizard.*mega:
146
- mode: 'instruct'
147
- instruction_template: 'Wizard-Mega'
148
- custom_stopping_strings: '"</s>"'
149
- .*ziya-:
150
- mode: 'instruct'
151
- instruction_template: 'Ziya'
152
- .*koalpaca:
153
- mode: 'instruct'
154
- instruction_template: 'KoAlpaca'
155
- .*openbuddy:
156
- mode: 'instruct'
157
- instruction_template: 'OpenBuddy'
158
- (?!.*chat).*vigogne:
159
- mode: 'instruct'
160
- instruction_template: 'Vigogne-Instruct'
161
- .*vigogne.*chat:
162
- mode: 'instruct'
163
- instruction_template: 'Vigogne-Chat'
164
- .*(llama-deus|supercot|llama-natural-instructions|open-llama-0.3t-7b-instruct-dolly-hhrlhf|open-llama-0.3t-7b-open-instruct):
165
- mode: 'instruct'
166
- instruction_template: 'Alpaca'
167
- .*bactrian:
168
- mode: 'instruct'
169
- instruction_template: 'Bactrian'
170
- .*(h2ogpt-oig-|h2ogpt-oasst1-|h2ogpt-research-oasst1-):
171
- mode: 'instruct'
172
- instruction_template: 'H2O-human_bot'
173
- .*h2ogpt-gm-:
174
- mode: 'instruct'
175
- instruction_template: 'H2O-prompt_answer'
176
- .*manticore:
177
- mode: 'instruct'
178
- instruction_template: 'Manticore Chat'
179
- .*bluemoonrp-(30|13)b:
180
- mode: 'instruct'
181
- instruction_template: 'Bluemoon'
182
- truncation_length: 4096
183
- .*Nous-Hermes-13b:
184
- mode: 'instruct'
185
- instruction_template: 'Alpaca'
186
- .*airoboros:
187
- mode: 'instruct'
188
- instruction_template: 'Vicuna-v1.1'
189
- .*airoboros.*1.2:
190
- mode: 'instruct'
191
- instruction_template: 'Airoboros-v1.2'
192
- .*WizardLM-30B-V1.0:
193
- mode: 'instruct'
194
- instruction_template: 'Vicuna-v1.1'
195
- TheBloke_WizardLM-30B-GPTQ:
196
- mode: 'instruct'
197
- instruction_template: 'Vicuna-v1.1'
198
- .*alpa(cino|sta):
199
- mode: 'instruct'
200
- instruction_template: 'Alpaca'
201
- .*hippogriff:
202
- mode: 'instruct'
203
- instruction_template: 'Hippogriff'
204
- .*gpt4all-.*-snoozy:
205
- mode: 'instruct'
206
- instruction_template: 'WizardLM'
207
- .*lazarus:
208
- mode: 'instruct'
209
- instruction_template: 'Alpaca'
210
- .*guanaco-.*(7|13|33|65)b:
211
- mode: 'instruct'
212
- instruction_template: 'Guanaco'
213
- .*hypermantis:
214
- mode: 'instruct'
215
- instruction_template: 'Alpaca'
216
- .*open-llama-.*-open-instruct:
217
- mode: 'instruct'
218
- instruction_template: 'Alpaca'
219
- .*starcoder-gpteacher-code-instruct:
220
- mode: 'instruct'
221
- instruction_template: 'Alpaca'
222
- .*tulu:
223
- mode: 'instruct'
224
- instruction_template: 'Tulu'
225
- .*chronos:
226
- mode: 'instruct'
227
- instruction_template: 'Alpaca'
228
- .*samantha:
229
- mode: 'instruct'
230
- instruction_template: 'Samantha'
231
- .*wizardcoder:
232
- mode: 'instruct'
233
- instruction_template: 'Alpaca'
234
- .*starchat-beta:
235
- mode: 'instruct'
236
- instruction_template: 'Starchat-Beta'
237
- .*minotaur:
238
- mode: 'instruct'
239
- instruction_template: 'Minotaur'
240
- .*minotaur-15b:
241
- truncation_length: 8192
242
- .*orca_mini:
243
- mode: 'instruct'
244
- instruction_template: 'Orca Mini'
245
- .*landmark:
246
- truncation_length: 8192
247
- .*superhot-8k:
248
- truncation_length: 8192
249
- .*xgen.*-inst:
250
- truncation_length: 8192
251
- instruction_template: 'Vicuna-v0'
252
- .*(platypus|gplatty|superplatty):
253
- mode: 'instruct'
254
- instruction_template: 'Alpaca'
255
- .*longchat:
256
- mode: 'instruct'
257
- instruction_template: 'Vicuna-v1.1'
258
- .*vicuna-33b:
259
- mode: 'instruct'
260
- instruction_template: 'Vicuna-v1.1'
261
- .*redmond-hermes-coder:
262
- mode: 'instruct'
263
- instruction_template: 'Alpaca'
264
- truncation_length: 8192
265
- .*wizardcoder-15b:
266
- mode: 'instruct'
267
- instruction_template: 'Alpaca'
268
- truncation_length: 8192
269
- .*wizardlm-.*-v1.1:
270
- mode: 'instruct'
271
- instruction_template: 'Vicuna-v1.1'
272
- .*godzilla:
273
- mode: 'instruct'
274
- instruction_template: 'Alpaca'
275
- .*llama-(2|v2):
276
- truncation_length: 4096
277
- .*llama-(2|v2).*chat:
278
- mode: 'instruct'
279
- instruction_template: 'Llama-v2'
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
models/orca/orca-mini-3b.bin DELETED
@@ -1,3 +0,0 @@
1
- version https://git-lfs.github.com/spec/v1
2
- oid sha256:e84705205f71dd55be7b24a778f248f0eda9999a125d313358c087e092d83148
3
- size 1928446208
 
 
 
 
nohup.out DELETED
@@ -1 +0,0 @@
1
- Waiting for Emacs...