Rishabh56 commited on
Commit
85ef4c2
·
verified ·
1 Parent(s): ac08d68

Training in progress, step 126

Browse files
adapter_config.json CHANGED
@@ -4,7 +4,7 @@
4
  "base_model_class": "GPTNeoXForCausalLM",
5
  "parent_library": "transformers.models.gpt_neox.modeling_gpt_neox"
6
  },
7
- "base_model_name_or_path": "databricks/dolly-v2-3b",
8
  "bias": "none",
9
  "eva_config": null,
10
  "exclude_modules": null,
 
4
  "base_model_class": "GPTNeoXForCausalLM",
5
  "parent_library": "transformers.models.gpt_neox.modeling_gpt_neox"
6
  },
7
+ "base_model_name_or_path": "OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5",
8
  "bias": "none",
9
  "eva_config": null,
10
  "exclude_modules": null,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:9a4c316dc9cdd75c32ff0a3ec1d00e9492b07fd42d4acaddf6b977665fddf607
3
- size 10494984
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1ec86f9fbcb545a428256ceb8b677a48e2a0c2f4d3b25931c209df3f99b88de
3
+ size 23603456
special_tokens_map.json CHANGED
@@ -1,8 +1,10 @@
1
  {
2
  "additional_special_tokens": [
3
- "### End",
4
- "### Instruction:",
5
- "### Response:"
 
 
6
  ],
7
  "bos_token": {
8
  "content": "<|endoftext|>",
@@ -19,6 +21,13 @@
19
  "single_word": false
20
  },
21
  "pad_token": {
 
 
 
 
 
 
 
22
  "content": "<|endoftext|>",
23
  "lstrip": false,
24
  "normalized": false,
 
1
  {
2
  "additional_special_tokens": [
3
+ "<|system|>",
4
+ "<|assistant|>",
5
+ "<|prefix_begin|>",
6
+ "<|prefix_end|>",
7
+ "<|prompter|>"
8
  ],
9
  "bos_token": {
10
  "content": "<|endoftext|>",
 
21
  "single_word": false
22
  },
23
  "pad_token": {
24
+ "content": "<|padding|>",
25
+ "lstrip": false,
26
+ "normalized": false,
27
+ "rstrip": false,
28
+ "single_word": false
29
+ },
30
+ "sep_token": {
31
  "content": "<|endoftext|>",
32
  "lstrip": false,
33
  "normalized": false,
tokenizer.json CHANGED
@@ -1,7 +1,21 @@
1
  {
2
  "version": "1.0",
3
- "truncation": null,
4
- "padding": null,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
5
  "added_tokens": [
6
  {
7
  "id": 0,
@@ -230,7 +244,7 @@
230
  },
231
  {
232
  "id": 50277,
233
- "content": "### End",
234
  "single_word": false,
235
  "lstrip": false,
236
  "rstrip": false,
@@ -239,7 +253,7 @@
239
  },
240
  {
241
  "id": 50278,
242
- "content": "### Instruction:",
243
  "single_word": false,
244
  "lstrip": false,
245
  "rstrip": false,
@@ -248,7 +262,25 @@
248
  },
249
  {
250
  "id": 50279,
251
- "content": "### Response:",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
252
  "single_word": false,
253
  "lstrip": false,
254
  "rstrip": false,
 
1
  {
2
  "version": "1.0",
3
+ "truncation": {
4
+ "direction": "Right",
5
+ "max_length": 512,
6
+ "strategy": "LongestFirst",
7
+ "stride": 0
8
+ },
9
+ "padding": {
10
+ "strategy": {
11
+ "Fixed": 512
12
+ },
13
+ "direction": "Right",
14
+ "pad_to_multiple_of": null,
15
+ "pad_id": 1,
16
+ "pad_type_id": 0,
17
+ "pad_token": "<|padding|>"
18
+ },
19
  "added_tokens": [
20
  {
21
  "id": 0,
 
244
  },
245
  {
246
  "id": 50277,
247
+ "content": "<|prefix_end|>",
248
  "single_word": false,
249
  "lstrip": false,
250
  "rstrip": false,
 
253
  },
254
  {
255
  "id": 50278,
256
+ "content": "<|prefix_begin|>",
257
  "single_word": false,
258
  "lstrip": false,
259
  "rstrip": false,
 
262
  },
263
  {
264
  "id": 50279,
265
+ "content": "<|prompter|>",
266
+ "single_word": false,
267
+ "lstrip": false,
268
+ "rstrip": false,
269
+ "normalized": false,
270
+ "special": true
271
+ },
272
+ {
273
+ "id": 50280,
274
+ "content": "<|system|>",
275
+ "single_word": false,
276
+ "lstrip": false,
277
+ "rstrip": false,
278
+ "normalized": false,
279
+ "special": true
280
+ },
281
+ {
282
+ "id": 50281,
283
+ "content": "<|assistant|>",
284
  "single_word": false,
285
  "lstrip": false,
286
  "rstrip": false,
tokenizer_config.json CHANGED
@@ -204,7 +204,7 @@
204
  "special": false
205
  },
206
  "50277": {
207
- "content": "### End",
208
  "lstrip": false,
209
  "normalized": false,
210
  "rstrip": false,
@@ -212,7 +212,7 @@
212
  "special": true
213
  },
214
  "50278": {
215
- "content": "### Instruction:",
216
  "lstrip": false,
217
  "normalized": false,
218
  "rstrip": false,
@@ -220,7 +220,23 @@
220
  "special": true
221
  },
222
  "50279": {
223
- "content": "### Response:",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
224
  "lstrip": false,
225
  "normalized": false,
226
  "rstrip": false,
@@ -229,16 +245,19 @@
229
  }
230
  },
231
  "additional_special_tokens": [
232
- "### End",
233
- "### Instruction:",
234
- "### Response:"
 
 
235
  ],
236
  "bos_token": "<|endoftext|>",
237
- "clean_up_tokenization_spaces": false,
238
  "eos_token": "<|endoftext|>",
239
  "extra_special_tokens": {},
240
  "model_max_length": 1000000000000000019884624838656,
241
- "pad_token": "<|endoftext|>",
 
242
  "tokenizer_class": "GPTNeoXTokenizer",
243
  "unk_token": "<|endoftext|>"
244
  }
 
204
  "special": false
205
  },
206
  "50277": {
207
+ "content": "<|prefix_end|>",
208
  "lstrip": false,
209
  "normalized": false,
210
  "rstrip": false,
 
212
  "special": true
213
  },
214
  "50278": {
215
+ "content": "<|prefix_begin|>",
216
  "lstrip": false,
217
  "normalized": false,
218
  "rstrip": false,
 
220
  "special": true
221
  },
222
  "50279": {
223
+ "content": "<|prompter|>",
224
+ "lstrip": false,
225
+ "normalized": false,
226
+ "rstrip": false,
227
+ "single_word": false,
228
+ "special": true
229
+ },
230
+ "50280": {
231
+ "content": "<|system|>",
232
+ "lstrip": false,
233
+ "normalized": false,
234
+ "rstrip": false,
235
+ "single_word": false,
236
+ "special": true
237
+ },
238
+ "50281": {
239
+ "content": "<|assistant|>",
240
  "lstrip": false,
241
  "normalized": false,
242
  "rstrip": false,
 
245
  }
246
  },
247
  "additional_special_tokens": [
248
+ "<|system|>",
249
+ "<|assistant|>",
250
+ "<|prefix_begin|>",
251
+ "<|prefix_end|>",
252
+ "<|prompter|>"
253
  ],
254
  "bos_token": "<|endoftext|>",
255
+ "clean_up_tokenization_spaces": true,
256
  "eos_token": "<|endoftext|>",
257
  "extra_special_tokens": {},
258
  "model_max_length": 1000000000000000019884624838656,
259
+ "pad_token": "<|padding|>",
260
+ "sep_token": "<|endoftext|>",
261
  "tokenizer_class": "GPTNeoXTokenizer",
262
  "unk_token": "<|endoftext|>"
263
  }
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ab309e2baded7ed3fde47e810fc7a42c14faec1dcf2f80ca8fc30fd855adfab9
3
  size 5496
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:231481c050bd17627c2d07ca8e01a64183cead3d867fb3b758abc0dd0fcccd61
3
  size 5496