Training in progress, step 126
Browse files- adapter_config.json +1 -1
- adapter_model.safetensors +2 -2
- special_tokens_map.json +12 -3
- tokenizer.json +37 -5
- tokenizer_config.json +27 -8
- training_args.bin +1 -1
adapter_config.json
CHANGED
@@ -4,7 +4,7 @@
|
|
4 |
"base_model_class": "GPTNeoXForCausalLM",
|
5 |
"parent_library": "transformers.models.gpt_neox.modeling_gpt_neox"
|
6 |
},
|
7 |
-
"base_model_name_or_path": "
|
8 |
"bias": "none",
|
9 |
"eva_config": null,
|
10 |
"exclude_modules": null,
|
|
|
4 |
"base_model_class": "GPTNeoXForCausalLM",
|
5 |
"parent_library": "transformers.models.gpt_neox.modeling_gpt_neox"
|
6 |
},
|
7 |
+
"base_model_name_or_path": "OpenAssistant/oasst-sft-4-pythia-12b-epoch-3.5",
|
8 |
"bias": "none",
|
9 |
"eva_config": null,
|
10 |
"exclude_modules": null,
|
adapter_model.safetensors
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c1ec86f9fbcb545a428256ceb8b677a48e2a0c2f4d3b25931c209df3f99b88de
|
3 |
+
size 23603456
|
special_tokens_map.json
CHANGED
@@ -1,8 +1,10 @@
|
|
1 |
{
|
2 |
"additional_special_tokens": [
|
3 |
-
"
|
4 |
-
"
|
5 |
-
"
|
|
|
|
|
6 |
],
|
7 |
"bos_token": {
|
8 |
"content": "<|endoftext|>",
|
@@ -19,6 +21,13 @@
|
|
19 |
"single_word": false
|
20 |
},
|
21 |
"pad_token": {
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
22 |
"content": "<|endoftext|>",
|
23 |
"lstrip": false,
|
24 |
"normalized": false,
|
|
|
1 |
{
|
2 |
"additional_special_tokens": [
|
3 |
+
"<|system|>",
|
4 |
+
"<|assistant|>",
|
5 |
+
"<|prefix_begin|>",
|
6 |
+
"<|prefix_end|>",
|
7 |
+
"<|prompter|>"
|
8 |
],
|
9 |
"bos_token": {
|
10 |
"content": "<|endoftext|>",
|
|
|
21 |
"single_word": false
|
22 |
},
|
23 |
"pad_token": {
|
24 |
+
"content": "<|padding|>",
|
25 |
+
"lstrip": false,
|
26 |
+
"normalized": false,
|
27 |
+
"rstrip": false,
|
28 |
+
"single_word": false
|
29 |
+
},
|
30 |
+
"sep_token": {
|
31 |
"content": "<|endoftext|>",
|
32 |
"lstrip": false,
|
33 |
"normalized": false,
|
tokenizer.json
CHANGED
@@ -1,7 +1,21 @@
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
-
"truncation":
|
4 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
5 |
"added_tokens": [
|
6 |
{
|
7 |
"id": 0,
|
@@ -230,7 +244,7 @@
|
|
230 |
},
|
231 |
{
|
232 |
"id": 50277,
|
233 |
-
"content": "
|
234 |
"single_word": false,
|
235 |
"lstrip": false,
|
236 |
"rstrip": false,
|
@@ -239,7 +253,7 @@
|
|
239 |
},
|
240 |
{
|
241 |
"id": 50278,
|
242 |
-
"content": "
|
243 |
"single_word": false,
|
244 |
"lstrip": false,
|
245 |
"rstrip": false,
|
@@ -248,7 +262,25 @@
|
|
248 |
},
|
249 |
{
|
250 |
"id": 50279,
|
251 |
-
"content": "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
252 |
"single_word": false,
|
253 |
"lstrip": false,
|
254 |
"rstrip": false,
|
|
|
1 |
{
|
2 |
"version": "1.0",
|
3 |
+
"truncation": {
|
4 |
+
"direction": "Right",
|
5 |
+
"max_length": 512,
|
6 |
+
"strategy": "LongestFirst",
|
7 |
+
"stride": 0
|
8 |
+
},
|
9 |
+
"padding": {
|
10 |
+
"strategy": {
|
11 |
+
"Fixed": 512
|
12 |
+
},
|
13 |
+
"direction": "Right",
|
14 |
+
"pad_to_multiple_of": null,
|
15 |
+
"pad_id": 1,
|
16 |
+
"pad_type_id": 0,
|
17 |
+
"pad_token": "<|padding|>"
|
18 |
+
},
|
19 |
"added_tokens": [
|
20 |
{
|
21 |
"id": 0,
|
|
|
244 |
},
|
245 |
{
|
246 |
"id": 50277,
|
247 |
+
"content": "<|prefix_end|>",
|
248 |
"single_word": false,
|
249 |
"lstrip": false,
|
250 |
"rstrip": false,
|
|
|
253 |
},
|
254 |
{
|
255 |
"id": 50278,
|
256 |
+
"content": "<|prefix_begin|>",
|
257 |
"single_word": false,
|
258 |
"lstrip": false,
|
259 |
"rstrip": false,
|
|
|
262 |
},
|
263 |
{
|
264 |
"id": 50279,
|
265 |
+
"content": "<|prompter|>",
|
266 |
+
"single_word": false,
|
267 |
+
"lstrip": false,
|
268 |
+
"rstrip": false,
|
269 |
+
"normalized": false,
|
270 |
+
"special": true
|
271 |
+
},
|
272 |
+
{
|
273 |
+
"id": 50280,
|
274 |
+
"content": "<|system|>",
|
275 |
+
"single_word": false,
|
276 |
+
"lstrip": false,
|
277 |
+
"rstrip": false,
|
278 |
+
"normalized": false,
|
279 |
+
"special": true
|
280 |
+
},
|
281 |
+
{
|
282 |
+
"id": 50281,
|
283 |
+
"content": "<|assistant|>",
|
284 |
"single_word": false,
|
285 |
"lstrip": false,
|
286 |
"rstrip": false,
|
tokenizer_config.json
CHANGED
@@ -204,7 +204,7 @@
|
|
204 |
"special": false
|
205 |
},
|
206 |
"50277": {
|
207 |
-
"content": "
|
208 |
"lstrip": false,
|
209 |
"normalized": false,
|
210 |
"rstrip": false,
|
@@ -212,7 +212,7 @@
|
|
212 |
"special": true
|
213 |
},
|
214 |
"50278": {
|
215 |
-
"content": "
|
216 |
"lstrip": false,
|
217 |
"normalized": false,
|
218 |
"rstrip": false,
|
@@ -220,7 +220,23 @@
|
|
220 |
"special": true
|
221 |
},
|
222 |
"50279": {
|
223 |
-
"content": "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
224 |
"lstrip": false,
|
225 |
"normalized": false,
|
226 |
"rstrip": false,
|
@@ -229,16 +245,19 @@
|
|
229 |
}
|
230 |
},
|
231 |
"additional_special_tokens": [
|
232 |
-
"
|
233 |
-
"
|
234 |
-
"
|
|
|
|
|
235 |
],
|
236 |
"bos_token": "<|endoftext|>",
|
237 |
-
"clean_up_tokenization_spaces":
|
238 |
"eos_token": "<|endoftext|>",
|
239 |
"extra_special_tokens": {},
|
240 |
"model_max_length": 1000000000000000019884624838656,
|
241 |
-
"pad_token": "<|
|
|
|
242 |
"tokenizer_class": "GPTNeoXTokenizer",
|
243 |
"unk_token": "<|endoftext|>"
|
244 |
}
|
|
|
204 |
"special": false
|
205 |
},
|
206 |
"50277": {
|
207 |
+
"content": "<|prefix_end|>",
|
208 |
"lstrip": false,
|
209 |
"normalized": false,
|
210 |
"rstrip": false,
|
|
|
212 |
"special": true
|
213 |
},
|
214 |
"50278": {
|
215 |
+
"content": "<|prefix_begin|>",
|
216 |
"lstrip": false,
|
217 |
"normalized": false,
|
218 |
"rstrip": false,
|
|
|
220 |
"special": true
|
221 |
},
|
222 |
"50279": {
|
223 |
+
"content": "<|prompter|>",
|
224 |
+
"lstrip": false,
|
225 |
+
"normalized": false,
|
226 |
+
"rstrip": false,
|
227 |
+
"single_word": false,
|
228 |
+
"special": true
|
229 |
+
},
|
230 |
+
"50280": {
|
231 |
+
"content": "<|system|>",
|
232 |
+
"lstrip": false,
|
233 |
+
"normalized": false,
|
234 |
+
"rstrip": false,
|
235 |
+
"single_word": false,
|
236 |
+
"special": true
|
237 |
+
},
|
238 |
+
"50281": {
|
239 |
+
"content": "<|assistant|>",
|
240 |
"lstrip": false,
|
241 |
"normalized": false,
|
242 |
"rstrip": false,
|
|
|
245 |
}
|
246 |
},
|
247 |
"additional_special_tokens": [
|
248 |
+
"<|system|>",
|
249 |
+
"<|assistant|>",
|
250 |
+
"<|prefix_begin|>",
|
251 |
+
"<|prefix_end|>",
|
252 |
+
"<|prompter|>"
|
253 |
],
|
254 |
"bos_token": "<|endoftext|>",
|
255 |
+
"clean_up_tokenization_spaces": true,
|
256 |
"eos_token": "<|endoftext|>",
|
257 |
"extra_special_tokens": {},
|
258 |
"model_max_length": 1000000000000000019884624838656,
|
259 |
+
"pad_token": "<|padding|>",
|
260 |
+
"sep_token": "<|endoftext|>",
|
261 |
"tokenizer_class": "GPTNeoXTokenizer",
|
262 |
"unk_token": "<|endoftext|>"
|
263 |
}
|
training_args.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5496
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:231481c050bd17627c2d07ca8e01a64183cead3d867fb3b758abc0dd0fcccd61
|
3 |
size 5496
|