bhaviktheslider commited on
Commit
935f0e9
·
verified ·
1 Parent(s): 63c5799

Training in progress, step 450

Browse files
pytorch_model-00001-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:266824cc3d1dc0ff5a818d48fe79a79fa533a90abca92bdbc0b496aac9b281a9
3
  size 4996733620
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4ac95e9dd6742520c26302b9d0057e2ebc40ad31df3e31e73eb9f7e0834556f0
3
  size 4996733620
pytorch_model-00002-of-00002.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d0ef0181d58c23616f8c7b93a96f712384f847e85ac234bf0c5d951c63b42a68
3
  size 1178243330
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:090d295208ef1093b5c85611049a7818a2096ae344c9414d2b9d87f769c8901c
3
  size 1178243330
runs/Feb01_20-35-12_7bf1f721fedf/events.out.tfevents.1738442162.7bf1f721fedf.25154.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b06d4dcdafc20f344254c91cdf6ba0a571bdc5537b72fcd7fc25516aa86dfb8b
3
- size 125686
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5402ca37a7db7c26427ce5a115fcc01f4e57a3aabaeffbab76c8611d77387f40
3
+ size 133450
special_tokens_map.json CHANGED
@@ -21,11 +21,5 @@
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
- "pad_token": {
25
- "content": "<|endoftext|>",
26
- "lstrip": false,
27
- "normalized": false,
28
- "rstrip": false,
29
- "single_word": false
30
- }
31
  }
 
21
  "rstrip": false,
22
  "single_word": false
23
  },
24
+ "pad_token": "<|endoftext|>"
 
 
 
 
 
 
25
  }
tokenizer.json CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:5eee858c5123a4279c3e1f7b81247343f356ac767940b2692a928ad929543214
3
- size 11422063
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9c5ae00e602b8860cbd784ba82a8aa14e8feecec692e7076590d014d7b7fdafa
3
+ size 11421896
tokenizer_config.json CHANGED
@@ -202,7 +202,6 @@
202
  "extra_special_tokens": {},
203
  "model_max_length": 131072,
204
  "pad_token": "<|endoftext|>",
205
- "padding_side": "left",
206
  "split_special_tokens": false,
207
  "tokenizer_class": "Qwen2Tokenizer",
208
  "unk_token": null
 
202
  "extra_special_tokens": {},
203
  "model_max_length": 131072,
204
  "pad_token": "<|endoftext|>",
 
205
  "split_special_tokens": false,
206
  "tokenizer_class": "Qwen2Tokenizer",
207
  "unk_token": null