Kinsleykinsley commited on
Commit
2a5764b
·
verified ·
1 Parent(s): ea7d115

Upload folder using huggingface_hub

Browse files
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:391873f7bd3f2fd90e1b32839b1ff568cf27ea009815c8ba2d3c4c1c376bcd0d
3
  size 604510743
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0fe561ece2a3a4d02534788071ac40015db9fafd7a50da77c570580c34798aeb
3
  size 604510743
special_tokens_map.json CHANGED
@@ -1,36 +1,24 @@
1
  {
2
  "additional_special_tokens": [
3
- {
4
- "content": "<SMILES>",
5
- "lstrip": false,
6
- "normalized": false,
7
- "rstrip": false,
8
- "single_word": false
9
- },
10
- {
11
- "content": "</SMILES>",
12
- "lstrip": false,
13
- "normalized": false,
14
- "rstrip": false,
15
- "single_word": false
16
- },
17
- {
18
- "content": "<SEQ>",
19
- "lstrip": false,
20
- "normalized": false,
21
- "rstrip": false,
22
- "single_word": false
23
- },
24
- {
25
- "content": "</SEQ>",
26
- "lstrip": false,
27
- "normalized": false,
28
- "rstrip": false,
29
- "single_word": false
30
- }
31
  ],
32
- "bos_token": "<|endoftext|>",
33
- "eos_token": "<|endoftext|>",
 
 
 
 
 
 
 
 
 
 
 
 
34
  "pad_token": {
35
  "content": "[PAD]",
36
  "lstrip": false,
@@ -38,5 +26,11 @@
38
  "rstrip": false,
39
  "single_word": false
40
  },
41
- "unk_token": "<|endoftext|>"
 
 
 
 
 
 
42
  }
 
1
  {
2
  "additional_special_tokens": [
3
+ "<SMILES>",
4
+ "</SMILES>",
5
+ "<SEQ>",
6
+ "</SEQ>"
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
7
  ],
8
+ "bos_token": {
9
+ "content": "<|endoftext|>",
10
+ "lstrip": false,
11
+ "normalized": false,
12
+ "rstrip": false,
13
+ "single_word": false
14
+ },
15
+ "eos_token": {
16
+ "content": "<|endoftext|>",
17
+ "lstrip": false,
18
+ "normalized": false,
19
+ "rstrip": false,
20
+ "single_word": false
21
+ },
22
  "pad_token": {
23
  "content": "[PAD]",
24
  "lstrip": false,
 
26
  "rstrip": false,
27
  "single_word": false
28
  },
29
+ "unk_token": {
30
+ "content": "<|endoftext|>",
31
+ "lstrip": false,
32
+ "normalized": false,
33
+ "rstrip": false,
34
+ "single_word": false
35
+ }
36
  }
tokenizer_config.json CHANGED
@@ -84,8 +84,15 @@
84
  "clean_up_tokenization_spaces": false,
85
  "eos_token": "<|endoftext|>",
86
  "extra_special_tokens": {},
 
87
  "model_max_length": 1000000000000000019884624838656,
 
88
  "pad_token": "[PAD]",
 
 
 
89
  "tokenizer_class": "GPT2Tokenizer",
 
 
90
  "unk_token": "<|endoftext|>"
91
  }
 
84
  "clean_up_tokenization_spaces": false,
85
  "eos_token": "<|endoftext|>",
86
  "extra_special_tokens": {},
87
+ "max_length": 512,
88
  "model_max_length": 1000000000000000019884624838656,
89
+ "pad_to_multiple_of": null,
90
  "pad_token": "[PAD]",
91
+ "pad_token_type_id": 0,
92
+ "padding_side": "right",
93
+ "stride": 0,
94
  "tokenizer_class": "GPT2Tokenizer",
95
+ "truncation_side": "right",
96
+ "truncation_strategy": "longest_first",
97
  "unk_token": "<|endoftext|>"
98
  }