toy-multistep-v3-test / tokenizer.json
cfpark00's picture
Upload tokenizer
eefe2c7 verified
raw
history blame contribute delete
903 Bytes
{
"version": "1.0",
"truncation": null,
"padding": null,
"added_tokens": [
{
"id": 2,
"content": "<pad>",
"single_word": false,
"lstrip": false,
"rstrip": false,
"normalized": false,
"special": true
}
],
"normalizer": null,
"pre_tokenizer": {
"type": "Whitespace"
},
"post_processor": null,
"decoder": null,
"model": {
"type": "WordLevel",
"vocab": {
"<unk>": 0,
"<bos>": 1,
"<pad>": 2,
":": 3,
"S": 4,
"P": 5,
"C": 6,
"A": 7,
"0": 8,
"1": 9,
"2": 10,
"3": 11,
"4": 12,
"5": 13,
"6": 14,
"7": 15,
"8": 16,
"9": 17,
"10": 18,
"11": 19,
"12": 20,
"13": 21,
"14": 22,
"15": 23,
"16": 24,
"17": 25,
"18": 26,
"19": 27
},
"unk_token": "<unk>"
}
}