Upload tokenizer
Browse files- special_tokens_map.json +7 -0
- tokenizer_config.json +20 -0
- vocab.txt +0 -0
    	
        special_tokens_map.json
    ADDED
    
    | @@ -0,0 +1,7 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "cls_token": "[CLS]",
         | 
| 3 | 
            +
              "mask_token": "[MASK]",
         | 
| 4 | 
            +
              "pad_token": "[PAD]",
         | 
| 5 | 
            +
              "sep_token": "[SEP]",
         | 
| 6 | 
            +
              "unk_token": "[UNK]"
         | 
| 7 | 
            +
            }
         | 
    	
        tokenizer_config.json
    ADDED
    
    | @@ -0,0 +1,20 @@ | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | |
|  | 
|  | |
| 1 | 
            +
            {
         | 
| 2 | 
            +
              "cls_token": "[CLS]",
         | 
| 3 | 
            +
              "do_lower_case": false,
         | 
| 4 | 
            +
              "do_subword_tokenize": true,
         | 
| 5 | 
            +
              "do_word_tokenize": true,
         | 
| 6 | 
            +
              "jumanpp_kwargs": null,
         | 
| 7 | 
            +
              "mask_token": "[MASK]",
         | 
| 8 | 
            +
              "mecab_kwargs": null,
         | 
| 9 | 
            +
              "model_max_length": 1000000000000000019884624838656,
         | 
| 10 | 
            +
              "name_or_path": "bandainamco-mirai/distilbert-base-japanese",
         | 
| 11 | 
            +
              "never_split": null,
         | 
| 12 | 
            +
              "pad_token": "[PAD]",
         | 
| 13 | 
            +
              "sep_token": "[SEP]",
         | 
| 14 | 
            +
              "special_tokens_map_file": null,
         | 
| 15 | 
            +
              "subword_tokenizer_type": "wordpiece",
         | 
| 16 | 
            +
              "sudachi_kwargs": null,
         | 
| 17 | 
            +
              "tokenizer_class": "BertJapaneseTokenizer",
         | 
| 18 | 
            +
              "unk_token": "[UNK]",
         | 
| 19 | 
            +
              "word_tokenizer_type": "basic"
         | 
| 20 | 
            +
            }
         | 
    	
        vocab.txt
    ADDED
    
    | The diff for this file is too large to render. 
		See raw diff | 
|  | 
