{"tokenizer_class": "GPT2Tokenizer", "model_max_length": 128}