{ | |
"add_att_soft_cap": false, | |
"architectures": [ | |
"PLM" | |
], | |
"attention_soft_cap": 2.0, | |
"expansion_ratio": 2.6667, | |
"hidden_size": 768, | |
"mlm": true, | |
"num_att_tokens": 512, | |
"num_attention_heads": 6, | |
"num_hidden_layers": 12, | |
"p_attention": false, | |
"sliding_window_size": 2048, | |
"soft_logit_cap": 32.0, | |
"tie_embeddings": false, | |
"torch_dtype": "bfloat16", | |
"transformers_version": "4.53.0", | |
"unet": true, | |
"vocab_size": 33 | |
} | |