speedrun / config.json
lhallee's picture
Upload PLM
5393461 verified
{
"add_att_soft_cap": false,
"architectures": [
"PLM"
],
"attention_soft_cap": 2.0,
"expansion_ratio": 2.6667,
"hidden_size": 768,
"mlm": true,
"num_att_tokens": 512,
"num_attention_heads": 6,
"num_hidden_layers": 12,
"p_attention": false,
"sliding_window_size": 2048,
"soft_logit_cap": 32.0,
"tie_embeddings": false,
"torch_dtype": "bfloat16",
"transformers_version": "4.53.0",
"unet": true,
"vocab_size": 33
}