File size: 306 Bytes
d7316c8 2ba4536 d7316c8 |
1 2 3 4 5 6 7 8 9 10 11 12 13 |
{
"_name_or_path": "microsoft/phi-4",
"architectures": ["AutoModelForCausalLM"],
"model_type": "causal_lm",
"num_hidden_layers": 32,
"num_attention_heads": 16,
"num_key_value_heads": 8,
"hidden_size": 4096,
"vocab_size": 50257,
"torch_dtype": "bfloat16",
"tie_word_embeddings": false
}
|