{ | |
"hidden_size": 4096, | |
"num_attention_heads": 32, | |
"num_hidden_layers": 31, | |
"intermediate_size": 14336, | |
"model_type": "llama", | |
"vocab_size": 32000 | |
} | |
{ | |
"hidden_size": 4096, | |
"num_attention_heads": 32, | |
"num_hidden_layers": 31, | |
"intermediate_size": 14336, | |
"model_type": "llama", | |
"vocab_size": 32000 | |
} | |