|
{ |
|
"apply_layernorm": true, |
|
"architectures": [ |
|
"Dinov2Model" |
|
], |
|
"attention_probs_dropout_prob": 0.0, |
|
"block_chunks": 4, |
|
"drop_path_rate": 0.0, |
|
"hidden_act": "gelu", |
|
"hidden_dropout_prob": 0.0, |
|
"hidden_size": 1536, |
|
"image_size": 224, |
|
"initializer_range": 0.02, |
|
"layer_norm_eps": 1e-06, |
|
"layerscale_value": 1.0, |
|
"mlp_ratio": 4, |
|
"model_type": "dinov2", |
|
"num_attention_heads": 24, |
|
"num_channels": 3, |
|
"num_hidden_layers": 40, |
|
"num_layers_per_block": 10, |
|
"out_features": [ |
|
"stage40" |
|
], |
|
"out_indices": [ |
|
40 |
|
], |
|
"patch_size": 14, |
|
"qkv_bias": true, |
|
"reshape_hidden_states": true, |
|
"stage_names": [ |
|
"stem", |
|
"stage1", |
|
"stage2", |
|
"stage3", |
|
"stage4", |
|
"stage5", |
|
"stage6", |
|
"stage7", |
|
"stage8", |
|
"stage9", |
|
"stage10", |
|
"stage11", |
|
"stage12", |
|
"stage13", |
|
"stage14", |
|
"stage15", |
|
"stage16", |
|
"stage17", |
|
"stage18", |
|
"stage19", |
|
"stage20", |
|
"stage21", |
|
"stage22", |
|
"stage23", |
|
"stage24", |
|
"stage25", |
|
"stage26", |
|
"stage27", |
|
"stage28", |
|
"stage29", |
|
"stage30", |
|
"stage31", |
|
"stage32", |
|
"stage33", |
|
"stage34", |
|
"stage35", |
|
"stage36", |
|
"stage37", |
|
"stage38", |
|
"stage39", |
|
"stage40" |
|
], |
|
"torch_dtype": "float32", |
|
"transformers_version": "4.46.2", |
|
"use_swiglu_ffn": true |
|
} |
|
|