narugo1992 commited on
Commit
07c7dec
·
verified ·
1 Parent(s): 9aaae6e

Export model 'vit_base_patch16_siglip_384.webli', on 2025-01-20 05:10:50 UTC

Browse files
README.md CHANGED
@@ -29,7 +29,6 @@ base_model:
29
  - timm/fastvit_s12.apple_dist_in1k
30
  - timm/fastvit_t12.apple_in1k
31
  - timm/ghostnetv2_100.in1k
32
- - timm/hardcorenas_a.miil_green_in1k
33
  - timm/hardcorenas_b.miil_green_in1k
34
  - timm/hardcorenas_f.miil_green_in1k
35
  - timm/hgnet_small.ssld_in1k
@@ -94,6 +93,7 @@ base_model:
94
  - timm/vit_base_patch14_reg4_dinov2.lvd142m
95
  - timm/vit_base_patch16_224.orig_in21k
96
  - timm/vit_base_patch16_rope_reg1_gap_256.sbb_in1k
 
97
  - timm/vit_base_patch16_siglip_gap_256.webli_i18n
98
  - timm/vit_betwixt_patch16_reg1_gap_256.sbb_in1k
99
  - timm/vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k
@@ -114,7 +114,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
114
 
115
  # Models
116
 
117
- 234 models exported from TIMM in total.
118
 
119
  ## Beit
120
 
@@ -636,11 +636,12 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
636
 
637
  ## VisionTransformer
638
 
639
- 27 models with model class `VisionTransformer`.
640
 
641
  | Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
642
  |:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:-------------|:------------------|:---------------------------------|:-------------|
643
  | [vit_base_patch14_reg4_dinov2.lvd142m](https://huggingface.co/timm/vit_base_patch14_reg4_dinov2.lvd142m) | 85.5M | 117.4G | 518 | False | 768 | 768 | | VisionTransformer | vit_base_patch14_reg4_dinov2 | 2023-10-30 |
 
644
  | [vit_base_r50_s16_384.orig_in21k_ft_in1k](https://huggingface.co/timm/vit_base_r50_s16_384.orig_in21k_ft_in1k) | 86.6M | 49.5G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_r50_s16_384 | 2022-12-23 |
645
  | [vit_base_patch16_clip_384.openai_ft_in12k_in1k](https://huggingface.co/timm/vit_base_patch16_clip_384.openai_ft_in12k_in1k) | 86.4M | 49.4G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch16_clip_384 | 2022-11-30 |
646
  | [vit_base_patch16_clip_384.laion2b_ft_in12k_in1k](https://huggingface.co/timm/vit_base_patch16_clip_384.laion2b_ft_in12k_in1k) | 86.4M | 49.4G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch16_clip_384 | 2022-11-11 |
 
29
  - timm/fastvit_s12.apple_dist_in1k
30
  - timm/fastvit_t12.apple_in1k
31
  - timm/ghostnetv2_100.in1k
 
32
  - timm/hardcorenas_b.miil_green_in1k
33
  - timm/hardcorenas_f.miil_green_in1k
34
  - timm/hgnet_small.ssld_in1k
 
93
  - timm/vit_base_patch14_reg4_dinov2.lvd142m
94
  - timm/vit_base_patch16_224.orig_in21k
95
  - timm/vit_base_patch16_rope_reg1_gap_256.sbb_in1k
96
+ - timm/vit_base_patch16_siglip_384.webli
97
  - timm/vit_base_patch16_siglip_gap_256.webli_i18n
98
  - timm/vit_betwixt_patch16_reg1_gap_256.sbb_in1k
99
  - timm/vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k
 
114
 
115
  # Models
116
 
117
+ 235 models exported from TIMM in total.
118
 
119
  ## Beit
120
 
 
636
 
637
  ## VisionTransformer
638
 
639
+ 28 models with model class `VisionTransformer`.
640
 
641
  | Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
642
  |:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:-------------|:------------------|:---------------------------------|:-------------|
643
  | [vit_base_patch14_reg4_dinov2.lvd142m](https://huggingface.co/timm/vit_base_patch14_reg4_dinov2.lvd142m) | 85.5M | 117.4G | 518 | False | 768 | 768 | | VisionTransformer | vit_base_patch14_reg4_dinov2 | 2023-10-30 |
644
+ | [vit_base_patch16_siglip_384.webli](https://huggingface.co/timm/vit_base_patch16_siglip_384.webli) | 92.7M | 50.0G | 384 | False | 768 | 768 | | VisionTransformer | vit_base_patch16_siglip_384 | 2024-12-24 |
645
  | [vit_base_r50_s16_384.orig_in21k_ft_in1k](https://huggingface.co/timm/vit_base_r50_s16_384.orig_in21k_ft_in1k) | 86.6M | 49.5G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_r50_s16_384 | 2022-12-23 |
646
  | [vit_base_patch16_clip_384.openai_ft_in12k_in1k](https://huggingface.co/timm/vit_base_patch16_clip_384.openai_ft_in12k_in1k) | 86.4M | 49.4G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch16_clip_384 | 2022-11-30 |
647
  | [vit_base_patch16_clip_384.laion2b_ft_in12k_in1k](https://huggingface.co/timm/vit_base_patch16_clip_384.laion2b_ft_in12k_in1k) | 86.4M | 49.4G | 384 | True | 768 | 1000 | imagenet-1k | VisionTransformer | vit_base_patch16_clip_384 | 2022-11-11 |
models.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:70c603c7baa4ed19f270bb2b3417c83cbe668f6aa2b5eb4a7dede99754aa8eca
3
- size 22992
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:55979f5cb4537ed0cad7388a2e79952acb594023eb0a964580d9ad77c2377295
3
+ size 23031
vit_base_patch16_siglip_384.webli/meta.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4912c2c26ded1d9d00b42dae9194791f675719f8be81f90763b7796e68f2cc13
3
+ size 484
vit_base_patch16_siglip_384.webli/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:078864b9e92087d6660867bc17faafb0a5ccdb75320c10a96289a12e88adc2bd
3
+ size 372886550
vit_base_patch16_siglip_384.webli/preprocess.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:12eb69f461d904bd37771631fbd78ee5e3f973ce8269097856c99756d57dd898
3
+ size 642