narugo1992 commited on
Commit
48722d1
·
verified ·
1 Parent(s): 4039b01

Export model 'vit_medium_patch16_reg4_gap_256.sbb_in1k', on 2025-01-20 08:26:16 UTC

Browse files
README.md CHANGED
@@ -76,7 +76,6 @@ base_model:
76
  - timm/test_resnet.r160_in1k
77
  - timm/test_vit2.r160_in1k
78
  - timm/test_vit.r160_in1k
79
- - timm/tiny_vit_21m_384.dist_in22k_ft_in1k
80
  - timm/vit_base_patch14_reg4_dinov2.lvd142m
81
  - timm/vit_base_patch16_224.orig_in21k
82
  - timm/vit_base_patch16_rope_reg1_gap_256.sbb_in1k
@@ -94,6 +93,7 @@ base_model:
94
  - timm/vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k
95
  - timm/vit_little_patch16_reg4_gap_256.sbb_in1k
96
  - timm/vit_medium_patch16_reg1_gap_256.sbb_in1k
 
97
  - timm/vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k
98
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k
99
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k
@@ -114,7 +114,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
114
 
115
  # Models
116
 
117
- 575 models exported from TIMM in total.
118
 
119
  ## Beit
120
 
@@ -963,7 +963,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
963
 
964
  ## VisionTransformer
965
 
966
- 73 models with model class `VisionTransformer`.
967
 
968
  | Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
969
  |:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:------------------|:------------------|:---------------------------------|:-------------|
@@ -1011,6 +1011,7 @@ ONNX export version from [TIMM](https://huggingface.co/timm).
1011
  | [vit_small_patch16_384.augreg_in21k_ft_in1k](https://huggingface.co/timm/vit_small_patch16_384.augreg_in21k_ft_in1k) | 22.0M | 12.4G | 384 | True | 384 | 1000 | imagenet-1k | VisionTransformer | vit_small_patch16_384 | 2022-12-22 |
1012
  | [vit_small_patch16_384.augreg_in1k](https://huggingface.co/timm/vit_small_patch16_384.augreg_in1k) | 22.0M | 12.4G | 384 | True | 384 | 1000 | imagenet-1k | VisionTransformer | vit_small_patch16_384 | 2022-12-22 |
1013
  | [vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k) | 38.7M | 9.9G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_reg4_gap_256 | 2024-05-20 |
 
1014
  | [vit_medium_patch16_reg1_gap_256.sbb_in1k](https://huggingface.co/timm/vit_medium_patch16_reg1_gap_256.sbb_in1k) | 38.7M | 9.8G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_reg1_gap_256 | 2024-05-10 |
1015
  | [vit_medium_patch16_gap_256.sw_in12k_ft_in1k](https://huggingface.co/timm/vit_medium_patch16_gap_256.sw_in12k_ft_in1k) | 38.7M | 9.8G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_gap_256 | 2022-12-02 |
1016
  | [deit3_medium_patch16_224.fb_in1k](https://huggingface.co/timm/deit3_medium_patch16_224.fb_in1k) | 38.7M | 7.5G | 224 | True | 512 | 1000 | imagenet-1k | VisionTransformer | deit3_medium_patch16_224 | 2023-03-28 |
 
76
  - timm/test_resnet.r160_in1k
77
  - timm/test_vit2.r160_in1k
78
  - timm/test_vit.r160_in1k
 
79
  - timm/vit_base_patch14_reg4_dinov2.lvd142m
80
  - timm/vit_base_patch16_224.orig_in21k
81
  - timm/vit_base_patch16_rope_reg1_gap_256.sbb_in1k
 
93
  - timm/vit_little_patch16_reg1_gap_256.sbb_in12k_ft_in1k
94
  - timm/vit_little_patch16_reg4_gap_256.sbb_in1k
95
  - timm/vit_medium_patch16_reg1_gap_256.sbb_in1k
96
+ - timm/vit_medium_patch16_reg4_gap_256.sbb_in1k
97
  - timm/vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k
98
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k
99
  - timm/vit_mediumd_patch16_reg4_gap_256.sbb2_e200_in12k_ft_in1k
 
114
 
115
  # Models
116
 
117
+ 576 models exported from TIMM in total.
118
 
119
  ## Beit
120
 
 
963
 
964
  ## VisionTransformer
965
 
966
+ 74 models with model class `VisionTransformer`.
967
 
968
  | Name | Params | Flops | Input Size | Can Classify | Features | Classes | Dataset | Model | Architecture | Created At |
969
  |:-------------------------------------------------------------------------------------------------------------------------------------------------|:---------|:--------|-------------:|:---------------|-----------:|----------:|:------------------|:------------------|:---------------------------------|:-------------|
 
1011
  | [vit_small_patch16_384.augreg_in21k_ft_in1k](https://huggingface.co/timm/vit_small_patch16_384.augreg_in21k_ft_in1k) | 22.0M | 12.4G | 384 | True | 384 | 1000 | imagenet-1k | VisionTransformer | vit_small_patch16_384 | 2022-12-22 |
1012
  | [vit_small_patch16_384.augreg_in1k](https://huggingface.co/timm/vit_small_patch16_384.augreg_in1k) | 22.0M | 12.4G | 384 | True | 384 | 1000 | imagenet-1k | VisionTransformer | vit_small_patch16_384 | 2022-12-22 |
1013
  | [vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k](https://huggingface.co/timm/vit_medium_patch16_reg4_gap_256.sbb_in12k_ft_in1k) | 38.7M | 9.9G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_reg4_gap_256 | 2024-05-20 |
1014
+ | [vit_medium_patch16_reg4_gap_256.sbb_in1k](https://huggingface.co/timm/vit_medium_patch16_reg4_gap_256.sbb_in1k) | 38.7M | 9.9G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_reg4_gap_256 | 2024-05-10 |
1015
  | [vit_medium_patch16_reg1_gap_256.sbb_in1k](https://huggingface.co/timm/vit_medium_patch16_reg1_gap_256.sbb_in1k) | 38.7M | 9.8G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_reg1_gap_256 | 2024-05-10 |
1016
  | [vit_medium_patch16_gap_256.sw_in12k_ft_in1k](https://huggingface.co/timm/vit_medium_patch16_gap_256.sw_in12k_ft_in1k) | 38.7M | 9.8G | 256 | True | 512 | 1000 | imagenet-1k | VisionTransformer | vit_medium_patch16_gap_256 | 2022-12-02 |
1017
  | [deit3_medium_patch16_224.fb_in1k](https://huggingface.co/timm/deit3_medium_patch16_224.fb_in1k) | 38.7M | 7.5G | 224 | True | 512 | 1000 | imagenet-1k | VisionTransformer | deit3_medium_patch16_224 | 2023-03-28 |
models.parquet CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d3c15adc4232455e9416adf13db878e29e14930ceecae4002a979c06ad88d5dc
3
- size 39114
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8f8cb82c49d8c8b4221963b7111d9af92bfece6b2a93e7b416ae4a1efe692ec8
3
+ size 39127
vit_medium_patch16_reg4_gap_256.sbb_in1k/meta.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1720c5f18c9ed2d3104efefaa24127f5694faaafbfb31c8a09d4903a208d8ba3
3
+ size 169865
vit_medium_patch16_reg4_gap_256.sbb_in1k/model.onnx ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ccfdfc25373d8963a053f932ccd3b408a58151c18fd3a51d71c33c8702c75750
3
+ size 155706889
vit_medium_patch16_reg4_gap_256.sbb_in1k/preprocess.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f4c3ceed99ae49aa023c8c7cbc224be5ba023f409004ac0091862f1e398d784e
3
+ size 642