metadata
library_name: litert
base_model: timm/deit_tiny_patch16_224.fb_in1k
tags:
- vision
- image-classification
datasets:
- imagenet-1k
deit_tiny_patch16_224
Converted TIMM image classification model for LiteRT.
- Source architecture: deit_tiny_patch16_224
- File: model.tflite
Model Details
- Model Type: Image classification / feature backbone
- Model Stats:
- Params (M): 5.7
- GMACs: 1.3
- Activations (M): 6.0
- Image size: 224 x 224
- Papers:
- Training data-efficient image transformers & distillation through attention: https://arxiv.org/abs/2012.12877
- Original: https://github.com/facebookresearch/deit
- Dataset: ImageNet-1k
Citation
@InProceedings{pmlr-v139-touvron21a,
title = {Training data-efficient image transformers & distillation through attention},
author = {Touvron, Hugo and Cord, Matthieu and Douze, Matthijs and Massa, Francisco and Sablayrolles, Alexandre and Jegou, Herve},
booktitle = {International Conference on Machine Learning},
pages = {10347--10357},
year = {2021},
volume = {139},
month = {July}
}
@misc{rw2019timm,
author = {Ross Wightman},
title = {PyTorch Image Models},
year = {2019},
publisher = {GitHub},
journal = {GitHub repository},
doi = {10.5281/zenodo.4414861},
howpublished = {\url{https://github.com/huggingface/pytorch-image-models}}
}