Edit model card

Model card for phikon-distil-vit-tiny-patch16-224-kather2016

This model is a distilled version of owkin/phikon to a TinyViT on the 1aurent/Kather-texture-2016 dataset.

Model Usage

Image Classification

from transformers import AutoModelForImageClassification, AutoImageProcessor
from urllib.request import urlopen
from PIL import Image

# get example histology image
img = Image.open(
  urlopen(
    "https://datasets-server.huggingface.co/assets/1aurent/Kather-texture-2016/--/default/train/0/image/image.jpg"
  )
)

# load image_processor and model from the hub
model_name = "1aurent/phikon-distil-vit-tiny-patch16-224-kather2016"
image_processor = AutoImageProcessor.from_pretrained(model_name)
model = AutoModelForImageClassification.from_pretrained(model_name)

inputs = image_processor(img, return_tensors="pt")
outputs = model(**inputs)

Citation

@article{Filiot2023.07.21.23292757,
  author       = {Alexandre Filiot and Ridouane Ghermi and Antoine Olivier and Paul Jacob and Lucas Fidon and Alice Mac Kain and Charlie Saillard and Jean-Baptiste Schiratti},
  title        = {Scaling Self-Supervised Learning for Histopathology with Masked Image Modeling},
  elocation-id = {2023.07.21.23292757},
  year         = {2023},
  doi          = {10.1101/2023.07.21.23292757},
  publisher    = {Cold Spring Harbor Laboratory Press},
  url          = {https://www.medrxiv.org/content/early/2023/09/14/2023.07.21.23292757},
  eprint       = {https://www.medrxiv.org/content/early/2023/09/14/2023.07.21.23292757.full.pdf},
  journal      = {medRxiv}
}
Downloads last month
3
Safetensors
Model size
5.53M params
Tensor type
F32
·
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social visibility and check back later, or deploy to Inference Endpoints (dedicated) instead.

Finetuned from

Dataset used to train 1aurent/phikon-distil-vit-tiny-patch16-224-kather2016

Evaluation results