Datasets:
Languages:
English
Multilinguality:
monolingual
Size Categories:
1K<n<10K
Language Creators:
expert-generated
Annotations Creators:
expert-generated
Source Datasets:
original
License:
import json | |
import os | |
from collections import OrderedDict, defaultdict | |
from math import ceil | |
import numpy as np | |
import pandas as pd | |
import datasets | |
logger = datasets.logging.get_logger(__name__) | |
CCAGT_CLASSES = OrderedDict( | |
{ | |
1: "NUCLEUS", | |
2: "CLUSTER", | |
3: "SATELLITE", | |
4: "NUCLEUS_OUT_OF_FOCUS", | |
5: "OVERLAPPED_NUCLEI", | |
6: "NON_VIABLE_NUCLEUS", | |
7: "LEUKOCYTE_NUCLEUS", | |
} | |
) | |
_LICENSE = "CC BY NC 3.0 License" | |
_CITATION = """\ | |
@misc{CCAgTDataset, | |
doi = {10.17632/WG4BPM33HJ.2}, | |
url = {https://data.mendeley.com/datasets/wg4bpm33hj/2}, | |
author = {Jo{\\~{a}}o Gustavo Atkinson Amorim and Andr{\'{e}} Vict{\'{o}}ria Matias and Tainee Bottamedi and Vinícius Sanches and Ane Francyne Costa and Fabiana Botelho De Miranda Onofre and Alexandre Sherlley Casimiro Onofre and Aldo von Wangenheim}, | |
title = {CCAgT: Images of Cervical Cells with AgNOR Stain Technique}, | |
publisher = {Mendeley}, | |
year = {2022}, | |
copyright = {Attribution-NonCommercial 3.0 Unported} | |
} | |
""" | |
_HOMEPAGE = "https://data.mendeley.com/datasets/wg4bpm33hj" | |
_DESCRIPTION = """\ | |
The CCAgT (Images of Cervical Cells with AgNOR Stain Technique) dataset contains 9339 images (1600x1200 resolution where each pixel is 0.111µmX0.111µm) from 15 different slides stained using the AgNOR technique. | |
Each image has at least one label. In total, this dataset has more than 63K instances of annotated object. | |
The images are from the patients of the Gynecology and Colonoscopy Outpatient Clinic of the Polydoro Ernani de São Thiago University Hospital of the Universidade Federal de Santa Catarina (HU-UFSC). | |
""" | |
_DATA_URL = "https://md-datasets-cache-zipfiles-prod.s3.eu-west-1.amazonaws.com/wg4bpm33hj-2.zip" | |
def tvt(ids, tvt_size, seed=1609): | |
"""From a list of indexes/ids (int) will generate the train-validation-test data. | |
Based on `github.com/scikit-learn/scikit-learn/blob/37ac6788c9504ee409b75e5e24ff7d86c90c2ffb/sklearn/model_selection/_split.py#L2321` | |
""" | |
n_samples = len(ids) | |
qtd = { | |
"valid": ceil(n_samples * tvt_size[1]), | |
"test": ceil(n_samples * tvt_size[2]), | |
} | |
qtd["train"] = int(n_samples - qtd["valid"] - qtd["test"]) | |
rng = np.random.RandomState(seed) | |
permutatation = rng.permutation(ids) | |
out = { | |
"train": set(permutatation[: qtd["train"]]), | |
"valid": set(permutatation[qtd["train"] : qtd["train"] + qtd["valid"]]), | |
"test": set(permutatation[qtd["train"] + qtd["valid"] :]), | |
} | |
return out["train"], out["valid"], out["test"] | |
def annotations_per_image(df): | |
""" | |
based on: https://github.com/johnnv1/CCAgT-utils/blob/54ade78e4ddb2e2ed9507b8a1633940897767cac/CCAgT_utils/describe.py#L152 | |
""" | |
df_describe_images = df.groupby(["image_id", "category_id"]).size().reset_index().rename(columns={0: "count"}) | |
df_describe_images = df_describe_images.pivot(columns=["category_id"], index="image_id") | |
df_describe_images = df_describe_images.rename(CCAGT_CLASSES, axis=1) | |
df_describe_images["qtd_annotations"] = df_describe_images.sum(axis=1) | |
df_describe_images = df_describe_images.fillna(0) | |
df_describe_images["NORs"] = ( | |
df_describe_images[ | |
"count", | |
CCAGT_CLASSES[2], | |
] | |
+ df_describe_images[ | |
"count", | |
CCAGT_CLASSES[3], | |
] | |
) | |
return df_describe_images | |
def tvt_by_nors(df, tvt_size=(0.7, 0.15, 0.15), **kwargs): | |
"""This will split the CCAgT annotations based on the number of NORs | |
into each image. With a silly separation, first will split | |
between each fold images with one or less NORs, after will split | |
images with the amount of NORs is between 2 and 7, and at least will | |
split images that have more than 7 NORs. | |
based on `https://github.com/johnnv1/CCAgT-utils/blob/54ade78e4ddb2e2ed9507b8a1633940897767cac/CCAgT_utils/split.py#L64` | |
""" | |
if sum(tvt_size) != 1: | |
raise ValueError("The sum of `tvt_size` need to be equal to 1!") | |
df_describe_imgs = annotations_per_image(df) | |
img_ids = {} | |
img_ids["low_nors"] = df_describe_imgs.loc[(df_describe_imgs["NORs"] < 2)].index | |
img_ids["medium_nors"] = df_describe_imgs[(df_describe_imgs["NORs"] >= 2) * (df_describe_imgs["NORs"] <= 7)].index | |
img_ids["high_nors"] = df_describe_imgs[(df_describe_imgs["NORs"] > 7)].index | |
train_ids = set({}) | |
valid_ids = set({}) | |
test_ids = set({}) | |
for k, ids in img_ids.items(): | |
logger.info(f"Splitting {len(ids)} images with {k} quantity...") | |
if len(ids) == 0: | |
continue | |
_train, _valid, _test = tvt(ids, tvt_size, **kwargs) | |
logger.info(f">T: {len(_train)} V: {len(_valid)} T: {len(_test)}") | |
train_ids = train_ids.union(_train) | |
valid_ids = valid_ids.union(_valid) | |
test_ids = test_ids.union(_test) | |
return train_ids, valid_ids, test_ids | |
def get_basename(path): | |
return os.path.splitext(os.path.basename(path))[0] | |
def get_slide_id(path): | |
bn = get_basename(path) | |
slide_id = bn.split("_")[0] | |
return slide_id | |
class CCAgTConfig(datasets.BuilderConfig): | |
"""BuilderConfig for CCAgT.""" | |
seed = 1609 | |
tvt_size = (0.7, 0.15, 0.15) | |
class CCAgT(datasets.GeneratorBasedBuilder): | |
"""Images of Cervical Cells with AgNOR Stain Technique (CCAgT) dataset""" | |
test_dummy_data = False | |
VERSION = datasets.Version("2.0.0") | |
BUILDER_CONFIG_CLASS = CCAgTConfig | |
BUILDER_CONFIGS = [ | |
CCAgTConfig(name="semantic_segmentation", version=VERSION, description="The semantic segmentation variant."), | |
CCAgTConfig(name="object_detection", version=VERSION, description="The object detection variant."), | |
CCAgTConfig(name="instance_segmentation", version=VERSION, description="The instance segmentation variant."), | |
] | |
DEFAULT_CONFIG_NAME = "semantic_segmentation" | |
def _info(self): | |
assert len(CCAGT_CLASSES) == 7 | |
if self.config.name == "semantic_segmentation": | |
features = datasets.Features( | |
{ | |
"image": datasets.Image(), | |
"annotation": datasets.Image(), | |
} | |
) | |
elif self.config.name == "object_detection": | |
features = datasets.Features( | |
{ | |
"image": datasets.Image(), | |
"objects": datasets.Sequence( | |
{ | |
"bbox": datasets.Sequence(datasets.Value("float32"), length=4), | |
"label": datasets.ClassLabel(names=list(CCAGT_CLASSES.values())), | |
} | |
), | |
} | |
) | |
elif self.config.name == "instance_segmentation": | |
features = datasets.Features( | |
{ | |
"image": datasets.Image(), | |
"objects": datasets.Sequence( | |
{ | |
"bbox": datasets.Sequence(datasets.Value("float32"), length=4), | |
"segment": datasets.Sequence(datasets.Sequence(datasets.Value("float32"))), | |
"label": datasets.ClassLabel(names=list(CCAGT_CLASSES.values())), | |
} | |
), | |
} | |
) | |
else: | |
raise NotImplementedError | |
return datasets.DatasetInfo( | |
description=_DESCRIPTION, | |
features=features, | |
homepage=_HOMEPAGE, | |
license=_LICENSE, | |
citation=_CITATION, | |
) | |
def _download_and_extract_all(self, dl_manager): | |
def extracted_by_slide(paths): | |
return {get_slide_id(path): dl_manager.extract(path) for path in paths} | |
data_dir = dl_manager.download_and_extract(_DATA_URL) | |
base_path = os.path.join(data_dir, "wg4bpm33hj-2") | |
logger.info("Extracting images...") | |
self.images_base_dir = os.path.join(base_path, "images") | |
images_to_extract = [ | |
os.path.join(self.images_base_dir, fn) for fn in os.listdir(self.images_base_dir) if fn.endswith(".zip") | |
] | |
self.images_extracted = extracted_by_slide(images_to_extract) | |
if self.config.name == "semantic_segmentation": | |
logger.info("Extracting masks...") | |
self.masks_base_dir = os.path.join(base_path, "masks") | |
masks_to_extract = [ | |
os.path.join(self.masks_base_dir, fn) for fn in os.listdir(self.masks_base_dir) if fn.endswith(".zip") | |
] | |
self.masks_extracted = extracted_by_slide(masks_to_extract) | |
elif self.config.name in {"object_detection", "instance_segmentation"}: | |
logger.info("Reading COCO OD file...") | |
ccagt_OD_COCO_path = os.path.join(base_path, "CCAgT_COCO_OD.json") | |
with open(ccagt_OD_COCO_path, "r", encoding="utf-8") as json_file: | |
coco_OD = json.load(json_file) | |
self._imageid_to_coco_OD_annotations = defaultdict(list) | |
for labels in coco_OD["annotations"]: | |
self._imageid_to_coco_OD_annotations[labels["image_id"]].append(labels) | |
logger.info("Loading dataset info...") | |
ccagt_raw_path = os.path.join(base_path, "CCAgT.parquet.gzip") | |
with open(ccagt_raw_path, "rb") as f: | |
self._ccagt_info = pd.read_parquet(f, columns=["image_name", "category_id", "image_id", "slide_id"]) | |
self._bn_to_imageid = pd.Series( | |
self._ccagt_info["image_id"].values, index=self._ccagt_info["image_name"] | |
).to_dict() | |
def _split_generators(self, dl_manager): | |
"""Returns SplitGenerators.""" | |
def build_path(basename, tp="images"): | |
slide = basename.split("_")[0] | |
if tp == "images": | |
dir_path = self.images_extracted[slide] | |
ext = ".jpg" | |
else: | |
dir_path = self.masks_extracted[slide] | |
ext = ".png" | |
return os.path.join(dir_path, slide, basename + ext) | |
def images_and_masks(basenames): | |
for bn in basenames: | |
yield build_path(bn), build_path(bn, "masks") | |
def images_and_boxes(basenames): | |
for bn in basenames: | |
image_id = self._bn_to_imageid[bn] | |
labels = [ | |
{"bbox": annotation["bbox"], "label": annotation["category_id"] - 1} | |
for annotation in self._imageid_to_coco_OD_annotations[image_id] | |
] | |
yield build_path(bn), labels | |
def images_and_instances(basenames): | |
for bn in basenames: | |
image_id = self._bn_to_imageid[bn] | |
instances = [ | |
{ | |
"bbox": annotation["bbox"], | |
"label": annotation["category_id"] - 1, | |
"segment": annotation["segmentation"], | |
} | |
for annotation in self._imageid_to_coco_OD_annotations[image_id] | |
] | |
yield build_path(bn), instances | |
self._download_and_extract_all(dl_manager) | |
logger.info("Splitting dataset based on the NORs quantity by image...") | |
train_ids, valid_ids, test_ids = tvt_by_nors( | |
self._ccagt_info, tvt_size=self.config.tvt_size, seed=self.config.seed | |
) | |
train_bn_images = self._ccagt_info.loc[self._ccagt_info["image_id"].isin(train_ids), "image_name"].unique() | |
valid_bn_images = self._ccagt_info.loc[self._ccagt_info["image_id"].isin(valid_ids), "image_name"].unique() | |
test_bn_images = self._ccagt_info.loc[self._ccagt_info["image_id"].isin(test_ids), "image_name"].unique() | |
if self.config.name == "semantic_segmentation": | |
train_data = images_and_masks(train_bn_images) | |
valid_data = images_and_masks(valid_bn_images) | |
test_data = images_and_masks(test_bn_images) | |
elif self.config.name == "object_detection": | |
train_data = images_and_boxes(train_bn_images) | |
valid_data = images_and_boxes(valid_bn_images) | |
test_data = images_and_boxes(test_bn_images) | |
elif self.config.name == "instance_segmentation": | |
train_data = images_and_instances(train_bn_images) | |
valid_data = images_and_instances(valid_bn_images) | |
test_data = images_and_instances(test_bn_images) | |
else: | |
raise NotImplementedError | |
return [ | |
datasets.SplitGenerator( | |
name=datasets.Split.TRAIN, | |
gen_kwargs={"data": train_data}, | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.TEST, | |
gen_kwargs={"data": test_data}, | |
), | |
datasets.SplitGenerator( | |
name=datasets.Split.VALIDATION, | |
gen_kwargs={"data": valid_data}, | |
), | |
] | |
def _generate_examples(self, data): | |
if self.config.name == "semantic_segmentation": | |
for img_path, msk_path in data: | |
img_basename = get_basename(img_path) | |
image_id = self._bn_to_imageid[img_basename] | |
yield image_id, { | |
"image": img_path, | |
"annotation": msk_path, | |
} | |
elif self.config.name == "object_detection": | |
for img_path, labels in data: | |
img_basename = get_basename(img_path) | |
image_id = self._bn_to_imageid[img_basename] | |
yield image_id, {"image": img_path, "objects": labels} | |
elif self.config.name == "instance_segmentation": | |
for img_path, instances in data: | |
img_basename = get_basename(img_path) | |
image_id = self._bn_to_imageid[img_basename] | |
yield image_id, {"image": img_path, "objects": instances} | |
else: | |
raise NotImplementedError | |