| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| | |
| |
|
| | """ |
| | Generate rich statistics for object detection datasets on Hugging Face Hub. |
| | |
| | Mirrors panlabel's stats command. Computes: |
| | |
| | - Summary counts (images, annotations, categories) |
| | - Label distribution histogram (top-N) |
| | - Bounding box statistics (area, aspect ratio, out-of-bounds) |
| | - Annotation density per image |
| | - Per-category bbox statistics |
| | - Category co-occurrence pairs |
| | - Image resolution distribution |
| | |
| | Supports COCO-style (xywh), XYXY/VOC, YOLO (normalized center xywh), |
| | TFOD (normalized xyxy), and Label Studio (percentage xywh) bbox formats. |
| | Supports streaming for large datasets. Outputs text or JSON. |
| | |
| | Examples: |
| | uv run stats-hf-dataset.py merve/test-coco-dataset |
| | uv run stats-hf-dataset.py merve/test-coco-dataset --top 20 --report json |
| | uv run stats-hf-dataset.py merve/test-coco-dataset --bbox-format tfod |
| | uv run stats-hf-dataset.py merve/test-coco-dataset --streaming --max-samples 5000 |
| | """ |
| |
|
| | import argparse |
| | import json |
| | import logging |
| | import math |
| | import os |
| | import sys |
| | import time |
| | from collections import Counter, defaultdict |
| | from datetime import datetime |
| | from typing import Any |
| |
|
| | from datasets import load_dataset |
| | from huggingface_hub import DatasetCard, login |
| | from tqdm.auto import tqdm |
| |
|
| | logging.basicConfig(level=logging.INFO) |
| | logger = logging.getLogger(__name__) |
| |
|
| | BBOX_FORMATS = ["coco_xywh", "xyxy", "voc", "yolo", "tfod", "label_studio"] |
| |
|
| |
|
| | def to_xyxy(bbox: list[float], fmt: str, img_w: float = 1.0, img_h: float = 1.0) -> tuple[float, float, float, float]: |
| | """Convert any bbox format to (xmin, ymin, xmax, ymax) in pixel space.""" |
| | if fmt == "coco_xywh": |
| | x, y, w, h = bbox |
| | return (x, y, x + w, y + h) |
| | elif fmt in ("xyxy", "voc"): |
| | return tuple(bbox[:4]) |
| | elif fmt == "yolo": |
| | cx, cy, w, h = bbox |
| | return (cx - w / 2) * img_w, (cy - h / 2) * img_h, (cx + w / 2) * img_w, (cy + h / 2) * img_h |
| | elif fmt == "tfod": |
| | xmin_n, ymin_n, xmax_n, ymax_n = bbox |
| | return (xmin_n * img_w, ymin_n * img_h, xmax_n * img_w, ymax_n * img_h) |
| | elif fmt == "label_studio": |
| | x_pct, y_pct, w_pct, h_pct = bbox |
| | return ( |
| | x_pct / 100.0 * img_w, |
| | y_pct / 100.0 * img_h, |
| | (x_pct + w_pct) / 100.0 * img_w, |
| | (y_pct + h_pct) / 100.0 * img_h, |
| | ) |
| | else: |
| | raise ValueError(f"Unknown bbox format: {fmt}") |
| |
|
| |
|
| | def percentile(sorted_vals: list[float], p: float) -> float: |
| | """Compute percentile from sorted values.""" |
| | if not sorted_vals: |
| | return 0.0 |
| | k = (len(sorted_vals) - 1) * p / 100.0 |
| | f = int(k) |
| | c = f + 1 |
| | if c >= len(sorted_vals): |
| | return sorted_vals[-1] |
| | return sorted_vals[f] + (k - f) * (sorted_vals[c] - sorted_vals[f]) |
| |
|
| |
|
| | def main( |
| | input_dataset: str, |
| | bbox_column: str = "bbox", |
| | category_column: str = "category", |
| | bbox_format: str = "coco_xywh", |
| | width_column: str | None = "width", |
| | height_column: str | None = "height", |
| | split: str = "train", |
| | max_samples: int | None = None, |
| | streaming: bool = False, |
| | top: int = 10, |
| | report_format: str = "text", |
| | tolerance: float = 0.5, |
| | hf_token: str | None = None, |
| | output_dataset: str | None = None, |
| | private: bool = False, |
| | ): |
| | """Compute statistics for an object detection dataset.""" |
| |
|
| | start_time = datetime.now() |
| |
|
| | HF_TOKEN = hf_token or os.environ.get("HF_TOKEN") |
| | if HF_TOKEN: |
| | login(token=HF_TOKEN) |
| |
|
| | logger.info(f"Loading dataset: {input_dataset} (split={split}, streaming={streaming})") |
| | dataset = load_dataset(input_dataset, split=split, streaming=streaming) |
| |
|
| | |
| | total_images = 0 |
| | total_annotations = 0 |
| | category_counts = Counter() |
| | annotations_per_image = [] |
| | areas = [] |
| | aspect_ratios = [] |
| | widths = [] |
| | heights = [] |
| | out_of_bounds_count = 0 |
| | zero_area_count = 0 |
| | per_category_areas = defaultdict(list) |
| | co_occurrence_pairs = Counter() |
| | images_without_annotations = 0 |
| |
|
| | iterable = dataset |
| | if max_samples: |
| | if streaming: |
| | iterable = dataset.take(max_samples) |
| | else: |
| | iterable = dataset.select(range(min(max_samples, len(dataset)))) |
| |
|
| | for idx, example in enumerate(tqdm(iterable, desc="Computing stats", total=max_samples)): |
| | total_images += 1 |
| |
|
| | objects = example.get("objects", example) |
| | bboxes = objects.get(bbox_column, []) or [] |
| | categories = objects.get(category_column, []) or [] |
| |
|
| | |
| | img_w = None |
| | img_h = None |
| | if width_column: |
| | img_w = example.get(width_column) or (objects.get(width_column) if isinstance(objects, dict) else None) |
| | if height_column: |
| | img_h = example.get(height_column) or (objects.get(height_column) if isinstance(objects, dict) else None) |
| |
|
| | if img_w is not None and img_h is not None: |
| | widths.append(img_w) |
| | heights.append(img_h) |
| |
|
| | num_anns = len(bboxes) |
| | annotations_per_image.append(num_anns) |
| | total_annotations += num_anns |
| |
|
| | if num_anns == 0: |
| | images_without_annotations += 1 |
| | continue |
| |
|
| | |
| | image_cats = set() |
| | for ann_idx, bbox in enumerate(bboxes): |
| | cat = categories[ann_idx] if ann_idx < len(categories) else None |
| | cat_str = str(cat) if cat is not None else "<unknown>" |
| | category_counts[cat_str] += 1 |
| | image_cats.add(cat_str) |
| |
|
| | if bbox is None or len(bbox) < 4: |
| | continue |
| | if not all(math.isfinite(v) for v in bbox[:4]): |
| | continue |
| |
|
| | w_for_conv = img_w if img_w else 1.0 |
| | h_for_conv = img_h if img_h else 1.0 |
| | xmin, ymin, xmax, ymax = to_xyxy(bbox[:4], bbox_format, w_for_conv, h_for_conv) |
| |
|
| | bw = xmax - xmin |
| | bh = ymax - ymin |
| | area = bw * bh |
| |
|
| | if area <= 0: |
| | zero_area_count += 1 |
| | else: |
| | areas.append(area) |
| | per_category_areas[cat_str].append(area) |
| |
|
| | if bh > 0: |
| | aspect_ratios.append(bw / bh) |
| |
|
| | |
| | if img_w is not None and img_h is not None: |
| | if xmin < -tolerance or ymin < -tolerance or xmax > img_w + tolerance or ymax > img_h + tolerance: |
| | out_of_bounds_count += 1 |
| |
|
| | |
| | sorted_cats = sorted(image_cats) |
| | for i in range(len(sorted_cats)): |
| | for j in range(i + 1, len(sorted_cats)): |
| | co_occurrence_pairs[(sorted_cats[i], sorted_cats[j])] += 1 |
| |
|
| | processing_time = datetime.now() - start_time |
| |
|
| | |
| | areas.sort() |
| | aspect_ratios.sort() |
| | annotations_per_image.sort() |
| |
|
| | def dist_stats(vals: list[float]) -> dict: |
| | if not vals: |
| | return {"count": 0, "min": 0, "max": 0, "mean": 0, "median": 0, "p25": 0, "p75": 0} |
| | return { |
| | "count": len(vals), |
| | "min": round(vals[0], 2), |
| | "max": round(vals[-1], 2), |
| | "mean": round(sum(vals) / len(vals), 2), |
| | "median": round(percentile(vals, 50), 2), |
| | "p25": round(percentile(vals, 25), 2), |
| | "p75": round(percentile(vals, 75), 2), |
| | } |
| |
|
| | |
| | top_categories = category_counts.most_common(top) |
| |
|
| | |
| | top_cooccurrences = co_occurrence_pairs.most_common(top) |
| |
|
| | |
| | per_cat_stats = {} |
| | for cat, cat_areas in sorted(per_category_areas.items(), key=lambda x: -len(x[1])): |
| | cat_areas.sort() |
| | per_cat_stats[cat] = dist_stats(cat_areas) |
| |
|
| | report = { |
| | "dataset": input_dataset, |
| | "split": split, |
| | "summary": { |
| | "total_images": total_images, |
| | "total_annotations": total_annotations, |
| | "unique_categories": len(category_counts), |
| | "images_without_annotations": images_without_annotations, |
| | "out_of_bounds_bboxes": out_of_bounds_count, |
| | "zero_area_bboxes": zero_area_count, |
| | }, |
| | "label_distribution": {cat: count for cat, count in top_categories}, |
| | "annotation_density": dist_stats([float(x) for x in annotations_per_image]), |
| | "bbox_area": dist_stats(areas), |
| | "bbox_aspect_ratio": dist_stats(aspect_ratios), |
| | "image_resolution": { |
| | "width": dist_stats([float(w) for w in sorted(widths)]) if widths else {}, |
| | "height": dist_stats([float(h) for h in sorted(heights)]) if heights else {}, |
| | }, |
| | "per_category_area": {cat: per_cat_stats[cat] for cat in list(per_cat_stats)[:top]}, |
| | "co_occurrence_pairs": [ |
| | {"pair": list(pair), "count": count} for pair, count in top_cooccurrences |
| | ], |
| | "processing_time_seconds": processing_time.total_seconds(), |
| | "timestamp": datetime.now().isoformat(), |
| | } |
| |
|
| | if report_format == "json": |
| | print(json.dumps(report, indent=2)) |
| | else: |
| | print("\n" + "=" * 60) |
| | print(f"Dataset Statistics: {input_dataset}") |
| | print("=" * 60) |
| |
|
| | s = report["summary"] |
| | print(f"\n Images: {s['total_images']:,}") |
| | print(f" Annotations: {s['total_annotations']:,}") |
| | print(f" Categories: {s['unique_categories']:,}") |
| | print(f" Empty images: {s['images_without_annotations']:,}") |
| | print(f" Out-of-bounds: {s['out_of_bounds_bboxes']:,}") |
| | print(f" Zero-area bboxes: {s['zero_area_bboxes']:,}") |
| |
|
| | if total_images > 0: |
| | print(f"\n Annotations/image: {total_annotations / total_images:.1f} avg") |
| |
|
| | d = report["annotation_density"] |
| | if d["count"]: |
| | print(f" min={d['min']}, median={d['median']}, max={d['max']}") |
| |
|
| | print(f"\n Label Distribution (top {top}):") |
| | for cat, count in top_categories: |
| | pct = 100.0 * count / total_annotations if total_annotations else 0 |
| | bar = "#" * int(pct / 2) |
| | print(f" {cat:30s} {count:>8,} ({pct:5.1f}%) {bar}") |
| |
|
| | a = report["bbox_area"] |
| | if a["count"]: |
| | print(f"\n Bbox Area:") |
| | print(f" min={a['min']}, median={a['median']}, mean={a['mean']}, max={a['max']}") |
| |
|
| | ar = report["bbox_aspect_ratio"] |
| | if ar["count"]: |
| | print(f"\n Bbox Aspect Ratio (w/h):") |
| | print(f" min={ar['min']}, median={ar['median']}, mean={ar['mean']}, max={ar['max']}") |
| |
|
| | if top_cooccurrences: |
| | print(f"\n Category Co-occurrence (top {top}):") |
| | for pair, count in top_cooccurrences: |
| | print(f" {pair[0]} + {pair[1]}: {count:,}") |
| |
|
| | print(f"\n Processing time: {processing_time.total_seconds():.1f}s") |
| | print("=" * 60) |
| |
|
| | |
| | if output_dataset: |
| | from datasets import Dataset as HFDataset |
| |
|
| | report_ds = HFDataset.from_dict({ |
| | "report_json": [json.dumps(report)], |
| | "dataset": [input_dataset], |
| | "total_images": [total_images], |
| | "total_annotations": [total_annotations], |
| | "unique_categories": [len(category_counts)], |
| | "timestamp": [datetime.now().isoformat()], |
| | }) |
| |
|
| | logger.info(f"Pushing stats report to {output_dataset}") |
| | max_retries = 3 |
| | for attempt in range(1, max_retries + 1): |
| | try: |
| | if attempt > 1: |
| | os.environ["HF_HUB_DISABLE_XET"] = "1" |
| | report_ds.push_to_hub(output_dataset, private=private, token=HF_TOKEN) |
| | break |
| | except Exception as e: |
| | logger.error(f"Upload attempt {attempt}/{max_retries} failed: {e}") |
| | if attempt < max_retries: |
| | time.sleep(30 * (2 ** (attempt - 1))) |
| | else: |
| | logger.error("All upload attempts failed.") |
| | sys.exit(1) |
| |
|
| | logger.info(f"Stats pushed to: https://huggingface.co/datasets/{output_dataset}") |
| |
|
| |
|
| | if __name__ == "__main__": |
| | parser = argparse.ArgumentParser( |
| | description="Generate statistics for object detection datasets on HF Hub", |
| | formatter_class=argparse.RawDescriptionHelpFormatter, |
| | epilog=""" |
| | Bbox formats: |
| | coco_xywh [x, y, width, height] in pixels (default) |
| | xyxy [xmin, ymin, xmax, ymax] in pixels |
| | voc [xmin, ymin, xmax, ymax] in pixels (alias for xyxy) |
| | yolo [cx, cy, w, h] normalized 0-1 |
| | tfod [xmin, ymin, xmax, ymax] normalized 0-1 |
| | label_studio [x, y, w, h] percentage 0-100 |
| | |
| | Examples: |
| | uv run stats-hf-dataset.py merve/coco-dataset |
| | uv run stats-hf-dataset.py merve/coco-dataset --top 20 --report json |
| | uv run stats-hf-dataset.py merve/coco-dataset --streaming --max-samples 5000 |
| | """, |
| | ) |
| |
|
| | parser.add_argument("input_dataset", help="Input dataset ID on HF Hub") |
| | parser.add_argument("--bbox-column", default="bbox", help="Column containing bboxes (default: bbox)") |
| | parser.add_argument("--category-column", default="category", help="Column containing categories (default: category)") |
| | parser.add_argument("--bbox-format", choices=BBOX_FORMATS, default="coco_xywh", help="Bbox format (default: coco_xywh)") |
| | parser.add_argument("--width-column", default="width", help="Column for image width (default: width)") |
| | parser.add_argument("--height-column", default="height", help="Column for image height (default: height)") |
| | parser.add_argument("--split", default="train", help="Dataset split (default: train)") |
| | parser.add_argument("--max-samples", type=int, help="Max samples to process") |
| | parser.add_argument("--streaming", action="store_true", help="Use streaming mode") |
| | parser.add_argument("--top", type=int, default=10, help="Top-N items for histograms (default: 10)") |
| | parser.add_argument("--report", choices=["text", "json"], default="text", help="Report format (default: text)") |
| | parser.add_argument("--tolerance", type=float, default=0.5, help="Out-of-bounds tolerance in pixels (default: 0.5)") |
| | parser.add_argument("--hf-token", help="HF API token") |
| | parser.add_argument("--output-dataset", help="Push stats report to this HF dataset") |
| | parser.add_argument("--private", action="store_true", help="Make output dataset private") |
| |
|
| | args = parser.parse_args() |
| |
|
| | main( |
| | input_dataset=args.input_dataset, |
| | bbox_column=args.bbox_column, |
| | category_column=args.category_column, |
| | bbox_format=args.bbox_format, |
| | width_column=args.width_column, |
| | height_column=args.height_column, |
| | split=args.split, |
| | max_samples=args.max_samples, |
| | streaming=args.streaming, |
| | top=args.top, |
| | report_format=args.report, |
| | tolerance=args.tolerance, |
| | hf_token=args.hf_token, |
| | output_dataset=args.output_dataset, |
| | private=args.private, |
| | ) |
| |
|