File size: 4,676 Bytes
3b99a71
 
 
 
 
 
 
 
 
 
 
 
 
64cc260
 
3b99a71
64cc260
 
3b99a71
 
 
 
64cc260
3b99a71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
64cc260
3b99a71
 
 
 
 
 
 
 
 
 
 
 
 
b70b9e6
3b99a71
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
988d049
3b99a71
 
 
 
 
 
 
988d049
3b99a71
988d049
3b99a71
988d049
3b99a71
6c00f1e
 
988d049
 
3b99a71
988d049
 
3b99a71
 
 
988d049
3b99a71
 
 
988d049
3b99a71
988d049
3b99a71
988d049
3b99a71
988d049
3b99a71
5563a17
3b99a71
988d049
3b99a71
988d049
3b99a71
988d049
3b99a71
988d049
3b99a71
5563a17
988d049
 
 
 
 
5563a17
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
---
annotations_creators: []
language: en
size_categories:
- 1K<n<10K
task_categories:
- image-classification
task_ids: []
pretty_name: ImageNet-D
tags:
- fiftyone
- image
- image-classification
- synthetic
dataset_summary: >

  This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 4838
  samples.

  ## Installation


  If you haven't already, install FiftyOne:


  ```bash

  pip install -U fiftyone

  ```
  ## Usage


  ```python

  import fiftyone as fo

  import fiftyone.utils.huggingface as fouh


  # Load the dataset

  # Note: other available arguments include 'max_samples', etc

  dataset = fouh.load_from_hub("harpreetsahota/ImageNet-D")


  # Launch the App

  session = fo.launch_app(dataset)

  ```
---

# Dataset Card for ImageNet-D

![image/png](imagenet-d.gif)

This is a [FiftyOne](https://github.com/voxel51/fiftyone) dataset with 4838 samples.

## Installation

If you haven't already, install FiftyOne:

```bash
pip install -U fiftyone
```

## Usage

```python
import fiftyone as fo
import fiftyone.utils.huggingface as fouh

# Load the dataset
# Note: other available arguments include 'max_samples', etc
dataset = fouh.load_from_hub("Voxel51/ImageNet-D")

# Launch the App
session = fo.launch_app(dataset)
```

### Dataset Description

ImageNet-D is a new benchmark created using diffusion models to generate realistic synthetic images with diverse backgrounds, textures, and materials[1]. The dataset contains 4,835 hard images that cause significant accuracy drops of up to 60% for a range of vision models, including ResNet, ViT, CLIP, LLaVa, and MiniGPT-4[1].

To create ImageNet-D, a large pool of synthetic images is generated by combining object categories with various nuisance attributes using Stable Diffusion[1]. The most challenging images that cause shared failures across multiple surrogate models are selected for the final dataset[1]. Human labelling via Amazon Mechanical Turk is used for quality control to ensure the images are valid and high-quality[1].

Experiments show that ImageNet-D reveals significant robustness gaps in current vision models[1]. The synthetic images transfer well to unseen models, uncovering common failure modes[1]. ImageNet-D provides a more diverse and challenging test set than prior synthetic benchmarks like ImageNet-C, ImageNet-9, and Stylized ImageNet[1].

The recipe notebook for creating this dataset can be found [here](https://colab.research.google.com/drive/1iiiXN8B36YhjtOH2PDbHevHTXH736It_?usp=sharing)

Citations:
[1] https://arxiv.org/html/2403.18775v1

- **Funded by :** KAIST, University of Michigan, Ann Arbor, McGill University, MILA
- **License:** MIT License

### Source Data

See the [original repo](https://github.com/chenshuang-zhang/imagenet_d) for details

#### Data Collection and Processing

The ImageNet-D dataset was constructed using diffusion models to generate a large pool of realistic synthetic images covering various combinations of object categories and nuisance attributes. The key steps in the data collection and generation process were:

1. **Image generation**: The Stable Diffusion model was used to generate high-fidelity images based on user-defined text prompts specifying the desired object category (C) and nuisance attributes (N) such as background, material, and texture. The image generation is formulated as:

   Image(C, N) = StableDiffusion(Prompt(C, N)) 

   For example, to generate an image of a backpack, the prompt might specify "a backpack in a wheat field" to control both the object category and background nuisance.

2. **Prompt design**: A set of prompts was carefully designed to cover a matrix of object categories and nuisance attributes (see [Table 1 in the paper](https://arxiv.org/html/2403.18775v1#S3) for an overview). This allows generating images with a much broader range of category-nuisance combinations compared to existing test sets.

3. **Labeling**: Each generated image is automatically labeled with the object category (C) specified in its generation prompt. This category label serves as the ground truth for evaluating classification models on the ImageNet-D dataset. A classification is considered incorrect if the model's predicted class does not match the ground truth category.

#### Who are the source data producers?

Chenshuang Zhang, Fei Pan, Junmo Kim, In So Kweon, Chengzhi Mao

## Citation 
**BibTeX:**
```bibtex
@article{zhang2024imagenet_d,
  author    = {Zhang, Chenshuang and Pan, Fei and Kim, Junmo and Kweon, In So and Mao, Chengzhi},
  title     = {ImageNet-D: Benchmarking Neural Network Robustness on Diffusion Synthetic Object},
  journal   = {CVPR},
  year      = {2024},
}
```