harpreetsahota commited on
Commit
5563a17
1 Parent(s): 988d049

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -2
README.md CHANGED
@@ -108,7 +108,7 @@ The ImageNet-D dataset was constructed using diffusion models to generate a larg
108
 
109
  For example, to generate an image of a backpack, the prompt might specify "a backpack in a wheat field" to control both the object category and background nuisance.
110
 
111
- 2. **Prompt design**: A set of prompts was carefully designed to cover a matrix of object categories and nuisance attributes (see Table 1 in the paper for an overview). This allows generating images with a much broader range of category-nuisance combinations compared to existing test sets.
112
 
113
  3. **Labeling**: Each generated image is automatically labeled with the object category (C) specified in its generation prompt. This category label serves as the ground truth for evaluating classification models on the ImageNet-D dataset. A classification is considered incorrect if the model's predicted class does not match the ground truth category.
114
 
@@ -118,9 +118,11 @@ Chenshuang Zhang, Fei Pan, Junmo Kim, In So Kweon, Chengzhi Mao
118
 
119
  ## Citation
120
  **BibTeX:**
 
121
  @article{zhang2024imagenet_d,
122
  author = {Zhang, Chenshuang and Pan, Fei and Kim, Junmo and Kweon, In So and Mao, Chengzhi},
123
  title = {ImageNet-D: Benchmarking Neural Network Robustness on Diffusion Synthetic Object},
124
  journal = {CVPR},
125
  year = {2024},
126
- }
 
 
108
 
109
  For example, to generate an image of a backpack, the prompt might specify "a backpack in a wheat field" to control both the object category and background nuisance.
110
 
111
+ 2. **Prompt design**: A set of prompts was carefully designed to cover a matrix of object categories and nuisance attributes (see [Table 1 in the paper](https://arxiv.org/html/2403.18775v1#S3) for an overview). This allows generating images with a much broader range of category-nuisance combinations compared to existing test sets.
112
 
113
  3. **Labeling**: Each generated image is automatically labeled with the object category (C) specified in its generation prompt. This category label serves as the ground truth for evaluating classification models on the ImageNet-D dataset. A classification is considered incorrect if the model's predicted class does not match the ground truth category.
114
 
 
118
 
119
  ## Citation
120
  **BibTeX:**
121
+ ```bibtex
122
  @article{zhang2024imagenet_d,
123
  author = {Zhang, Chenshuang and Pan, Fei and Kim, Junmo and Kweon, In So and Mao, Chengzhi},
124
  title = {ImageNet-D: Benchmarking Neural Network Robustness on Diffusion Synthetic Object},
125
  journal = {CVPR},
126
  year = {2024},
127
+ }
128
+ ```