File size: 2,745 Bytes
779b904 9f9c8a1 09353ae 9f9c8a1 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 |
---
configs:
- config_name: default
data_files:
- split: validation
path: data/validation-*
- split: test
path: data/test-*
dataset_info:
features:
- name: instruction
dtype: string
- name: image
dtype: image
- name: image_url
dtype: string
- name: task
dtype: string
- name: split
dtype: string
- name: idx
dtype: int64
- name: hash
dtype: string
- name: input_caption
dtype: string
- name: output_caption
dtype: string
splits:
- name: validation
num_bytes: 766577760.29
num_examples: 2022
- name: test
num_bytes: 1353975788.0
num_examples: 3589
download_size: 1904999338
dataset_size: 2120553548.29
---
# Dataset Card for the Emu Edit Test Set
## Table of Contents
- [Table of Contents](#table-of-contents)
- [Dataset Description](#dataset-description)
- [Dataset Summary](#dataset-summary)
- [Additional Information](#additional-information)
- [Licensing Information](#licensing-information)
- [Citation Information](#citation-information)
## Dataset Description
- **Homepage: https://emu-edit.metademolab.com/**
- **Paper: TODO**
### Dataset Summary
To create a benchmark for image editing we first define seven different categories of potential image editing operations: background alteration (background), comprehensive image changes (global), style alteration (style), object removal (remove), object addition (add), localized modifications (local), and color/texture alterations (texture).
Then, we utilize the diverse set of input images from the [MagicBrush benchmark](https://huggingface.co/datasets/osunlp/MagicBrush), and for each editing operation, we task crowd workers to devise relevant, creative, and challenging instructions.
Moreover, to increase the quality of the collected examples, we apply a post-verification stage, in which crowd workers filter examples with irrelevant instructions.
Finally, to support evaluation for methods that require input and output captions (e.g. prompt2prompt and pnp), we additionally collect an input caption and output caption for each example.
When doing so, we ask annotators to ensure that the captions capture both important elements in the image, and elements that should change based on the instruction.
Additionally, to support proper comparison with Emu Edit with publicly release the model generations on the test set [here](https://huggingface.co/datasets/facebook/emu_edit_test_set_generations).
For more details please see our [paper](TODO) and [project page](https://emu-edit.metademolab.com/).
### Licensing Information
Creative Commons License This work is licensed under a Creative Commons Attribution 4.0 International License.
### Citation Information
TODO |