Datasets:
Tasks:
Image-to-Text
Sub-tasks:
image-captioning
Languages:
English
Size:
10M<n<100M
ArXiv:
License:
annotations_creators: | |
- found | |
language_creators: | |
- found | |
language: | |
- en | |
license: | |
- other | |
multilinguality: | |
- monolingual | |
size_categories: | |
- 10M<n<100M | |
source_datasets: | |
- original | |
task_categories: | |
- image-to-text | |
task_ids: | |
- image-captioning | |
paperswithcode_id: cc12m | |
pretty_name: Conceptual 12M | |
dataset_info: | |
features: | |
- name: image_url | |
dtype: string | |
- name: caption | |
dtype: string | |
splits: | |
- name: train | |
num_bytes: 2794168030 | |
num_examples: 12423374 | |
download_size: 2707204412 | |
dataset_size: 2794168030 | |
# Dataset Card for Conceptual 12M | |
## Table of Contents | |
- [Dataset Description](#dataset-description) | |
- [Dataset Summary](#dataset-summary) | |
- [Dataset Preprocessing](#dataset-preprocessing) | |
- [Supported Tasks](#supported-tasks-and-leaderboards) | |
- [Languages](#languages) | |
- [Dataset Structure](#dataset-structure) | |
- [Data Instances](#data-instances) | |
- [Data Fields](#data-instances) | |
- [Data Splits](#data-instances) | |
- [Dataset Creation](#dataset-creation) | |
- [Curation Rationale](#curation-rationale) | |
- [Source Data](#source-data) | |
- [Annotations](#annotations) | |
- [Personal and Sensitive Information](#personal-and-sensitive-information) | |
- [Considerations for Using the Data](#considerations-for-using-the-data) | |
- [Social Impact of Dataset](#social-impact-of-dataset) | |
- [Discussion of Biases](#discussion-of-biases) | |
- [Other Known Limitations](#other-known-limitations) | |
- [Additional Information](#additional-information) | |
- [Dataset Curators](#dataset-curators) | |
- [Licensing Information](#licensing-information) | |
- [Citation Information](#citation-information) | |
## Dataset Description | |
- **Repository:** [Conceptual 12M repository](https://github.com/google-research-datasets/conceptual-12m) | |
- **Paper:** [Conceptual 12M: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts](https://arxiv.org/abs/2102.08981) | |
- **Point of Contact:** [Conceptual Captions e-mail](mailto:conceptual-captions@google.com) | |
### Dataset Summary | |
Conceptual 12M (CC12M) is a dataset with 12 million image-text pairs specifically meant to be used for visionand-language pre-training. | |
Its data collection pipeline is a relaxed version of the one used in Conceptual Captions 3M (CC3M). | |
### Dataset Preprocessing | |
This dataset doesn't download the images locally by default. Instead, it exposes URLs to the images. To fetch the images, use the following code: | |
```python | |
from concurrent.futures import ThreadPoolExecutor | |
from functools import partial | |
import io | |
import urllib | |
import PIL.Image | |
from datasets import load_dataset | |
from datasets.utils.file_utils import get_datasets_user_agent | |
USER_AGENT = get_datasets_user_agent() | |
def fetch_single_image(image_url, timeout=None, retries=0): | |
for _ in range(retries + 1): | |
try: | |
request = urllib.request.Request( | |
image_url, | |
data=None, | |
headers={"user-agent": USER_AGENT}, | |
) | |
with urllib.request.urlopen(request, timeout=timeout) as req: | |
image = PIL.Image.open(io.BytesIO(req.read())) | |
break | |
except Exception: | |
image = None | |
return image | |
def fetch_images(batch, num_threads, timeout=None, retries=0): | |
fetch_single_image_with_args = partial(fetch_single_image, timeout=timeout, retries=retries) | |
with ThreadPoolExecutor(max_workers=num_threads) as executor: | |
batch["image"] = list(executor.map(fetch_single_image_with_args, batch["image_url"])) | |
return batch | |
num_threads = 20 | |
dset = load_dataset("conceptual_12m") | |
dset = dset.map(fetch_images, batched=True, batch_size=100, fn_kwargs={"num_threads": num_threads}) | |
``` | |
### Supported Tasks and Leaderboards | |
- `image-captioning`: This dataset can be used to train model for the Image Captioning task. | |
### Languages | |
All captions are in English. | |
## Dataset Structure | |
### Data Instances | |
Each instance represents a single image with a caption: | |
``` | |
{ | |
'image_url': 'http://lh6.ggpht.com/-IvRtNLNcG8o/TpFyrudaT6I/AAAAAAAAM6o/_11MuAAKalQ/IMG_3422.JPG?imgmax=800', | |
'caption': 'a very typical bus station' | |
} | |
``` | |
### Data Fields | |
- `image_url`: Static URL for downloading the image associated with the post. | |
- `caption`: Textual description of the image. | |
### Data Splits | |
There is only training data, with a total of 12423374 rows | |
## Dataset Creation | |
### Curation Rationale | |
Conceptual 12M shares the same pipeline with Conceptual Captions (CC3M), but relaxes some processing steps. | |
### Source Data | |
#### Initial Data Collection and Normalization | |
From the paper: | |
> To arrive at CC12M, we keep | |
the image-text filtering intact, and relax the unimodal filters only. First, for image-based filtering, we set the maximum ratio of larger to smaller dimension to 2.5 instead of 2. | |
We still keep only JPEG images with size greater than | |
400 pixels, and still exclude images that trigger pornography detectors. Second, in text-based filtering, we allow text | |
between 3 and 256 words in the alt-text. We still discard | |
candidates with no noun or no determiner, but permit ones | |
without prepositions. We discard the heuristics regarding | |
high unique-word ratio covering various POS tags and word | |
capitalization. We set the maximum fraction of word repetition allowed to 0.2. Given a larger pool of text due to the | |
above relaxations, the threshold for counting a word type as | |
rare is increased from 5 to 20 | |
> The main motivation for CC3M to | |
perform text transformation is that a majority of candidate | |
captions contain ultrafine-grained entities such as proper | |
names (people, venues, locations, etc.), making it extremely | |
difficult to learn as part of the image captioning task. In | |
contrast, we are not restricted by the end task of image caption generation. Our intuition is that relatively more difficult pre-training data would lead to better transferability. | |
We thus do not perform hypernimization or digit substitution. [...] The only exception to the “keep alt-texts as | |
raw as possible” rule is performing person-name substitutions, which we identify as necessary to protect the privacy | |
of the individuals in these images. For this step, we use the | |
Google Cloud Natural Language APIs to detect all named | |
entities of type Person, and substitute them by a special token <PERSON>. Around 25% of all the alt-texts in CC12M | |
are transformed in this fashion. | |
#### Who are the source language producers? | |
Not specified. | |
### Annotations | |
#### Annotation process | |
Annotations are extracted jointly with the images using the automatic pipeline. | |
#### Who are the annotators? | |
Not specified. | |
### Personal and Sensitive Information | |
From the paper: | |
> The only exception to the “keep alt-texts as | |
raw as possible” rule is performing person-name substitutions, which we identify as necessary to protect the privacy | |
of the individuals in these images. For this step, we use the | |
Google Cloud Natural Language APIs to detect all named | |
entities of type Person, and substitute them by a special token <PERSON>. Around 25% of all the alt-texts in CC12M | |
are transformed in this fashion. | |
## Considerations for Using the Data | |
### Social Impact of Dataset | |
[More Information Needed] | |
### Discussion of Biases | |
[More Information Needed] | |
### Other Known Limitations | |
[More Information Needed] | |
## Additional Information | |
### Dataset Curators | |
Soravit Changpinyo, Piyush Sharma, Nan Ding and Radu Soricut. | |
### Licensing Information | |
The dataset may be freely used for any purpose, although acknowledgement of | |
Google LLC ("Google") as the data source would be appreciated. The dataset is | |
provided "AS IS" without any warranty, express or implied. Google disclaims all | |
liability for any damages, direct or indirect, resulting from the use of the | |
dataset. | |
### Citation Information | |
```bibtex | |
@inproceedings{changpinyo2021cc12m, | |
title = {{Conceptual 12M}: Pushing Web-Scale Image-Text Pre-Training To Recognize Long-Tail Visual Concepts}, | |
author = {Changpinyo, Soravit and Sharma, Piyush and Ding, Nan and Soricut, Radu}, | |
booktitle = {CVPR}, | |
year = {2021}, | |
} | |
``` | |
### Contributions | |
Thanks to [@thomasw21](https://github.com/thomasw21) for adding this dataset. |