Datasets:
annotations_creators:
- expert-generated
language_creators:
- found
language:
- en
license:
- cc-by-nc-nd-4.0
multilinguality:
- monolingual
size_categories:
- 10K<n<100K
source_datasets:
- extended|other-wider
task_categories:
- object-detection
task_ids:
- face-detection
paperswithcode_id: wider-face-1
pretty_name: WIDER FACE
dataset_info:
features:
- name: image
dtype: image
- name: faces
sequence:
- name: bbox
sequence: float32
length: 4
- name: blur
dtype:
class_label:
names:
'0': clear
'1': normal
'2': heavy
- name: expression
dtype:
class_label:
names:
'0': typical
'1': exaggerate
- name: illumination
dtype:
class_label:
names:
'0': normal
'1': 'exaggerate '
- name: occlusion
dtype:
class_label:
names:
'0': 'no'
'1': partial
'2': heavy
- name: pose
dtype:
class_label:
names:
'0': typical
'1': atypical
- name: invalid
dtype: bool
splits:
- name: train
num_bytes: 12049881
num_examples: 12880
- name: test
num_bytes: 3761103
num_examples: 16097
- name: validation
num_bytes: 2998735
num_examples: 3226
download_size: 3676086479
dataset_size: 18809719
Dataset Card for WIDER FACE
Table of Contents
- Table of Contents
- Dataset Description
- Dataset Structure
- Dataset Creation
- Considerations for Using the Data
- Additional Information
Dataset Description
- Homepage: http://shuoyang1213.me/WIDERFACE/index.html
- Repository:
- Paper: WIDER FACE: A Face Detection Benchmark
- Leaderboard: http://shuoyang1213.me/WIDERFACE/WiderFace_Results.html
- Point of Contact: shuoyang.1213@gmail.com
Dataset Summary
WIDER FACE dataset is a face detection benchmark dataset, of which images are selected from the publicly available WIDER dataset. We choose 32,203 images and label 393,703 faces with a high degree of variability in scale, pose and occlusion as depicted in the sample images. WIDER FACE dataset is organized based on 61 event classes. For each event class, we randomly select 40%/10%/50% data as training, validation and testing sets. We adopt the same evaluation metric employed in the PASCAL VOC dataset. Similar to MALF and Caltech datasets, we do not release bounding box ground truth for the test images. Users are required to submit final prediction files, which we shall proceed to evaluate.
Supported Tasks and Leaderboards
face-detection
: The dataset can be used to train a model for Face Detection. More information on evaluating the model's performance can be found here.
Languages
English
Dataset Structure
Data Instances
A data point comprises an image and its face annotations.
{
'image': <PIL.JpegImagePlugin.JpegImageFile image mode=RGB size=1024x755 at 0x19FA12186D8>, 'faces': {
'bbox': [
[178.0, 238.0, 55.0, 73.0],
[248.0, 235.0, 59.0, 73.0],
[363.0, 157.0, 59.0, 73.0],
[468.0, 153.0, 53.0, 72.0],
[629.0, 110.0, 56.0, 81.0],
[745.0, 138.0, 55.0, 77.0]
],
'blur': [2, 2, 2, 2, 2, 2],
'expression': [0, 0, 0, 0, 0, 0],
'illumination': [0, 0, 0, 0, 0, 0],
'occlusion': [1, 2, 1, 2, 1, 2],
'pose': [0, 0, 0, 0, 0, 0],
'invalid': [False, False, False, False, False, False]
}
}
Data Fields
image
: APIL.Image.Image
object containing the image. Note that when accessing the image column:dataset[0]["image"]
the image file is automatically decoded. Decoding of a large number of image files might take a significant amount of time. Thus it is important to first query the sample index before the"image"
column, i.e.dataset[0]["image"]
should always be preferred overdataset["image"][0]
faces
: a dictionary of face attributes for the faces present on the imagebbox
: the bounding box of each face (in the coco format)blur
: the blur level of each face, with possible values includingclear
(0),normal
(1) andheavy
expression
: the facial expression of each face, with possible values includingtypical
(0) andexaggerate
(1)illumination
: the lightning condition of each face, with possible values includingnormal
(0) andexaggerate
(1)occlusion
: the level of occlusion of each face, with possible values includingno
(0),partial
(1) andheavy
(2)pose
: the pose of each face, with possible values includingtypical
(0) andatypical
(1)invalid
: whether the image is valid or invalid.
Data Splits
The data is split into training, validation and testing set. WIDER FACE dataset is organized based on 61 event classes. For each event class, 40%/10%/50% data is randomly selected as training, validation and testing sets. The training set contains 12880 images, the validation set 3226 images and test set 16097 images.
Dataset Creation
Curation Rationale
The curators state that the current face detection datasets typically contain a few thousand faces, with limited variations in pose, scale, facial expression, occlusion, and background clutters, making it difficult to assess for real world performance. They argue that the limitations of datasets have partially contributed to the failure of some algorithms in coping with heavy occlusion, small scale, and atypical pose.
Source Data
Initial Data Collection and Normalization
WIDER FACE dataset is a subset of the WIDER dataset. The images in WIDER were collected in the following three steps: 1) Event categories were defined and chosen following the Large Scale Ontology for Multimedia (LSCOM) [22], which provides around 1000 concepts relevant to video event analysis. 2) Images are retrieved using search engines like Google and Bing. For each category, 1000-3000 images were collected. 3) The data were cleaned by manually examining all the images and filtering out images without human face. Then, similar images in each event category were removed to ensure large diversity in face appearance. A total of 32203 images are eventually included in the WIDER FACE dataset.
Who are the source language producers?
The images are selected from publicly available WIDER dataset.
Annotations
Annotation process
The curators label the bounding boxes for all the recognizable faces in the WIDER FACE dataset. The bounding box is required to tightly contain the forehead, chin, and cheek.. If a face is occluded, they still label it with a bounding box but with an estimation on the scale of occlusion. Similar to the PASCAL VOC dataset [6], they assign an ’Ignore’ flag to the face which is very difficult to be recognized due to low resolution and small scale (10 pixels or less). After annotating the face bounding boxes, they further annotate the following attributes: pose (typical, atypical) and occlusion level (partial, heavy). Each annotation is labeled by one annotator and cross-checked by two different people.
Who are the annotators?
Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang.
Personal and Sensitive Information
[More Information Needed]
Considerations for Using the Data
Social Impact of Dataset
[More Information Needed]
Discussion of Biases
[More Information Needed]
Other Known Limitations
[More Information Needed]
Additional Information
Dataset Curators
Shuo Yang, Ping Luo, Chen Change Loy and Xiaoou Tang
Licensing Information
Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International (CC BY-NC-ND 4.0).
Citation Information
@inproceedings{yang2016wider,
Author = {Yang, Shuo and Luo, Ping and Loy, Chen Change and Tang, Xiaoou},
Booktitle = {IEEE Conference on Computer Vision and Pattern Recognition (CVPR)},
Title = {WIDER FACE: A Face Detection Benchmark},
Year = {2016}}
Contributions
Thanks to @mariosasko for adding this dataset.