selfies_and_id / selfies_and_id.py
Vadzim Kashko
refactor: script and readme
2271465
raw
history blame contribute delete
No virus
4.55 kB
import io
import datasets
import pandas as pd
_CITATION = """\
@InProceedings{huggingface:dataset,
title = {selfies_and_id},
author = {TrainingDataPro},
year = {2023}
}
"""
_DESCRIPTION = """\
4083 sets, which includes 2 photos of a person from his documents and
13 selfies. 571 sets of Hispanics and 3512 sets of Caucasians.
Photo documents contains only a photo of a person.
All personal information from the document is hidden.
"""
_NAME = 'selfies_and_id'
_HOMEPAGE = f"https://huggingface.co/datasets/TrainingDataPro/{_NAME}"
_LICENSE = ""
_DATA = f"https://huggingface.co/datasets/TrainingDataPro/{_NAME}/resolve/main/data/"
class SelfiesAndId(datasets.GeneratorBasedBuilder):
"""Small sample of image-text pairs"""
def _info(self):
return datasets.DatasetInfo(
description=_DESCRIPTION,
features=datasets.Features({
'id_1': datasets.Image(),
'id_2': datasets.Image(),
'selfie_1': datasets.Image(),
'selfie_2': datasets.Image(),
'selfie_3': datasets.Image(),
'selfie_4': datasets.Image(),
'selfie_5': datasets.Image(),
'selfie_6': datasets.Image(),
'selfie_7': datasets.Image(),
'selfie_8': datasets.Image(),
'selfie_9': datasets.Image(),
'selfie_10': datasets.Image(),
'selfie_11': datasets.Image(),
'selfie_12': datasets.Image(),
'selfie_13': datasets.Image(),
'user_id': datasets.Value('string'),
'set_id': datasets.Value('string'),
'user_race': datasets.Value('string'),
'name': datasets.Value('string'),
'age': datasets.Value('int8'),
'country': datasets.Value('string'),
'gender': datasets.Value('string')
}),
supervised_keys=None,
homepage=_HOMEPAGE,
citation=_CITATION,
)
def _split_generators(self, dl_manager):
images = dl_manager.download(f"{_DATA}images.tar.gz")
annotations = dl_manager.download(f"{_DATA}{_NAME}.csv")
images = dl_manager.iter_archive(images)
return [
datasets.SplitGenerator(name=datasets.Split.TRAIN,
gen_kwargs={
"images": images,
'annotations': annotations
}),
]
def _generate_examples(self, images, annotations):
annotations_df = pd.read_csv(annotations, sep=';')
images_data = pd.DataFrame(columns=['URL', 'Bytes'])
for idx, (image_path, image) in enumerate(images):
images_data.loc[idx] = {'URL': image_path, 'Bytes': image.read()}
annotations_df = pd.merge(annotations_df,
images_data,
how='left',
on=['URL'])
for idx, worker_id in enumerate(pd.unique(annotations_df['UserId'])):
annotation = annotations_df.loc[annotations_df['UserId'] ==
worker_id]
annotation = annotation.sort_values(['FName'])
data = {
row[5].lower(): {
'path': row[6],
'bytes': row[10]
} for row in annotation.itertuples()
}
age = annotation.loc[annotation['FName'] ==
'ID_1']['Age'].values[0]
country = annotation.loc[annotation['FName'] ==
'ID_1']['Country'].values[0]
gender = annotation.loc[annotation['FName'] ==
'ID_1']['Gender'].values[0]
set_id = annotation.loc[annotation['FName'] ==
'ID_1']['SetId'].values[0]
user_race = annotation.loc[annotation['FName'] ==
'ID_1']['UserRace'].values[0]
name = annotation.loc[annotation['FName'] ==
'ID_1']['Name'].values[0]
data['user_id'] = worker_id
data['age'] = age
data['country'] = country
data['gender'] = gender
data['set_id'] = set_id
data['user_race'] = user_race
data['name'] = name
yield idx, data