Vadzim Kashko commited on
Commit
88677fe
1 Parent(s): 66daa11

feat: csript

Browse files
Files changed (1) hide show
  1. speech-emotion-recognition-dataset.py +147 -0
speech-emotion-recognition-dataset.py ADDED
@@ -0,0 +1,147 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import datasets
2
+ import PIL.Image
3
+ import PIL.ImageOps
4
+ import numpy as np
5
+
6
+ _CITATION = """\
7
+ @InProceedings{huggingface:dataset,
8
+ title = {generated-usa-passeports-dataset},
9
+ author = {TrainingDataPro},
10
+ year = {2023}
11
+ }
12
+ """
13
+
14
+ _DESCRIPTION = """\
15
+ Data generation in machine learning involves creating or manipulating data
16
+ to train and evaluate machine learning models. The purpose of data generation
17
+ is to provide diverse and representative examples that cover a wide range of
18
+ scenarios, ensuring the model's robustness and generalization.
19
+ Data augmentation techniques involve applying various transformations to
20
+ existing data samples to create new ones. These transformations include:
21
+ random rotations, translations, scaling, flips, and more. Augmentation helps
22
+ in increasing the dataset size, introducing natural variations, and improving
23
+ model performance by making it more invariant to specific transformations.
24
+ The dataset contains **GENERATED** USA passports, which are replicas of
25
+ official passports but with randomly generated details, such as name, date of
26
+ birth etc. The primary intention of generating these fake passports is to
27
+ demonstrate the structure and content of a typical passport document and to
28
+ train the neural network to identify this type of document.
29
+ Generated passports can assist in conducting research without accessing or
30
+ compromising real user data that is often sensitive and subject to privacy
31
+ regulations. Synthetic data generation allows researchers to develop and
32
+ refine models using simulated passport data without risking privacy leaks.
33
+ """
34
+ _NAME = 'generated-usa-passeports-dataset'
35
+
36
+ _HOMEPAGE = f"https://huggingface.co/datasets/TrainingDataPro/{_NAME}"
37
+
38
+ _LICENSE = "cc-by-nc-nd-4.0"
39
+
40
+ _DATA = f"https://huggingface.co/datasets/TrainingDataPro/{_NAME}/resolve/main/data/"
41
+
42
+
43
+ def exif_transpose(img):
44
+ if not img:
45
+ return img
46
+
47
+ exif_orientation_tag = 274
48
+
49
+ # Check for EXIF data (only present on some files)
50
+ if hasattr(img, "_getexif") and isinstance(
51
+ img._getexif(), dict) and exif_orientation_tag in img._getexif():
52
+ exif_data = img._getexif()
53
+ orientation = exif_data[exif_orientation_tag]
54
+
55
+ # Handle EXIF Orientation
56
+ if orientation == 1:
57
+ # Normal image - nothing to do!
58
+ pass
59
+ elif orientation == 2:
60
+ # Mirrored left to right
61
+ img = img.transpose(PIL.Image.FLIP_LEFT_RIGHT)
62
+ elif orientation == 3:
63
+ # Rotated 180 degrees
64
+ img = img.rotate(180)
65
+ elif orientation == 4:
66
+ # Mirrored top to bottom
67
+ img = img.rotate(180).transpose(PIL.Image.FLIP_LEFT_RIGHT)
68
+ elif orientation == 5:
69
+ # Mirrored along top-left diagonal
70
+ img = img.rotate(-90,
71
+ expand=True).transpose(PIL.Image.FLIP_LEFT_RIGHT)
72
+ elif orientation == 6:
73
+ # Rotated 90 degrees
74
+ img = img.rotate(-90, expand=True)
75
+ elif orientation == 7:
76
+ # Mirrored along top-right diagonal
77
+ img = img.rotate(90,
78
+ expand=True).transpose(PIL.Image.FLIP_LEFT_RIGHT)
79
+ elif orientation == 8:
80
+ # Rotated 270 degrees
81
+ img = img.rotate(90, expand=True)
82
+
83
+ return img
84
+
85
+
86
+ def load_image_file(file, mode='RGB'):
87
+ # Load the image with PIL
88
+ img = PIL.Image.open(file)
89
+
90
+ if hasattr(PIL.ImageOps, 'exif_transpose'):
91
+ # Very recent versions of PIL can do exit transpose internally
92
+ img = PIL.ImageOps.exif_transpose(img)
93
+ else:
94
+ # Otherwise, do the exif transpose ourselves
95
+ img = exif_transpose(img)
96
+
97
+ img = img.convert(mode)
98
+
99
+ return np.array(img)
100
+
101
+
102
+ class GeneratedUsaPasseportsDataset(datasets.GeneratorBasedBuilder):
103
+
104
+ def _info(self):
105
+ return datasets.DatasetInfo(
106
+ description=_DESCRIPTION,
107
+ features=datasets.Features({
108
+ 'original': datasets.Image(),
109
+ 'us_pass_augmentated_1': datasets.Image(),
110
+ 'us_pass_augmentated_2': datasets.Image(),
111
+ 'us_pass_augmentated_3': datasets.Image()
112
+ }),
113
+ supervised_keys=None,
114
+ homepage=_HOMEPAGE,
115
+ citation=_CITATION,
116
+ license=_LICENSE)
117
+
118
+ def _split_generators(self, dl_manager):
119
+ original = dl_manager.download_and_extract(f"{_DATA}original.zip")
120
+ augmentation = dl_manager.download_and_extract(
121
+ f"{_DATA}augmentation.zip")
122
+ annotations = dl_manager.download(f"{_DATA}{_NAME}.csv")
123
+ original = dl_manager.iter_files(original)
124
+ augmentation = dl_manager.iter_files(augmentation)
125
+ return [
126
+ datasets.SplitGenerator(name=datasets.Split.TRAIN,
127
+ gen_kwargs={
128
+ "original": original,
129
+ 'augmentation': augmentation,
130
+ 'annotations': annotations
131
+ }),
132
+ ]
133
+
134
+ def _generate_examples(self, original, augmentation, annotations):
135
+ original = list(original)
136
+ augmentation = list(augmentation)
137
+ augmentation = [
138
+ augmentation[i:i + 3] for i in range(0, len(augmentation), 3)
139
+ ]
140
+
141
+ for idx, (org, aug) in enumerate(zip(original, augmentation)):
142
+ yield idx, {
143
+ 'original': load_image_file(org),
144
+ 'us_pass_augmentated_1': load_image_file(aug[0]),
145
+ 'us_pass_augmentated_2': load_image_file(aug[1]),
146
+ 'us_pass_augmentated_3': load_image_file(aug[2])
147
+ }