Datasets:
carlosdanielhernandezmena
commited on
Commit
•
e48eac0
1
Parent(s):
d7cd368
Convert dataset to Parquet (#1)
Browse files- Convert dataset to Parquet (a2167d3bb5cfc104c770e1cb6eee8478100a8a92)
- Delete data file (7e09079fa369265df979818c84462636f564ae35)
- Delete data file (1183983d2949ed82c085a158c41786de4d008dff)
- Delete loading script (d114e20b7e23027fb02e173f6f387b1a80afe565)
- Delete data file (bd5964ebdaa4b063f77393fa105e4dddc584d5e1)
- README.md +36 -6
- ciempiess_balance.py +0 -123
- corpus/speech/train.tar.gz → ciempiess_balance/train-00000-of-00003.parquet +2 -2
- ciempiess_balance/train-00001-of-00003.parquet +3 -0
- ciempiess_balance/train-00002-of-00003.parquet +3 -0
- corpus/files/metadata_train.tsv +0 -0
- corpus/files/tars_train.paths +0 -1
README.md
CHANGED
@@ -1,28 +1,58 @@
|
|
1 |
---
|
2 |
annotations_creators:
|
3 |
- expert-generated
|
4 |
-
language:
|
5 |
-
- es
|
6 |
language_creators:
|
7 |
- other
|
|
|
|
|
8 |
license:
|
9 |
- cc-by-sa-4.0
|
10 |
multilinguality:
|
11 |
- monolingual
|
12 |
-
pretty_name: 'CIEMPIESS BALANCE CORPUS: Audio and Transcripts of Mexican Spanish Broadcast Conversations.'
|
13 |
size_categories:
|
14 |
- 10K<n<100K
|
15 |
source_datasets:
|
16 |
- original
|
|
|
|
|
|
|
|
|
|
|
17 |
tags:
|
18 |
- ciempiess
|
19 |
- spanish
|
20 |
- mexican spanish
|
21 |
- ciempiess project
|
22 |
- ciempiess-unam project
|
23 |
-
|
24 |
-
|
25 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
26 |
---
|
27 |
|
28 |
# Dataset Card for ciempiess_balance
|
|
|
1 |
---
|
2 |
annotations_creators:
|
3 |
- expert-generated
|
|
|
|
|
4 |
language_creators:
|
5 |
- other
|
6 |
+
language:
|
7 |
+
- es
|
8 |
license:
|
9 |
- cc-by-sa-4.0
|
10 |
multilinguality:
|
11 |
- monolingual
|
|
|
12 |
size_categories:
|
13 |
- 10K<n<100K
|
14 |
source_datasets:
|
15 |
- original
|
16 |
+
task_categories:
|
17 |
+
- automatic-speech-recognition
|
18 |
+
task_ids: []
|
19 |
+
pretty_name: 'CIEMPIESS BALANCE CORPUS: Audio and Transcripts of Mexican Spanish Broadcast
|
20 |
+
Conversations.'
|
21 |
tags:
|
22 |
- ciempiess
|
23 |
- spanish
|
24 |
- mexican spanish
|
25 |
- ciempiess project
|
26 |
- ciempiess-unam project
|
27 |
+
dataset_info:
|
28 |
+
config_name: ciempiess_balance
|
29 |
+
features:
|
30 |
+
- name: audio_id
|
31 |
+
dtype: string
|
32 |
+
- name: audio
|
33 |
+
dtype:
|
34 |
+
audio:
|
35 |
+
sampling_rate: 16000
|
36 |
+
- name: speaker_id
|
37 |
+
dtype: string
|
38 |
+
- name: gender
|
39 |
+
dtype: string
|
40 |
+
- name: duration
|
41 |
+
dtype: float32
|
42 |
+
- name: normalized_text
|
43 |
+
dtype: string
|
44 |
+
splits:
|
45 |
+
- name: train
|
46 |
+
num_bytes: 1125593189.655
|
47 |
+
num_examples: 8555
|
48 |
+
download_size: 1295436525
|
49 |
+
dataset_size: 1125593189.655
|
50 |
+
configs:
|
51 |
+
- config_name: ciempiess_balance
|
52 |
+
data_files:
|
53 |
+
- split: train
|
54 |
+
path: ciempiess_balance/train-*
|
55 |
+
default: true
|
56 |
---
|
57 |
|
58 |
# Dataset Card for ciempiess_balance
|
ciempiess_balance.py
DELETED
@@ -1,123 +0,0 @@
|
|
1 |
-
from collections import defaultdict
|
2 |
-
import os
|
3 |
-
import json
|
4 |
-
import csv
|
5 |
-
|
6 |
-
import datasets
|
7 |
-
|
8 |
-
_NAME="ciempiess_balance"
|
9 |
-
_VERSION="1.0.0"
|
10 |
-
_AUDIO_EXTENSIONS=".flac"
|
11 |
-
|
12 |
-
_DESCRIPTION = """
|
13 |
-
CIEMPIESS BALANCE is a Radio Corpus designed to create acoustic models for automatic speech recognition. It is "balance" because it is designed to balance the CIEMPIESS LIGHT, which means that if both corpora are combined, one will get a gender balanced corpus.
|
14 |
-
"""
|
15 |
-
|
16 |
-
_CITATION = """
|
17 |
-
@misc{carlosmenaciempiessbalance2018,
|
18 |
-
title={CIEMPIESS BALANCE CORPUS: Audio and Transcripts of Mexican Spanish Broadcast Conversations.},
|
19 |
-
ldc_catalog_no={LDC2018S11},
|
20 |
-
DOI={https://doi.org/10.35111/rfmw-n126},
|
21 |
-
author={Hernandez Mena, Carlos Daniel},
|
22 |
-
journal={Linguistic Data Consortium, Philadelphia},
|
23 |
-
year={2018},
|
24 |
-
url={https://catalog.ldc.upenn.edu/LDC2018S11},
|
25 |
-
}
|
26 |
-
"""
|
27 |
-
|
28 |
-
_HOMEPAGE = "https://catalog.ldc.upenn.edu/LDC2018S11"
|
29 |
-
|
30 |
-
_LICENSE = "CC-BY-SA-4.0, See https://creativecommons.org/licenses/by-sa/4.0/"
|
31 |
-
|
32 |
-
_BASE_DATA_DIR = "corpus/"
|
33 |
-
_METADATA_TRAIN = os.path.join(_BASE_DATA_DIR,"files", "metadata_train.tsv")
|
34 |
-
|
35 |
-
_TARS_TRAIN = os.path.join(_BASE_DATA_DIR,"files", "tars_train.paths")
|
36 |
-
|
37 |
-
class CiempiessBalanceConfig(datasets.BuilderConfig):
|
38 |
-
"""BuilderConfig for CIEMPIESS BALANCE Corpus"""
|
39 |
-
|
40 |
-
def __init__(self, name, **kwargs):
|
41 |
-
name=_NAME
|
42 |
-
super().__init__(name=name, **kwargs)
|
43 |
-
|
44 |
-
class CiempiessBalance(datasets.GeneratorBasedBuilder):
|
45 |
-
"""CIEMPIESS BALANCE Corpus"""
|
46 |
-
|
47 |
-
VERSION = datasets.Version(_VERSION)
|
48 |
-
BUILDER_CONFIGS = [
|
49 |
-
CiempiessBalanceConfig(
|
50 |
-
name=_NAME,
|
51 |
-
version=datasets.Version(_VERSION),
|
52 |
-
)
|
53 |
-
]
|
54 |
-
|
55 |
-
def _info(self):
|
56 |
-
features = datasets.Features(
|
57 |
-
{
|
58 |
-
"audio_id": datasets.Value("string"),
|
59 |
-
"audio": datasets.Audio(sampling_rate=16000),
|
60 |
-
"speaker_id": datasets.Value("string"),
|
61 |
-
"gender": datasets.Value("string"),
|
62 |
-
"duration": datasets.Value("float32"),
|
63 |
-
"normalized_text": datasets.Value("string"),
|
64 |
-
}
|
65 |
-
)
|
66 |
-
return datasets.DatasetInfo(
|
67 |
-
description=_DESCRIPTION,
|
68 |
-
features=features,
|
69 |
-
homepage=_HOMEPAGE,
|
70 |
-
license=_LICENSE,
|
71 |
-
citation=_CITATION,
|
72 |
-
)
|
73 |
-
|
74 |
-
def _split_generators(self, dl_manager):
|
75 |
-
|
76 |
-
metadata_train=dl_manager.download_and_extract(_METADATA_TRAIN)
|
77 |
-
|
78 |
-
tars_train=dl_manager.download_and_extract(_TARS_TRAIN)
|
79 |
-
|
80 |
-
hash_tar_files=defaultdict(dict)
|
81 |
-
|
82 |
-
with open(tars_train,'r') as f:
|
83 |
-
hash_tar_files['train']=[path.replace('\n','') for path in f]
|
84 |
-
|
85 |
-
hash_meta_paths={"train":metadata_train}
|
86 |
-
audio_paths = dl_manager.download(hash_tar_files)
|
87 |
-
|
88 |
-
splits=["train"]
|
89 |
-
local_extracted_audio_paths = (
|
90 |
-
dl_manager.extract(audio_paths) if not dl_manager.is_streaming else
|
91 |
-
{
|
92 |
-
split:[None] * len(audio_paths[split]) for split in splits
|
93 |
-
}
|
94 |
-
)
|
95 |
-
|
96 |
-
return [
|
97 |
-
datasets.SplitGenerator(
|
98 |
-
name=datasets.Split.TRAIN,
|
99 |
-
gen_kwargs={
|
100 |
-
"audio_archives": [dl_manager.iter_archive(archive) for archive in audio_paths["train"]],
|
101 |
-
"local_extracted_archives_paths": local_extracted_audio_paths["train"],
|
102 |
-
"metadata_paths": hash_meta_paths["train"],
|
103 |
-
}
|
104 |
-
),
|
105 |
-
]
|
106 |
-
|
107 |
-
def _generate_examples(self, audio_archives, local_extracted_archives_paths, metadata_paths):
|
108 |
-
|
109 |
-
features = ["speaker_id","gender","duration","normalized_text"]
|
110 |
-
|
111 |
-
with open(metadata_paths) as f:
|
112 |
-
metadata = {x["audio_id"]: x for x in csv.DictReader(f, delimiter="\t")}
|
113 |
-
|
114 |
-
for audio_archive, local_extracted_archive_path in zip(audio_archives, local_extracted_archives_paths):
|
115 |
-
for audio_filename, audio_file in audio_archive:
|
116 |
-
audio_id = audio_filename.split(os.sep)[-1].split(_AUDIO_EXTENSIONS)[0]
|
117 |
-
path = os.path.join(local_extracted_archive_path, audio_filename) if local_extracted_archive_path else audio_filename
|
118 |
-
|
119 |
-
yield audio_id, {
|
120 |
-
"audio_id": audio_id,
|
121 |
-
**{feature: metadata[audio_id][feature] for feature in features},
|
122 |
-
"audio": {"path": path, "bytes": audio_file.read()},
|
123 |
-
}
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
corpus/speech/train.tar.gz → ciempiess_balance/train-00000-of-00003.parquet
RENAMED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ce15ebda86cd7483d88bc0b864874d17b16c102c81e3a0d7abdd9a2a4954ade6
|
3 |
+
size 392456833
|
ciempiess_balance/train-00001-of-00003.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:211aad4adc4d96e458fff376c2888d7f6055a377b0d34fd8b9fef37b753d1cf3
|
3 |
+
size 440405225
|
ciempiess_balance/train-00002-of-00003.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d81a5057a442ddbfa074fa86aa93dea92f3b8bc72c408e3bc2cf1853b659aab1
|
3 |
+
size 462574467
|
corpus/files/metadata_train.tsv
DELETED
The diff for this file is too large to render.
See raw diff
|
|
corpus/files/tars_train.paths
DELETED
@@ -1 +0,0 @@
|
|
1 |
-
corpus/speech/train.tar.gz
|
|
|
|