Datasets:
File size: 8,811 Bytes
91d859b 10316ef 91d859b 10316ef 91d859b 10316ef 91d859b 10316ef 91d859b 10316ef 91d859b 10316ef 91d859b df7f256 9d8a025 df7f256 1067bdf df7f256 1067bdf 9d8a025 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 163 164 165 166 167 168 169 170 171 172 173 174 175 176 177 178 179 180 181 182 183 184 185 186 187 188 189 190 191 192 193 194 195 196 197 198 199 200 201 202 203 204 205 206 207 208 209 210 211 212 213 214 215 216 217 218 219 220 221 |
---
dataset_info:
features:
- name: audio
dtype:
audio:
sampling_rate: 16000
- name: transcription
dtype: string
- name: duration
dtype: float32
- name: up_votes
dtype: int32
- name: down_votes
dtype: int32
- name: age
dtype: string
- name: gender
dtype: string
- name: accent
dtype: string
splits:
- name: train
num_bytes: 249774324
num_examples: 26501
- name: test
num_bytes: 90296575
num_examples: 9650
- name: validation
num_bytes: 78834938
num_examples: 8639
- name: validated
num_bytes: 412113612
num_examples: 46345
download_size: 818561949
dataset_size: 831019449
configs:
- config_name: default
data_files:
- split: train
path: data/train-*
- split: test
path: data/test-*
- split: validation
path: data/validation-*
- split: validated
path: data/validated-*
license: cc0-1.0
task_categories:
- automatic-speech-recognition
language:
- tr
---
# Improving CommonVoice 17 Turkish Dataset
I recently worked on enhancing the Mozilla CommonVoice 17 Turkish dataset to create a higher quality training set for speech recognition models.
Here's an overview of my process and findings.
## Initial Analysis and Split Organization
My first step was analyzing the dataset organization to understand its structure.
Through analysis of filename stems as unique keys, I revealed and documented an important aspect of CommonVoice's design that might not be immediately clear to all users:
- The validated set (113,699 total files) completely contained all samples from:
- Train split (35,035 files)
- Test split (11,290 files)
- Validation split (11,247 files)
- Additionally, the validated set had ~56K unique samples not present in any other split
This design follows CommonVoice's documentation, where dev/test/train are carefully reviewed subsets of the validated data.
However, this structure needs to be clearly understood to avoid potential data leakage when working with the dataset.
For example, using the validated set for training while evaluating on the test split would be problematic since the test data is already included in the validated set.
To create a clean dataset without overlaps, I:
1. Identified all overlapping samples using filename stems as unique keys
2. Removed samples that were already in train/test/validation splits from the validated set
3. Created a clean, non-overlapping validated split with unique samples only
This approach ensures that researchers can either:
- Use the original train/test/dev splits as curated by CommonVoice, OR
- Use my cleaned validated set with their own custom splits
Both approaches are valid, but mixing them could lead to evaluation issues.
## Audio Processing and Quality Improvements
### Audio Resampling
All audio files were resampled to 16 kHz to:
- Make the dataset directly compatible with Whisper and similar models
- Eliminate the need for runtime resampling during training
- Ensure consistent audio quality across the dataset
### Silence Trimming
I processed all audio files to remove unnecessary silence and noise:
- Used Silero VAD with a threshold of 0.6 to detect speech segments
- Trimmed leading and trailing silences
- Removed microphone noise and clicks at clip boundaries
### Duration Filtering and Analysis
I analyzed each split separately after trimming silences. Here are the detailed findings per split:
| Split | Files Before | Files After | Short Files | Duration Before (hrs) | Duration After (hrs) | Duration Reduction % | Short Files Duration (hrs) | Files Reduction % |
|---|--:|--:|--:|--:|--:|--:|--:|--:|
| Train | 11,290 | 9,651 | 1,626 | 13.01 | 7.34 | 43.6% | 0.37 | 14.5% |
| Validation | 11,247 | 8,640 | 2,609 | 11.17 | 6.27 | 43.9% | 0.60 | 23.2% |
| Test | 35,035 | 26,501 | 8,633 | 35.49 | 19.84 | 44.1% | 2.00 | 24.4% |
| Validated | 56,127 | 46,348 | 9,991 | 56.71 | 32.69 | 42.4% | 2.29 | 17.4% |
| **Total** | **113,699** | **91,140** | **22,859** | **116.38** | **66.14** | **43.2%** | **5.26** | **19.8%** |
Note: Files with duration shorter than 1.0 seconds were removed from the dataset.
#### Validation Split Analysis (formerly Eval)
- Original files: 11,247
- Found 2,609 files shorter than 1.0s
- Statistics for short files:
- Total duration: 26.26 minutes
- Average duration: 0.83 seconds
- Shortest file: 0.65 seconds
- Longest file: 0.97 seconds
#### Train Split Analysis
- Original files: 35,035
- Found 8,633 files shorter than 1.0s
- Statistics for short files:
- Total duration: 2.29 hours
- Average duration: 0.82 seconds
- Shortest file: 0.08 seconds
- Longest file: 0.97 seconds
#### Test Split Analysis
- Original files: 11,290
- Found 1,626 files shorter than 1.0s
- Statistics for short files:
- Total duration: 56.26 minutes
- Average duration: 0.85 seconds
- Shortest file: 0.65 seconds
- Longest file: 0.97 seconds
#### Validated Split Analysis
- Original files: 56,127
- Found 9,991 files shorter than 1.0s
- Statistics for short files:
- Total duration: 36.26 minutes
- Average duration: 0.83 seconds
- Shortest file: 0.65 seconds
- Longest file: 0.97 seconds
All short clips were removed from the dataset to ensure consistent quality. The final dataset maintains only clips longer than 1.0 seconds, with average durations between 2.54-2.69 seconds across splits.
### Final Split Statistics
The cleaned dataset was organized into:
- Train: 26,501 files (19.84 hours, avg duration: 2.69s, min: 1.04s, max: 9.58s)
- Test: 9,650 files (7.33 hours, avg duration: 2.74s, min: 1.08s, max: 9.29s)
- Validation: 8,639 files (6.27 hours, avg duration: 2.61s, min: 1.04s, max: 9.18s)
- Validated: 46,345 files (32.69 hours, avg duration: 2.54s, min: 1.04s, max: 9.07s)
### Final Dataset Split Metrics
| Split | Files | Duration (hours) | Avg Duration (s) | Min Duration (s) | Max Duration (s) |
|-------------|--------|------------------|------------------|------------------|------------------|
| TRAIN | 26501 | 19.84 | 2.69 | 1.04 | 9.58 |
| TEST | 9650 | 7.33 | 2.74 | 1.08 | 9.29 |
| VALIDATION | 8639 | 6.27 | 2.61 | 1.04 | 9.18 |
| VALIDATED | 46345 | 32.69 | 2.54 | 1.04 | 9.07 |
Total files processed: 91,135
Valid entries created: 91,135
Files skipped: 0
Total dataset duration: 66.13 hours
Average duration across all splits: 2.61 seconds
The dataset was processed in the following order:
1. Train split (26,501 files)
2. Test split (9,650 files)
3. Validation split (8,639 files) - Note: Also known as "eval" split in some CommonVoice versions
4. Validated split (46,348 files)
Note: The validation split (sometimes referred to as "eval" split in CommonVoice documentation) serves the same purpose - it's a held-out set for model validation during training.
We've standardized the naming to "validation" throughout this documentation for consistency with common machine learning terminology.
One text file in the validated split was flagged for being too short (2 characters), but was still included in the final dataset.
The processed dataset was saved as 'commonvoice_17_tr_fixed'.
## Text Processing and Standardization
### Character Set Optimization
- Created a comprehensive charset from all text labels
- Simplified the character set by:
- Standardizing quotation marks
- Removing infrequently used special characters
### Text Quality Improvements
- Generated word frequency metrics to identify potential issues
- Corrected common Turkish typos and grammar errors
- Standardized punctuation and spacing
## Results
The final dataset shows significant improvements:
- Removed unnecessary silence and noise from audio
- Consistent audio durations above 1.0 seconds
- Standardized text with corrected Turkish grammar and typography
- Maintained original metadata (age, upvotes, etc.)
These improvements make the dataset more suitable for training speech recognition models while maintaining the diversity and richness of the original CommonVoice collection.
## Tools Used
This dataset processing work was completed using [ASRTK (Automatic Speech Recognition Toolkit)](https://github.com/ysdede/asrtk), an open-source Python toolkit designed to streamline the development and enhancement of ASR systems. ASRTK provides utilities for:
- Audio processing with advanced splitting and resampling capabilities
- Text normalization and cleaning
- Forced alignment using Silero VAD models
- Efficient batch processing with multi-threading support
The toolkit is available under the MIT license and welcomes contributions from the community.
|