|
--- |
|
dataset_info: |
|
features: |
|
- name: audio |
|
dtype: |
|
audio: |
|
sampling_rate: 16000 |
|
- name: transcription |
|
dtype: string |
|
- name: duration |
|
dtype: float32 |
|
- name: up_votes |
|
dtype: int32 |
|
- name: down_votes |
|
dtype: int32 |
|
- name: age |
|
dtype: string |
|
- name: gender |
|
dtype: string |
|
- name: accent |
|
dtype: string |
|
splits: |
|
- name: train |
|
num_bytes: 249774324 |
|
num_examples: 26501 |
|
- name: test |
|
num_bytes: 90296575 |
|
num_examples: 9650 |
|
- name: validation |
|
num_bytes: 78834938 |
|
num_examples: 8639 |
|
- name: validated |
|
num_bytes: 412113612 |
|
num_examples: 46345 |
|
download_size: 818561949 |
|
dataset_size: 831019449 |
|
configs: |
|
- config_name: default |
|
data_files: |
|
- split: train |
|
path: data/train-* |
|
- split: test |
|
path: data/test-* |
|
- split: validation |
|
path: data/validation-* |
|
- split: validated |
|
path: data/validated-* |
|
license: cc0-1.0 |
|
task_categories: |
|
- automatic-speech-recognition |
|
language: |
|
- tr |
|
--- |
|
|
|
# Improving CommonVoice 17 Turkish Dataset |
|
|
|
I recently worked on enhancing the Mozilla CommonVoice 17 Turkish dataset to create a higher quality training set for speech recognition models. |
|
Here's an overview of my process and findings. |
|
|
|
## Initial Analysis and Split Organization |
|
|
|
My first step was analyzing the dataset organization to understand its structure. |
|
Through analysis of filename stems as unique keys, I revealed and documented an important aspect of CommonVoice's design that might not be immediately clear to all users: |
|
|
|
- The validated set (113,699 total files) completely contained all samples from: |
|
- Train split (35,035 files) |
|
- Test split (11,290 files) |
|
- Validation split (11,247 files) |
|
- Additionally, the validated set had ~56K unique samples not present in any other split |
|
|
|
This design follows CommonVoice's documentation, where dev/test/train are carefully reviewed subsets of the validated data. |
|
However, this structure needs to be clearly understood to avoid potential data leakage when working with the dataset. |
|
For example, using the validated set for training while evaluating on the test split would be problematic since the test data is already included in the validated set. |
|
|
|
To create a clean dataset without overlaps, I: |
|
|
|
1. Identified all overlapping samples using filename stems as unique keys |
|
2. Removed samples that were already in train/test/validation splits from the validated set |
|
3. Created a clean, non-overlapping validated split with unique samples only |
|
|
|
This approach ensures that researchers can either: |
|
- Use the original train/test/dev splits as curated by CommonVoice, OR |
|
- Use my cleaned validated set with their own custom splits |
|
|
|
Both approaches are valid, but mixing them could lead to evaluation issues. |
|
|
|
## Audio Processing and Quality Improvements |
|
|
|
### Audio Resampling |
|
All audio files were resampled to 16 kHz to: |
|
- Make the dataset directly compatible with Whisper and similar models |
|
- Eliminate the need for runtime resampling during training |
|
- Ensure consistent audio quality across the dataset |
|
|
|
### Silence Trimming |
|
I processed all audio files to remove unnecessary silence and noise: |
|
- Used Silero VAD with a threshold of 0.6 to detect speech segments |
|
- Trimmed leading and trailing silences |
|
- Removed microphone noise and clicks at clip boundaries |
|
|
|
### Duration Filtering and Analysis |
|
|
|
I analyzed each split separately after trimming silences. Here are the detailed findings per split: |
|
|
|
| Split | Files Before | Files After | Short Files | Duration Before (hrs) | Duration After (hrs) | Duration Reduction % | Short Files Duration (hrs) | Files Reduction % | |
|
|---|--:|--:|--:|--:|--:|--:|--:|--:| |
|
| Train | 11,290 | 9,651 | 1,626 | 13.01 | 7.34 | 43.6% | 0.37 | 14.5% | |
|
| Validation | 11,247 | 8,640 | 2,609 | 11.17 | 6.27 | 43.9% | 0.60 | 23.2% | |
|
| Test | 35,035 | 26,501 | 8,633 | 35.49 | 19.84 | 44.1% | 2.00 | 24.4% | |
|
| Validated | 56,127 | 46,348 | 9,991 | 56.71 | 32.69 | 42.4% | 2.29 | 17.4% | |
|
| **Total** | **113,699** | **91,140** | **22,859** | **116.38** | **66.14** | **43.2%** | **5.26** | **19.8%** | |
|
|
|
Note: Files with duration shorter than 1.0 seconds were removed from the dataset. |
|
|
|
#### Validation Split Analysis (formerly Eval) |
|
- Original files: 11,247 |
|
- Found 2,609 files shorter than 1.0s |
|
- Statistics for short files: |
|
- Total duration: 26.26 minutes |
|
- Average duration: 0.83 seconds |
|
- Shortest file: 0.65 seconds |
|
- Longest file: 0.97 seconds |
|
|
|
#### Train Split Analysis |
|
- Original files: 35,035 |
|
- Found 8,633 files shorter than 1.0s |
|
- Statistics for short files: |
|
- Total duration: 2.29 hours |
|
- Average duration: 0.82 seconds |
|
- Shortest file: 0.08 seconds |
|
- Longest file: 0.97 seconds |
|
|
|
#### Test Split Analysis |
|
- Original files: 11,290 |
|
- Found 1,626 files shorter than 1.0s |
|
- Statistics for short files: |
|
- Total duration: 56.26 minutes |
|
- Average duration: 0.85 seconds |
|
- Shortest file: 0.65 seconds |
|
- Longest file: 0.97 seconds |
|
|
|
#### Validated Split Analysis |
|
- Original files: 56,127 |
|
- Found 9,991 files shorter than 1.0s |
|
- Statistics for short files: |
|
- Total duration: 36.26 minutes |
|
- Average duration: 0.83 seconds |
|
- Shortest file: 0.65 seconds |
|
- Longest file: 0.97 seconds |
|
|
|
All short clips were removed from the dataset to ensure consistent quality. The final dataset maintains only clips longer than 1.0 seconds, with average durations between 2.54-2.69 seconds across splits. |
|
|
|
### Final Split Statistics |
|
The cleaned dataset was organized into: |
|
- Train: 26,501 files (19.84 hours, avg duration: 2.69s, min: 1.04s, max: 9.58s) |
|
- Test: 9,650 files (7.33 hours, avg duration: 2.74s, min: 1.08s, max: 9.29s) |
|
- Validation: 8,639 files (6.27 hours, avg duration: 2.61s, min: 1.04s, max: 9.18s) |
|
- Validated: 46,345 files (32.69 hours, avg duration: 2.54s, min: 1.04s, max: 9.07s) |
|
|
|
### Final Dataset Split Metrics |
|
|
|
| Split | Files | Duration (hours) | Avg Duration (s) | Min Duration (s) | Max Duration (s) | |
|
|-------------|--------|------------------|------------------|------------------|------------------| |
|
| TRAIN | 26501 | 19.84 | 2.69 | 1.04 | 9.58 | |
|
| TEST | 9650 | 7.33 | 2.74 | 1.08 | 9.29 | |
|
| VALIDATION | 8639 | 6.27 | 2.61 | 1.04 | 9.18 | |
|
| VALIDATED | 46345 | 32.69 | 2.54 | 1.04 | 9.07 | |
|
|
|
Total files processed: 91,135 |
|
Valid entries created: 91,135 |
|
Files skipped: 0 |
|
Total dataset duration: 66.13 hours |
|
Average duration across all splits: 2.61 seconds |
|
|
|
The dataset was processed in the following order: |
|
1. Train split (26,501 files) |
|
2. Test split (9,650 files) |
|
3. Validation split (8,639 files) - Note: Also known as "eval" split in some CommonVoice versions |
|
4. Validated split (46,348 files) |
|
|
|
Note: The validation split (sometimes referred to as "eval" split in CommonVoice documentation) serves the same purpose - it's a held-out set for model validation during training. |
|
We've standardized the naming to "validation" throughout this documentation for consistency with common machine learning terminology. |
|
|
|
One text file in the validated split was flagged for being too short (2 characters), but was still included in the final dataset. |
|
|
|
The processed dataset was saved as 'commonvoice_17_tr_fixed'. |
|
|
|
## Text Processing and Standardization |
|
|
|
### Character Set Optimization |
|
- Created a comprehensive charset from all text labels |
|
- Simplified the character set by: |
|
- Standardizing quotation marks |
|
- Removing infrequently used special characters |
|
|
|
### Text Quality Improvements |
|
- Generated word frequency metrics to identify potential issues |
|
- Corrected common Turkish typos and grammar errors |
|
- Standardized punctuation and spacing |
|
|
|
## Results |
|
|
|
The final dataset shows significant improvements: |
|
- Removed unnecessary silence and noise from audio |
|
- Consistent audio durations above 1.0 seconds |
|
- Standardized text with corrected Turkish grammar and typography |
|
- Maintained original metadata (age, upvotes, etc.) |
|
|
|
These improvements make the dataset more suitable for training speech recognition models while maintaining the diversity and richness of the original CommonVoice collection. |
|
|
|
## Tools Used |
|
|
|
This dataset processing work was completed using [ASRTK (Automatic Speech Recognition Toolkit)](https://github.com/ysdede/asrtk), an open-source Python toolkit designed to streamline the development and enhancement of ASR systems. ASRTK provides utilities for: |
|
|
|
- Audio processing with advanced splitting and resampling capabilities |
|
- Text normalization and cleaning |
|
- Forced alignment using Silero VAD models |
|
- Efficient batch processing with multi-threading support |
|
|
|
The toolkit is available under the MIT license and welcomes contributions from the community. |
|
|