Datasets:
Enrich card with TTS tag
#1
by
ylacombe
HF staff
- opened
README.md
CHANGED
@@ -17,6 +17,8 @@ source_datasets:
|
|
17 |
- original
|
18 |
task_categories:
|
19 |
- automatic-speech-recognition
|
|
|
|
|
20 |
task_ids: []
|
21 |
train-eval-index:
|
22 |
- config: main
|
@@ -98,9 +100,9 @@ The texts were published between 1884 and 1964, and are in the public domain. Th
|
|
98 |
### Supported Tasks and Leaderboards
|
99 |
|
100 |
The dataset can be used to train a model for Automatic Speech Recognition (ASR) or Text-to-Speech (TTS).
|
101 |
-
- `
|
102 |
The most common ASR evaluation metric is the word error rate (WER).
|
103 |
-
- `
|
104 |
A reasonable evaluation metric is the mean opinion score (MOS) of audio quality.
|
105 |
The dataset has an active leaderboard which can be found at https://paperswithcode.com/sota/text-to-speech-synthesis-on-ljspeech
|
106 |
|
|
|
17 |
- original
|
18 |
task_categories:
|
19 |
- automatic-speech-recognition
|
20 |
+
- text-to-speech
|
21 |
+
- text-to-audio
|
22 |
task_ids: []
|
23 |
train-eval-index:
|
24 |
- config: main
|
|
|
100 |
### Supported Tasks and Leaderboards
|
101 |
|
102 |
The dataset can be used to train a model for Automatic Speech Recognition (ASR) or Text-to-Speech (TTS).
|
103 |
+
- `automatic-speech-recognition`: An ASR model is presented with an audio file and asked to transcribe the audio file to written text.
|
104 |
The most common ASR evaluation metric is the word error rate (WER).
|
105 |
+
- `text-to-speech`, `text-to-audio`: A TTS model is given a written text in natural language and asked to generate a speech audio file.
|
106 |
A reasonable evaluation metric is the mean opinion score (MOS) of audio quality.
|
107 |
The dataset has an active leaderboard which can be found at https://paperswithcode.com/sota/text-to-speech-synthesis-on-ljspeech
|
108 |
|