nithinraok commited on
Commit
f907741
1 Parent(s): e0aba95

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -241,7 +241,7 @@ The tokenizers for these models were built using the text transcripts of the tra
241
 
242
  The model was trained on 65K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
243
 
244
- Dataset contains following Public English speech sets (25K hours)
245
 
246
  - Librispeech 960 hours of English speech
247
  - Fisher Corpus
 
241
 
242
  The model was trained on 65K hours of English speech collected and prepared by NVIDIA NeMo and Suno teams.
243
 
244
+ The training dataset consists of private subset with 40K hours of English speech plus 25K hours from the following public datasets:
245
 
246
  - Librispeech 960 hours of English speech
247
  - Fisher Corpus