Update README.md
Browse files
README.md
CHANGED
@@ -31,8 +31,7 @@ model-index:
|
|
31 |
|
32 |
# Wav2Vec2-Large-XLSR-53-ml
|
33 |
|
34 |
-
Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on ml using the [Indic TTS Malayalam Speech Corpus (via Kaggle)](https://www.kaggle.com/kavyamanohar/indic-tts-malayalam-speech-corpus), [Openslr Malayalam Speech Corpus](http://openslr.org/63/), [SMC Malayalam Speech Corpus](https://blog.smc.org.in/malayalam-speech-corpus/) and [IIIT-H Indic Speech Databases](http://speech.iiit.ac.in/index.php/research-svl/69.html).
|
35 |
-
When using this model, make sure that your speech input is sampled at 16kHz.
|
36 |
|
37 |
## Usage
|
38 |
|
@@ -135,8 +134,8 @@ resamplers = {
|
|
135 |
48000: torchaudio.transforms.Resample(48_000, 16_000),
|
136 |
}
|
137 |
|
138 |
-
chars_to_ignore_regex = '[
|
139 |
-
unicode_ignore_regex = r'[
|
140 |
|
141 |
# Preprocessing the datasets.
|
142 |
# We need to read the audio files as arrays
|
@@ -180,4 +179,4 @@ print("WER: {:2f}".format(100 * wer.compute(predictions=result["pred_strings"],
|
|
180 |
|
181 |
A combined dataset was created using [Indic TTS Malayalam Speech Corpus (via Kaggle)](https://www.kaggle.com/kavyamanohar/indic-tts-malayalam-speech-corpus), [Openslr Malayalam Speech Corpus](http://openslr.org/63/), [SMC Malayalam Speech Corpus](https://blog.smc.org.in/malayalam-speech-corpus/) and [IIIT-H Indic Speech Databases](http://speech.iiit.ac.in/index.php/research-svl/69.html). The datasets were downloaded and was converted to HF Dataset format using [this notebook](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/blob/main/make_hf_dataset.ipynb)
|
182 |
|
183 |
-
The notebook used for training and evaluation can be found [here](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/blob/main/fine-tune-xlsr-wav2vec2-on-malayalam-asr-with-
|
|
|
31 |
|
32 |
# Wav2Vec2-Large-XLSR-53-ml
|
33 |
|
34 |
+
Fine-tuned [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on ml (Malayalam) using the [Indic TTS Malayalam Speech Corpus (via Kaggle)](https://www.kaggle.com/kavyamanohar/indic-tts-malayalam-speech-corpus), [Openslr Malayalam Speech Corpus](http://openslr.org/63/), [SMC Malayalam Speech Corpus](https://blog.smc.org.in/malayalam-speech-corpus/) and [IIIT-H Indic Speech Databases](http://speech.iiit.ac.in/index.php/research-svl/69.html). The notebooks used to train model is available [here](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/). When using this model, make sure that your speech input is sampled at 16kHz.
|
|
|
35 |
|
36 |
## Usage
|
37 |
|
|
|
134 |
48000: torchaudio.transforms.Resample(48_000, 16_000),
|
135 |
}
|
136 |
|
137 |
+
chars_to_ignore_regex = '[\\,\\?\\.\\!\\-\\;\\:\\"\\“\\%\\‘\\”\\�Utrnle\\_]'
|
138 |
+
unicode_ignore_regex = r'[\\u200e]'
|
139 |
|
140 |
# Preprocessing the datasets.
|
141 |
# We need to read the audio files as arrays
|
|
|
179 |
|
180 |
A combined dataset was created using [Indic TTS Malayalam Speech Corpus (via Kaggle)](https://www.kaggle.com/kavyamanohar/indic-tts-malayalam-speech-corpus), [Openslr Malayalam Speech Corpus](http://openslr.org/63/), [SMC Malayalam Speech Corpus](https://blog.smc.org.in/malayalam-speech-corpus/) and [IIIT-H Indic Speech Databases](http://speech.iiit.ac.in/index.php/research-svl/69.html). The datasets were downloaded and was converted to HF Dataset format using [this notebook](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/blob/main/make_hf_dataset.ipynb)
|
181 |
|
182 |
+
The notebook used for training and evaluation can be found [here](https://github.com/gauthamsuresh09/wav2vec2-large-xlsr-53-malayalam/blob/main/fine-tune-xlsr-wav2vec2-on-malayalam-asr-with-transformers_v2.ipynb)
|