Titouan commited on
Commit
24632ba
1 Parent(s): 115e3dd
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -30,14 +30,14 @@ The performance of the model is the following:
30
 
31
  | Release | Test WER | GPUs |
32
  |:--------------:|:--------------:| :--------:|
33
- | 03-06-21 | 15.69 | 2xV100 32GB |
34
 
35
  ## Pipeline description
36
 
37
  This ASR system is composed of 2 different but linked blocks:
38
  - Tokenizer (unigram) that transforms words into subword units and trained with
39
  the train transcriptions (train.tsv) of CommonVoice (RW).
40
- - Acoustic model (wav2vec2.0 + CTC/Attention). A pretrained wav2vec 2.0 model ([wav2vec2-lv60-large](https://huggingface.co/facebook/wav2vec2-large-lv60)) is combined with two DNN layers and finetuned on CommonVoice En.
41
  The obtained final acoustic representation is given to the CTC and attention decoders.
42
 
43
 
@@ -81,7 +81,7 @@ pip install -e .
81
  3. Run Training:
82
  ```bash
83
  cd recipes/CommonVoice/ASR/seq2seq
84
- python train.py hparams/train_fr_with_wav2vec.yaml --data_folder=your_data_folder
85
  ```
86
 
87
  You can find our training results (models, logs, etc) [here](https://drive.google.com/drive/folders/1tjz6IZmVRkuRE97E7h1cXFoGTer7pT73?usp=sharing).
30
 
31
  | Release | Test WER | GPUs |
32
  |:--------------:|:--------------:| :--------:|
33
+ | 03-06-21 | 18.91 | 2xV100 32GB |
34
 
35
  ## Pipeline description
36
 
37
  This ASR system is composed of 2 different but linked blocks:
38
  - Tokenizer (unigram) that transforms words into subword units and trained with
39
  the train transcriptions (train.tsv) of CommonVoice (RW).
40
+ - Acoustic model (wav2vec2.0 + CTC/Attention). A pretrained wav2vec 2.0 model ([wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53)) is combined with two DNN layers and finetuned on CommonVoice En.
41
  The obtained final acoustic representation is given to the CTC and attention decoders.
42
 
43
 
81
  3. Run Training:
82
  ```bash
83
  cd recipes/CommonVoice/ASR/seq2seq
84
+ python train.py hparams/train_rw_with_wav2vec.yaml --data_folder=your_data_folder
85
  ```
86
 
87
  You can find our training results (models, logs, etc) [here](https://drive.google.com/drive/folders/1tjz6IZmVRkuRE97E7h1cXFoGTer7pT73?usp=sharing).