abdouaziiz commited on
Commit
3f7fb84
1 Parent(s): 6325747

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -9
README.md CHANGED
@@ -30,24 +30,29 @@ should probably proofread and complete it, then remove this comment. -->
30
  Wolof is a language spoken in Senegal and neighbouring countries, this language is not too well represented, there are few resources in the field of Text en speech
31
  In this sense we aim to bring our contribution to this, it is in this sense that enters this repo.
32
 
33
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) , that is trained with the largest available speech dataset of the [ALLFA project](https://github.com/besacier/ALFFA_PUBLIC/tree/master/ASR/WOLOF)
34
 
35
  It achieves the following results on the evaluation set:
36
  - Loss: 0.367826
37
  - Wer: 0.212565
38
 
39
  ## Model description
40
- More information needed
41
- ## Intended uses & limitations
42
- More information needed
43
  ## Training and evaluation data
44
- More information needed
45
- ## Training procedure
46
  ### Training hyperparameters
47
- More information needed
48
- ### Training results
 
 
 
 
 
 
 
 
49
 
50
-
51
 
52
  | Step | Training Loss | Validation Loss | Wer |
53
  |:-------:|:-------------:|:---------------:|:------:|
 
30
  Wolof is a language spoken in Senegal and neighbouring countries, this language is not too well represented, there are few resources in the field of Text en speech
31
  In this sense we aim to bring our contribution to this, it is in this sense that enters this repo.
32
 
33
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) , that is trained with the largest available speech dataset of the [ALFFA_PUBLIC project](https://github.com/besacier/ALFFA_PUBLIC/tree/master/ASR/WOLOF)
34
 
35
  It achieves the following results on the evaluation set:
36
  - Loss: 0.367826
37
  - Wer: 0.212565
38
 
39
  ## Model description
40
+ The duration of the training data is 16.8 hours, which we have divided into 10,000 audio files for the training and 3,339 for the test.
 
 
41
  ## Training and evaluation data
42
+ We eval the model at every 1500 step , and log it . and save at every 33340 step
 
43
  ### Training hyperparameters
44
+ The following hyperparameters were used during training:
45
+ - learning_rate: 1e-4
46
+ - train_batch_size: 3
47
+ - eval_batch_size : 8
48
+ - total_train_batch_size: 64
49
+ - total_eval_batch_size: 64
50
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ - lr_scheduler_type: linear
52
+ - lr_scheduler_warmup_steps: 1000
53
+ - num_epochs: 10.0
54
 
55
+ ### Training results
56
 
57
  | Step | Training Loss | Validation Loss | Wer |
58
  |:-------:|:-------------:|:---------------:|:------:|