pere commited on
Commit
2784ce7
1 Parent(s): 08171bc

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +66 -2
README.md CHANGED
@@ -1,7 +1,71 @@
1
  ---
2
- license: cc
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
4
- This is based on checkpoint 8500 of NbAiLab/wav2vec2-xlsr-300M-NPSC-OH
5
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
6
 
7
 
 
1
  ---
2
+ language:
3
+ - sv-SE
4
+ license: apache-2.0
5
+ tags:
6
+ - automatic-speech-recognition
7
+ - mozilla-foundation/common_voice_7_0
8
+ - generated_from_trainer
9
+ - no
10
+ - robust-speech-event
11
+ - model_for_talk
12
+ datasets:
13
+ - mozilla-foundation/common_voice_7_0
14
+ model-index:
15
+ - name: XLS-R-300M-LM - Norwegian
16
+ results:
17
+ - task:
18
+ name: Automatic Speech Recognition
19
+ type: automatic-speech-recognition
20
+ dataset:
21
+ name: NPSC
22
+ type: NbAiLab/NPSC
23
+ args: sv-SE
24
+ metrics:
25
+ - name: Eval WER
26
+ type: wer
27
+ value: 21.10
28
+ - name: Eval CER
29
+ type: cer
30
+ value: 0.06
31
+
32
  ---
 
33
 
34
+ # XLS-R-300M-LM - Norwegian
35
+
36
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_7_0 - SV-SE dataset.
37
+ ### Scores without Language Model
38
+ Without using a language model, it achieves the following scores on the NPSC Eval set
39
+ It achieves the following results on the evaluation set without a language model:
40
+ - Loss: 0.1992
41
+ - WER: 0.2110
42
+ - CER: 0.0622
43
+
44
+ ### Scores with Language Model
45
+ A 5-gram KenLM was added to boost the models performance. After
46
+
47
+ ## Model description
48
+ This current version is based on checkpoint 8500 of [NbAiLab/wav2vec2-xlsr-300M-NPSC-OH](https://huggingface.co/NbAiLab/wav2vec2-xlsr-300M-NPSC-OH)
49
+
50
+ ## Intended uses & limitations
51
+ Demo version only. The model will be updated later this week.
52
+
53
+ ## Training and evaluation data
54
+ The model is trained and evaluated on [NPSC](https://huggingface.co/datasets/NbAiLab/NPSC). Unfortunately there is no Norwegian test data in Common Voice, and currently the model is only evaluated on the validation set of NPSC..
55
+
56
+ ## Training procedure
57
+ ### Training hyperparameters
58
+ The following hyperparameters were used during training:
59
+ - learning_rate: 7.5e-05
60
+ - train_batch_size: 8
61
+ - eval_batch_size: 8
62
+ - seed: 42
63
+ - gradient_accumulation_steps: 4
64
+ - total_train_batch_size: 32
65
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
66
+ - lr_scheduler_type: linear
67
+ - lr_scheduler_warmup_steps: 2000
68
+ - num_epochs: 30.0 (But interrupted after 8500 steps, approx 6 epochs)
69
+ - mixed_precision_training: Native AMP
70
 
71