versae commited on
Commit
bab3af4
1 Parent(s): 7ca8c95

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +129 -115
README.md CHANGED
@@ -1,115 +1,129 @@
1
- ---
2
- license: apache-2.0
3
- tags:
4
- - automatic-speech-recognition
5
- - NbAiLab/NPSC
6
- - no
7
- - nn
8
- - nn-NO
9
- datasets:
10
- - NbAiLab/NPSC
11
- language:
12
- - nn-NO
13
- model-index:
14
- - name: nb-wav2vec2-300m-nynorsk
15
- results:
16
- - task:
17
- name: Automatic Speech Recognition
18
- type: automatic-speech-recognition
19
- dataset:
20
- name: NPSC
21
- type: NbAiLab/NPSC
22
- args: 16K_mp3_nynorsk
23
- metrics:
24
- - name: Test (Nynorsk) WER
25
- type: wer
26
- value: 0.1222
27
- - name: Test (Nynorsk) CER
28
- type: cer
29
- value: 0.0419
30
- ---
31
-
32
- # Norwegian Wav2Vec2 Model - 300M - VoxRex - Nynorsk
33
- This model is finetuned on top of feature extractor [VoxRex-model](https://huggingface.co/KBLab/wav2vec2-large-voxrex) from the National Library of Sweden. The finetuned model achieves the following results on the test set with a 5-gram KenLM. The numbers in parentheses are the results without the language model:
34
- - **WER: 0.1222** (0.1537)
35
- - **CER: 0.0419** (0.0468)
36
-
37
- ## Model description
38
- This is one of several Wav2Vec-models our team created during the 🤗 hosted [Robust Speech Event](https://discuss.huggingface.co/t/open-to-the-community-robust-speech-recognition-challenge/13614?s=09). This is the complete list of our models and their final scores:
39
-
40
- | Model | Final WER | |
41
- |:--------------|:------------|:------------:|
42
- | [NbAiLab/nb-wav2vec2-1b-bokmaal](https://huggingface.co/NbAiLab/nb-wav2vec2-1b-bokmaal) | 6.33 | |
43
- | [NbAiLab/nb-wav2vec2-300m-bokmaal](https://huggingface.co/NbAiLab/nb-wav2vec2-300m-bokmaal) | 7.03 | |
44
- | NbAiLab/nb-wav2vec2-300m-nynorsk (this model) | 12.22 | |
45
- ### Dataset
46
- In parallel with the event, the team also converted the [Norwegian Parliamentary Speech Corpus (NPSC)](https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-58/) to the [NbAiLab/NPSC](https://huggingface.co/datasets/NbAiLab/NPSC) in 🤗 Dataset format and used that as the main source for training.
47
-
48
- ## Code
49
- We have released all the code developed during the event so that the Norwegian NLP community can build upon it when developing even better Norwegian ASR models. The finetuning of these models is not very computationally demanding. After following the instructions here, you should be able to train your own automatic speech recognition system in less than a day with an average GPU.
50
-
51
- ## Team
52
- The following people contributed to building this model: Rolv-Arild Braaten, Per Egil Kummervold, Andre Kåsen, Javier de la Rosa, Per Erik Solberg, and Freddy Wetjen.
53
-
54
- ## Training procedure
55
- To reproduce these results, we strongly recommend that you follow the [instructions from 🤗](https://github.com/huggingface/transformers/tree/master/examples/research_projects/robust-speech-event#talks) to train a simple Swedish model.
56
-
57
- When you have verified that you are able to do this, create a fresh new repo. You can then start by copying the files ```run.sh``` and ```run_speech_recognition_ctc.py``` from our repo. Running these will create all the other necessary files, and should let you reproduce our results. With some tweaks to the hyperparameters, you might even be able to build an even better ASR. Good luck!
58
-
59
- ### Language Model
60
- As the scores indicate, adding even a simple 5-gram language will improve the results. 🤗 has provided another [very nice blog](https://huggingface.co/blog/wav2vec2-with-ngram) explaining how to add a 5-gram language model to improve the ASR model. You can build this from your own corpus, for instance by extracting some suitable text from the [Norwegian Colossal Corpus](https://huggingface.co/datasets/NbAiLab/NCC). You can also skip some of the steps in the guide, and copy the [5-gram model from this repo](https://huggingface.co/NbAiLab/XLSR-300M-bokmaal/tree/main/language_model).
61
-
62
-
63
- ### Parameters
64
- The final model was run using these parameters:
65
- ```
66
- --dataset_name="NbAiLab/NPSC"
67
- --model_name_or_path="KBLab/wav2vec2-large-voxrex"
68
- --dataset_config_name="16K_mp3_nynorsk"
69
- --output_dir="./"
70
- --overwrite_output_dir
71
- --num_train_epochs="80"
72
- --per_device_train_batch_size="16"
73
- --per_device_eval_batch_size="16"
74
- --gradient_accumulation_steps="2"
75
- --learning_rate="1e-4"
76
- --warmup_steps="2000"
77
- --length_column_name="input_length"
78
- --evaluation_strategy="steps"
79
- --text_column_name="text"
80
- --save_steps="500"
81
- --eval_steps="500"
82
- --logging_steps="100"
83
- --layerdrop="0.041"
84
- --attention_dropout="0.094"
85
- --activation_dropout="0.055"
86
- --hidden_dropout="0.047"
87
- --save_total_limit="3"
88
- --freeze_feature_encoder
89
- --feat_proj_dropout="0.04"
90
- --mask_time_prob="0.082"
91
- --mask_time_length="10"
92
- --mask_feature_prob="0.25"
93
- --mask_feature_length="64"
94
- --gradient_checkpointing
95
- --min_duration_in_seconds="0.5"
96
- --max_duration_in_seconds="30.0"
97
- --use_auth_token
98
- --seed="42"
99
- --fp16
100
- --group_by_length
101
- --do_train --do_eval
102
- --push_to_hub
103
- --preprocessing_num_workers="32"
104
- ```
105
-
106
- Using these settings, the training might take 3-4 days on an average GPU. You can, however, get a decent model and faster results by tweaking these parameters.
107
-
108
- | Parameter| Comment |
109
- |:-------------|:-----|
110
- | per_device_train_batch_size | Adjust this to the maximum of available memory. 16 or 24 might be good settings depending on your system |
111
- |gradient_accumulation_steps |Can be adjusted even further up to increase batch size and speed up training without running into memory issues |
112
- | learning_rate|Can be increased, maybe as high as 1e-4. Speeds up training but might add instability |
113
- | epochs| Can be decreased significantly. This is a huge dataset and you might get a decent result already after a couple of epochs|
114
-
115
-
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ tags:
4
+ - automatic-speech-recognition
5
+ datasets:
6
+ - NbAiLab/NPSC
7
+ language:
8
+ - nn
9
+ model-index:
10
+ - name: nb-wav2vec2-300m-nynorsk
11
+ results:
12
+ - task:
13
+ name: Automatic Speech Recognition
14
+ type: automatic-speech-recognition
15
+ dataset:
16
+ name: NPSC
17
+ type: NbAiLab/NPSC
18
+ args: 16K_mp3_nynorsk
19
+ metrics:
20
+ - name: Test (Nynorsk) WER
21
+ type: wer
22
+ value: 0.1222
23
+ - name: Test (Nynorsk) CER
24
+ type: cer
25
+ value: 0.0419
26
+ ---
27
+
28
+ # Norwegian Wav2Vec2 Model - 300M - VoxRex - Nynorsk
29
+ This model is finetuned on top of feature extractor [VoxRex-model](https://huggingface.co/KBLab/wav2vec2-large-voxrex) from the National Library of Sweden. The finetuned model achieves the following results on the test set with a 5-gram KenLM. The numbers in parentheses are the results without the language model:
30
+ - **WER: 0.1222** (0.1537)
31
+ - **CER: 0.0419** (0.0468)
32
+
33
+ ## Model description
34
+ This is one of several Wav2Vec-models our team created during the 🤗 hosted [Robust Speech Event](https://discuss.huggingface.co/t/open-to-the-community-robust-speech-recognition-challenge/13614?s=09). This is the complete list of our models and their final scores:
35
+
36
+ | Model | Final WER | |
37
+ |:--------------|:------------|:------------:|
38
+ | [NbAiLab/nb-wav2vec2-1b-bokmaal](https://huggingface.co/NbAiLab/nb-wav2vec2-1b-bokmaal) | 6.33 | |
39
+ | [NbAiLab/nb-wav2vec2-300m-bokmaal](https://huggingface.co/NbAiLab/nb-wav2vec2-300m-bokmaal) | 7.03 | |
40
+ | NbAiLab/nb-wav2vec2-300m-nynorsk (this model) | 12.22 | |
41
+ ### Dataset
42
+ In parallel with the event, the team also converted the [Norwegian Parliamentary Speech Corpus (NPSC)](https://www.nb.no/sprakbanken/en/resource-catalogue/oai-nb-no-sbr-58/) to the [NbAiLab/NPSC](https://huggingface.co/datasets/NbAiLab/NPSC) in 🤗 Dataset format and used that as the main source for training.
43
+
44
+ ## Code
45
+ We have released all the code developed during the event so that the Norwegian NLP community can build upon it when developing even better Norwegian ASR models. The finetuning of these models is not very computationally demanding. After following the instructions here, you should be able to train your own automatic speech recognition system in less than a day with an average GPU.
46
+
47
+ ## Team
48
+ The following people contributed to building this model: Rolv-Arild Braaten, Per Egil Kummervold, Andre Kåsen, Javier de la Rosa, Per Erik Solberg, and Freddy Wetjen.
49
+
50
+ ## Training procedure
51
+ To reproduce these results, we strongly recommend that you follow the [instructions from 🤗](https://github.com/huggingface/transformers/tree/master/examples/research_projects/robust-speech-event#talks) to train a simple Swedish model.
52
+
53
+ When you have verified that you are able to do this, create a fresh new repo. You can then start by copying the files ```run.sh``` and ```run_speech_recognition_ctc.py``` from our repo. Running these will create all the other necessary files, and should let you reproduce our results. With some tweaks to the hyperparameters, you might even be able to build an even better ASR. Good luck!
54
+
55
+ ### Language Model
56
+ As the scores indicate, adding even a simple 5-gram language will improve the results. 🤗 has provided another [very nice blog](https://huggingface.co/blog/wav2vec2-with-ngram) explaining how to add a 5-gram language model to improve the ASR model. You can build this from your own corpus, for instance by extracting some suitable text from the [Norwegian Colossal Corpus](https://huggingface.co/datasets/NbAiLab/NCC). You can also skip some of the steps in the guide, and copy the [5-gram model from this repo](https://huggingface.co/NbAiLab/XLSR-300M-bokmaal/tree/main/language_model).
57
+
58
+
59
+ ### Parameters
60
+ The final model was run using these parameters:
61
+ ```
62
+ --dataset_name="NbAiLab/NPSC"
63
+ --model_name_or_path="KBLab/wav2vec2-large-voxrex"
64
+ --dataset_config_name="16K_mp3_nynorsk"
65
+ --output_dir="./"
66
+ --overwrite_output_dir
67
+ --num_train_epochs="80"
68
+ --per_device_train_batch_size="16"
69
+ --per_device_eval_batch_size="16"
70
+ --gradient_accumulation_steps="2"
71
+ --learning_rate="1e-4"
72
+ --warmup_steps="2000"
73
+ --length_column_name="input_length"
74
+ --evaluation_strategy="steps"
75
+ --text_column_name="text"
76
+ --save_steps="500"
77
+ --eval_steps="500"
78
+ --logging_steps="100"
79
+ --layerdrop="0.041"
80
+ --attention_dropout="0.094"
81
+ --activation_dropout="0.055"
82
+ --hidden_dropout="0.047"
83
+ --save_total_limit="3"
84
+ --freeze_feature_encoder
85
+ --feat_proj_dropout="0.04"
86
+ --mask_time_prob="0.082"
87
+ --mask_time_length="10"
88
+ --mask_feature_prob="0.25"
89
+ --mask_feature_length="64"
90
+ --gradient_checkpointing
91
+ --min_duration_in_seconds="0.5"
92
+ --max_duration_in_seconds="30.0"
93
+ --use_auth_token
94
+ --seed="42"
95
+ --fp16
96
+ --group_by_length
97
+ --do_train --do_eval
98
+ --push_to_hub
99
+ --preprocessing_num_workers="32"
100
+ ```
101
+
102
+ Using these settings, the training might take 3-4 days on an average GPU. You can, however, get a decent model and faster results by tweaking these parameters.
103
+
104
+ | Parameter| Comment |
105
+ |:-------------|:-----|
106
+ | per_device_train_batch_size | Adjust this to the maximum of available memory. 16 or 24 might be good settings depending on your system |
107
+ |gradient_accumulation_steps |Can be adjusted even further up to increase batch size and speed up training without running into memory issues |
108
+ | learning_rate|Can be increased, maybe as high as 1e-4. Speeds up training but might add instability |
109
+ | epochs| Can be decreased significantly. This is a huge dataset and you might get a decent result already after a couple of epochs|
110
+
111
+ ## Citation
112
+
113
+ ```bibtex
114
+ @inproceedings{de-la-rosa-etal-2023-boosting,
115
+ title = "Boosting {N}orwegian Automatic Speech Recognition",
116
+ author = "De La Rosa, Javier and
117
+ Braaten, Rolv-Arild and
118
+ Kummervold, Per and
119
+ Wetjen, Freddy",
120
+ booktitle = "Proceedings of the 24th Nordic Conference on Computational Linguistics (NoDaLiDa)",
121
+ month = may,
122
+ year = "2023",
123
+ address = "T{\'o}rshavn, Faroe Islands",
124
+ publisher = "University of Tartu Library",
125
+ url = "https://aclanthology.org/2023.nodalida-1.55",
126
+ pages = "555--564",
127
+ abstract = "In this paper, we present several baselines for automatic speech recognition (ASR) models for the two official written languages in Norway: Bokm{\aa}l and Nynorsk. We compare the performance of models of varying sizes and pre-training approaches on multiple Norwegian speech datasets. Additionally, we measure the performance of these models against previous state-of-the-art ASR models, as well as on out-of-domain datasets. We improve the state of the art on the Norwegian Parliamentary Speech Corpus (NPSC) from a word error rate (WER) of 17.10{\%} to 7.60{\%}, with models achieving 5.81{\%} for Bokm{\aa}l and 11.54{\%} for Nynorsk. We also discuss the challenges and potential solutions for further improving ASR models for Norwegian.",
128
+ }
129
+ ```