lucio commited on
Commit
98e5ae5
1 Parent(s): 2a77a5b

Training in progress, step 200

Browse files
.ipynb_checkpoints/requirements-checkpoint.txt ADDED
@@ -0,0 +1,8 @@
 
 
 
 
 
 
 
 
 
1
+ unidecode
2
+ tensorboard
3
+ torch
4
+ torchaudio
5
+ jiwer~=2.3.0
6
+ soundfile~=0.10.3
7
+ transformers~=4.16.2
8
+ datasets~=1.18.3
.ipynb_checkpoints/run-checkpoint.sh CHANGED
@@ -1,29 +1,31 @@
1
- python xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
2
  --dataset_name="mozilla-foundation/common_voice_8_0" \
3
  --model_name_or_path="facebook/wav2vec2-xls-r-300m" \
4
  --dataset_config_name="uz" \
5
- --output_dir="./xls-r-uzbek-cv8" \
 
 
6
  --overwrite_output_dir \
7
- --num_train_epochs="30" \
8
- --per_device_train_batch_size="16" \
9
  --per_device_eval_batch_size="8" \
10
  --gradient_accumulation_steps="4" \
11
  --learning_rate="3e-5" \
12
- --warmup_steps="2000" \
13
  --length_column_name="input_length" \
14
  --evaluation_strategy="steps" \
15
  --text_column_name="sentence" \
16
- --eval_metrics="wer,cer" \
17
- --save_steps="500" \
18
- --eval_steps="500" \
19
  --logging_steps="100" \
20
- --min_duration_in_seconds="0.5" \
21
- --layerdrop="0.0" \
22
  --activation_dropout="0.1" \
23
  --save_total_limit="3" \
24
  --freeze_feature_encoder \
25
- --feat_proj_dropout="0.0" \
26
- --mask_time_prob="0.75" \
27
  --mask_time_length="10" \
28
  --mask_feature_prob="0.25" \
29
  --mask_feature_length="64" \
 
1
+ python ~/xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
2
  --dataset_name="mozilla-foundation/common_voice_8_0" \
3
  --model_name_or_path="facebook/wav2vec2-xls-r-300m" \
4
  --dataset_config_name="uz" \
5
+ --output_dir="~/xls-r-uzbek-cv8" \
6
+ --train_split_name="train[:30%]" \
7
+ --eval_split_name="validation[50%:]" \
8
  --overwrite_output_dir \
9
+ --num_train_epochs="100" \
10
+ --per_device_train_batch_size="32" \
11
  --per_device_eval_batch_size="8" \
12
  --gradient_accumulation_steps="4" \
13
  --learning_rate="3e-5" \
14
+ --warmup_steps="500" \
15
  --length_column_name="input_length" \
16
  --evaluation_strategy="steps" \
17
  --text_column_name="sentence" \
18
+ --eval_metrics="cer" \
19
+ --save_steps="200" \
20
+ --eval_steps="200" \
21
  --logging_steps="100" \
22
+ --min_duration_in_seconds="0.2" \
23
+ --layerdrop="0.01" \
24
  --activation_dropout="0.1" \
25
  --save_total_limit="3" \
26
  --freeze_feature_encoder \
27
+ --feat_proj_dropout="0.01" \
28
+ --mask_time_prob="0.50" \
29
  --mask_time_length="10" \
30
  --mask_feature_prob="0.25" \
31
  --mask_feature_length="64" \
config.json CHANGED
@@ -49,7 +49,7 @@
49
  "feat_extract_activation": "gelu",
50
  "feat_extract_dropout": 0.0,
51
  "feat_extract_norm": "layer",
52
- "feat_proj_dropout": 0.0,
53
  "feat_quantizer_dropout": 0.0,
54
  "final_dropout": 0.0,
55
  "hidden_act": "gelu",
@@ -58,13 +58,13 @@
58
  "initializer_range": 0.02,
59
  "intermediate_size": 4096,
60
  "layer_norm_eps": 1e-05,
61
- "layerdrop": 0.0,
62
  "mask_feature_length": 64,
63
  "mask_feature_min_masks": 0,
64
  "mask_feature_prob": 0.25,
65
  "mask_time_length": 10,
66
  "mask_time_min_masks": 2,
67
- "mask_time_prob": 0.75,
68
  "model_type": "wav2vec2",
69
  "num_adapter_layers": 3,
70
  "num_attention_heads": 16,
@@ -100,7 +100,7 @@
100
  1
101
  ],
102
  "torch_dtype": "float32",
103
- "transformers_version": "4.17.0.dev0",
104
  "use_weighted_layer_sum": false,
105
  "vocab_size": 33,
106
  "xvector_output_dim": 512,
 
49
  "feat_extract_activation": "gelu",
50
  "feat_extract_dropout": 0.0,
51
  "feat_extract_norm": "layer",
52
+ "feat_proj_dropout": 0.01,
53
  "feat_quantizer_dropout": 0.0,
54
  "final_dropout": 0.0,
55
  "hidden_act": "gelu",
 
58
  "initializer_range": 0.02,
59
  "intermediate_size": 4096,
60
  "layer_norm_eps": 1e-05,
61
+ "layerdrop": 0.01,
62
  "mask_feature_length": 64,
63
  "mask_feature_min_masks": 0,
64
  "mask_feature_prob": 0.25,
65
  "mask_time_length": 10,
66
  "mask_time_min_masks": 2,
67
+ "mask_time_prob": 0.5,
68
  "model_type": "wav2vec2",
69
  "num_adapter_layers": 3,
70
  "num_attention_heads": 16,
 
100
  1
101
  ],
102
  "torch_dtype": "float32",
103
+ "transformers_version": "4.16.2",
104
  "use_weighted_layer_sum": false,
105
  "vocab_size": 33,
106
  "xvector_output_dim": 512,
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:d7a4269673722dcfe01934e733ea692df4ab980becd9fbb7bf129b60ca1505af
3
  size 1262058993
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3f107f42a91fd5d39621347b43afb0c87ef3258c4bc185ec165f2870c082b35c
3
  size 1262058993
requirements.txt CHANGED
@@ -1,3 +1,8 @@
1
  unidecode
2
  tensorboard
3
-
 
 
 
 
 
 
1
  unidecode
2
  tensorboard
3
+ torch
4
+ torchaudio
5
+ jiwer~=2.3.0
6
+ soundfile~=0.10.3
7
+ transformers~=4.16.2
8
+ datasets~=1.18.3
run.sh CHANGED
@@ -1,29 +1,31 @@
1
- python xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
2
  --dataset_name="mozilla-foundation/common_voice_8_0" \
3
  --model_name_or_path="facebook/wav2vec2-xls-r-300m" \
4
  --dataset_config_name="uz" \
5
- --output_dir="./xls-r-uzbek-cv8" \
 
 
6
  --overwrite_output_dir \
7
- --num_train_epochs="30" \
8
- --per_device_train_batch_size="16" \
9
  --per_device_eval_batch_size="8" \
10
  --gradient_accumulation_steps="4" \
11
  --learning_rate="3e-5" \
12
- --warmup_steps="2000" \
13
  --length_column_name="input_length" \
14
  --evaluation_strategy="steps" \
15
  --text_column_name="sentence" \
16
- --eval_metrics="wer,cer" \
17
- --save_steps="500" \
18
- --eval_steps="500" \
19
  --logging_steps="100" \
20
- --min_duration_in_seconds="0.5" \
21
- --layerdrop="0.0" \
22
  --activation_dropout="0.1" \
23
  --save_total_limit="3" \
24
  --freeze_feature_encoder \
25
- --feat_proj_dropout="0.0" \
26
- --mask_time_prob="0.75" \
27
  --mask_time_length="10" \
28
  --mask_feature_prob="0.25" \
29
  --mask_feature_length="64" \
 
1
+ python ~/xls-r-uzbek-cv8/run_speech_recognition_ctc.py \
2
  --dataset_name="mozilla-foundation/common_voice_8_0" \
3
  --model_name_or_path="facebook/wav2vec2-xls-r-300m" \
4
  --dataset_config_name="uz" \
5
+ --output_dir="~/xls-r-uzbek-cv8" \
6
+ --train_split_name="train[:30%]" \
7
+ --eval_split_name="validation[50%:]" \
8
  --overwrite_output_dir \
9
+ --num_train_epochs="100" \
10
+ --per_device_train_batch_size="32" \
11
  --per_device_eval_batch_size="8" \
12
  --gradient_accumulation_steps="4" \
13
  --learning_rate="3e-5" \
14
+ --warmup_steps="500" \
15
  --length_column_name="input_length" \
16
  --evaluation_strategy="steps" \
17
  --text_column_name="sentence" \
18
+ --eval_metrics="cer" \
19
+ --save_steps="200" \
20
+ --eval_steps="200" \
21
  --logging_steps="100" \
22
+ --min_duration_in_seconds="0.2" \
23
+ --layerdrop="0.01" \
24
  --activation_dropout="0.1" \
25
  --save_total_limit="3" \
26
  --freeze_feature_encoder \
27
+ --feat_proj_dropout="0.01" \
28
+ --mask_time_prob="0.50" \
29
  --mask_time_length="10" \
30
  --mask_feature_prob="0.25" \
31
  --mask_feature_length="64" \
runs/Feb05_08-15-45_job-0a778896-a7e2-46e9-bcf5-016f91f242cf/1644049006.003602/events.out.tfevents.1644049006.job-0a778896-a7e2-46e9-bcf5-016f91f242cf.23457.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:84f165c6e40f5a1e87d21f624a9b4475d58842f270044fb50735b8f5a9d00101
3
+ size 4826
runs/Feb05_08-15-45_job-0a778896-a7e2-46e9-bcf5-016f91f242cf/events.out.tfevents.1644049005.job-0a778896-a7e2-46e9-bcf5-016f91f242cf.23457.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:314b3baf73fa8b9becdd35cfeecc2a0a871edb0fda354e78bfa86e375dcc81fd
3
+ size 5404
special_tokens_map.json CHANGED
@@ -1 +1 @@
1
- {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
 
1
+ {"bos_token": "<s>", "eos_token": "</s>", "unk_token": "[UNK]", "pad_token": "[PAD]", "additional_special_tokens": [{"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "<s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}, {"content": "</s>", "single_word": false, "lstrip": false, "rstrip": false, "normalized": true}]}
tokenizer_config.json CHANGED
@@ -1 +1 @@
1
- {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "./xls-r-uzbek-cv8", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
 
1
+ {"unk_token": "[UNK]", "bos_token": "<s>", "eos_token": "</s>", "pad_token": "[PAD]", "do_lower_case": false, "word_delimiter_token": "|", "special_tokens_map_file": null, "tokenizer_file": null, "name_or_path": "/workspace/xls-r-uzbek-cv8", "tokenizer_class": "Wav2Vec2CTCTokenizer"}
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:ebc2a615608b93ae5d470729cc3ef83ec5ad23e35190fdca59d878f4f9e7636a
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7a661f5e8e620b98c6c94d0bb8c2ddd7376f2412096a9ce2ff8088711cea618b
3
  size 3055