ivanlau commited on
Commit
db1ddb6
1 Parent(s): ba45084

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +19 -32
README.md CHANGED
@@ -1,10 +1,6 @@
1
  ---
2
- language:
3
- - zh-HK
4
  license: apache-2.0
5
  tags:
6
- - automatic-speech-recognition
7
- - mozilla-foundation/common_voice_8_0
8
  - generated_from_trainer
9
  datasets:
10
  - common_voice
@@ -18,10 +14,10 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  #
20
 
21
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_8_0 - ZH-HK dataset.
22
  It achieves the following results on the evaluation set:
23
- - Loss: 40.6968
24
- - Wer: 1.0
25
 
26
  ## Model description
27
 
@@ -41,40 +37,31 @@ More information needed
41
 
42
  The following hyperparameters were used during training:
43
  - learning_rate: 0.0003
44
- - train_batch_size: 8
45
- - eval_batch_size: 1
46
  - seed: 42
47
  - gradient_accumulation_steps: 2
48
- - total_train_batch_size: 16
49
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
50
  - lr_scheduler_type: linear
51
  - lr_scheduler_warmup_steps: 500
52
- - num_epochs: 1.0
53
  - mixed_precision_training: Native AMP
54
 
55
  ### Training results
56
 
57
- | Training Loss | Epoch | Step | Validation Loss | Wer |
58
- |:-------------:|:-----:|:----:|:---------------:|:---:|
59
- | No log | 0.05 | 10 | 239.0014 | 1.0 |
60
- | No log | 0.1 | 20 | 235.8207 | 1.0 |
61
- | No log | 0.15 | 30 | 226.9009 | 1.0 |
62
- | No log | 0.21 | 40 | 198.0769 | 1.0 |
63
- | No log | 0.26 | 50 | 166.6728 | 1.0 |
64
- | No log | 0.31 | 60 | 149.1445 | 1.0 |
65
- | No log | 0.36 | 70 | 138.4403 | 1.0 |
66
- | No log | 0.41 | 80 | 131.7249 | 1.0 |
67
- | No log | 0.46 | 90 | 125.5583 | 1.0 |
68
- | No log | 0.51 | 100 | 119.7515 | 1.0 |
69
- | No log | 0.56 | 110 | 113.7283 | 1.0 |
70
- | No log | 0.62 | 120 | 107.2455 | 1.0 |
71
- | No log | 0.67 | 130 | 100.2172 | 1.0 |
72
- | No log | 0.72 | 140 | 92.5585 | 1.0 |
73
- | No log | 0.77 | 150 | 84.2573 | 1.0 |
74
- | No log | 0.82 | 160 | 75.2953 | 1.0 |
75
- | No log | 0.87 | 170 | 65.6953 | 1.0 |
76
- | No log | 0.92 | 180 | 55.7544 | 1.0 |
77
- | No log | 0.97 | 190 | 45.7297 | 1.0 |
78
 
79
 
80
  ### Framework versions
 
1
  ---
 
 
2
  license: apache-2.0
3
  tags:
 
 
4
  - generated_from_trainer
5
  datasets:
6
  - common_voice
 
14
 
15
  #
16
 
17
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice dataset.
18
  It achieves the following results on the evaluation set:
19
+ - Loss: 2.6726
20
+ - Wer: 0.9815
21
 
22
  ## Model description
23
 
 
37
 
38
  The following hyperparameters were used during training:
39
  - learning_rate: 0.0003
40
+ - train_batch_size: 32
41
+ - eval_batch_size: 16
42
  - seed: 42
43
  - gradient_accumulation_steps: 2
44
+ - total_train_batch_size: 64
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 500
48
+ - num_epochs: 10.0
49
  - mixed_precision_training: Native AMP
50
 
51
  ### Training results
52
 
53
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
54
+ |:-------------:|:-----:|:----:|:---------------:|:------:|
55
+ | No log | 1.0 | 183 | 47.8442 | 1.0 |
56
+ | No log | 2.0 | 366 | 6.3109 | 1.0 |
57
+ | 41.8902 | 3.0 | 549 | 6.2392 | 1.0 |
58
+ | 41.8902 | 4.0 | 732 | 5.9739 | 1.1123 |
59
+ | 41.8902 | 5.0 | 915 | 4.9014 | 1.9474 |
60
+ | 5.5817 | 6.0 | 1098 | 3.9892 | 1.0188 |
61
+ | 5.5817 | 7.0 | 1281 | 3.5080 | 1.0104 |
62
+ | 5.5817 | 8.0 | 1464 | 3.0797 | 0.9905 |
63
+ | 3.5579 | 9.0 | 1647 | 2.8111 | 0.9836 |
64
+ | 3.5579 | 10.0 | 1830 | 2.6726 | 0.9815 |
 
 
 
 
 
 
 
 
 
65
 
66
 
67
  ### Framework versions