DylanonWic commited on
Commit
362c82f
·
1 Parent(s): d5b8f18

update model card README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -16
README.md CHANGED
@@ -1,6 +1,8 @@
1
  ---
2
  tags:
3
  - generated_from_trainer
 
 
4
  model-index:
5
  - name: wav2vec2-large-asr-th-2
6
  results: []
@@ -13,14 +15,9 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  This model was trained from scratch on the None dataset.
15
  It achieves the following results on the evaluation set:
16
- - eval_loss: 0.1819
17
- - eval_wer: 0.3346
18
- - eval_cer: 0.0910
19
- - eval_runtime: 735.0436
20
- - eval_samples_per_second: 13.822
21
- - eval_steps_per_second: 1.728
22
- - epoch: 0.95
23
- - step: 4000
24
 
25
  ## Model description
26
 
@@ -39,21 +36,31 @@ More information needed
39
  ### Training hyperparameters
40
 
41
  The following hyperparameters were used during training:
42
- - learning_rate: 0.0001
43
- - train_batch_size: 16
44
  - eval_batch_size: 8
45
  - seed: 42
46
  - gradient_accumulation_steps: 3
47
- - total_train_batch_size: 48
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
  - lr_scheduler_warmup_steps: 500
51
- - training_steps: 5000
52
  - mixed_precision_training: Native AMP
53
 
 
 
 
 
 
 
 
 
 
 
54
  ### Framework versions
55
 
56
- - Transformers 4.27.4
57
- - Pytorch 1.13.1+cu116
58
- - Datasets 2.11.0
59
- - Tokenizers 0.13.2
 
1
  ---
2
  tags:
3
  - generated_from_trainer
4
+ metrics:
5
+ - wer
6
  model-index:
7
  - name: wav2vec2-large-asr-th-2
8
  results: []
 
15
 
16
  This model was trained from scratch on the None dataset.
17
  It achieves the following results on the evaluation set:
18
+ - Loss: 0.2501
19
+ - Wer: 0.3205
20
+ - Cer: 0.0879
 
 
 
 
 
21
 
22
  ## Model description
23
 
 
36
  ### Training hyperparameters
37
 
38
  The following hyperparameters were used during training:
39
+ - learning_rate: 2e-05
40
+ - train_batch_size: 12
41
  - eval_batch_size: 8
42
  - seed: 42
43
  - gradient_accumulation_steps: 3
44
+ - total_train_batch_size: 36
45
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
46
  - lr_scheduler_type: linear
47
  - lr_scheduler_warmup_steps: 500
48
+ - training_steps: 4000
49
  - mixed_precision_training: Native AMP
50
 
51
+ ### Training results
52
+
53
+ | Training Loss | Epoch | Step | Validation Loss | Wer | Cer |
54
+ |:-------------:|:-----:|:----:|:---------------:|:------:|:------:|
55
+ | 0.1461 | 0.18 | 1000 | 0.4394 | 0.3238 | 0.0888 |
56
+ | 0.1817 | 0.36 | 2000 | 0.4112 | 0.3261 | 0.0895 |
57
+ | 0.2881 | 0.53 | 3000 | 0.3429 | 0.3241 | 0.0890 |
58
+ | 0.7837 | 0.71 | 4000 | 0.2501 | 0.3205 | 0.0879 |
59
+
60
+
61
  ### Framework versions
62
 
63
+ - Transformers 4.28.1
64
+ - Pytorch 2.0.0+cu118
65
+ - Datasets 2.12.0
66
+ - Tokenizers 0.13.3