kingabzpro commited on
Commit
63b779a
1 Parent(s): eccc472

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +23 -23
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  language:
3
- - sv-SE
4
  license: apache-2.0
5
  tags:
6
  - automatic-speech-recognition
@@ -26,32 +26,32 @@ model-index:
26
  value: 24.73
27
  name: Test WER
28
  args:
29
- - learning_rate: 7.5e-05
30
- - train_batch_size: 64
31
- - eval_batch_size: 8
32
- - seed: 42
33
- - gradient_accumulation_steps: 2
34
- - total_train_batch_size: 128
35
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
36
- - lr_scheduler_type: linear
37
- - lr_scheduler_warmup_steps: 1000
38
- - num_epochs: 50
39
- - mixed_precision_training: Native AMP
40
  - type: cer
41
  value: 7.58
42
  name: Test CER
43
  args:
44
- - learning_rate: 7.5e-05
45
- - train_batch_size: 64
46
- - eval_batch_size: 8
47
- - seed: 42
48
- - gradient_accumulation_steps: 2
49
- - total_train_batch_size: 128
50
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
- - lr_scheduler_type: linear
52
- - lr_scheduler_warmup_steps: 1000
53
- - num_epochs: 50
54
- - mixed_precision_training: Native AMP
55
  ---
56
 
57
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
1
  ---
2
  language:
3
+ - sv
4
  license: apache-2.0
5
  tags:
6
  - automatic-speech-recognition
 
26
  value: 24.73
27
  name: Test WER
28
  args:
29
+ learning_rate: 7.5e-05
30
+ train_batch_size: 64
31
+ eval_batch_size: 8
32
+ seed: 42
33
+ gradient_accumulation_steps: 2
34
+ total_train_batch_size: 128
35
+ optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
36
+ lr_scheduler_type: linear
37
+ lr_scheduler_warmup_steps: 1000
38
+ num_epochs: 50
39
+ mixed_precision_training: Native AMP
40
  - type: cer
41
  value: 7.58
42
  name: Test CER
43
  args:
44
+ learning_rate: 7.5e-05
45
+ train_batch_size: 64
46
+ eval_batch_size: 8
47
+ seed: 42
48
+ gradient_accumulation_steps: 2
49
+ total_train_batch_size: 128
50
+ optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
51
+ lr_scheduler_type: linear
52
+ lr_scheduler_warmup_steps: 1000
53
+ num_epochs: 50
54
+ mixed_precision_training: Native AMP
55
  ---
56
 
57
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You