VladS159 commited on
Commit
28dbd60
1 Parent(s): 50c6715

End of training

Browse files
README.md CHANGED
@@ -1,40 +1,24 @@
1
  ---
2
- language:
3
- - ro
4
  license: apache-2.0
5
  base_model: openai/whisper-medium
6
  tags:
7
- - hf-asr-leaderboard
8
  - generated_from_trainer
9
- datasets:
10
- - VladS159/common_voice_romanian_speech_synthesis
11
  metrics:
12
  - wer
13
  model-index:
14
- - name: Whisper Medium Ro - Sarbu Vlad - multi gpu
15
- results:
16
- - task:
17
- name: Automatic Speech Recognition
18
- type: automatic-speech-recognition
19
- dataset:
20
- name: Common Voice 16.1 + Romanian speech synthesis
21
- type: VladS159/common_voice_romanian_speech_synthesis
22
- args: 'config: ro, split: test'
23
- metrics:
24
- - name: Wer
25
- type: wer
26
- value: 12.181988686208669
27
  ---
28
 
29
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
30
  should probably proofread and complete it, then remove this comment. -->
31
 
32
- # Whisper Medium Ro - Sarbu Vlad - multi gpu
33
 
34
- This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the Common Voice 16.1 + Romanian speech synthesis dataset.
35
  It achieves the following results on the evaluation set:
36
- - Loss: 0.1620
37
- - Wer: 12.1820
38
 
39
  ## Model description
40
 
@@ -54,31 +38,27 @@ More information needed
54
 
55
  The following hyperparameters were used during training:
56
  - learning_rate: 1e-05
57
- - train_batch_size: 16
58
- - eval_batch_size: 16
59
  - seed: 42
60
  - distributed_type: multi-GPU
61
  - num_devices: 3
62
- - total_train_batch_size: 48
63
- - total_eval_batch_size: 48
64
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
65
  - lr_scheduler_type: linear
66
- - lr_scheduler_warmup_steps: 200
67
- - training_steps: 2000
68
  - mixed_precision_training: Native AMP
69
 
70
  ### Training results
71
 
72
  | Training Loss | Epoch | Step | Validation Loss | Wer |
73
  |:-------------:|:-----:|:----:|:---------------:|:-------:|
74
- | 0.148 | 0.98 | 250 | 0.1494 | 14.1574 |
75
- | 0.0875 | 1.96 | 500 | 0.1295 | 12.9080 |
76
- | 0.0404 | 2.94 | 750 | 0.1285 | 11.8734 |
77
- | 0.0227 | 3.92 | 1000 | 0.1353 | 12.1094 |
78
- | 0.0139 | 4.9 | 1250 | 0.1409 | 11.9702 |
79
- | 0.0076 | 5.88 | 1500 | 0.1539 | 12.0459 |
80
- | 0.005 | 6.86 | 1750 | 0.1599 | 12.1880 |
81
- | 0.0039 | 7.84 | 2000 | 0.1620 | 12.1820 |
82
 
83
 
84
  ### Framework versions
 
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: openai/whisper-medium
4
  tags:
 
5
  - generated_from_trainer
 
 
6
  metrics:
7
  - wer
8
  model-index:
9
+ - name: Whisper_medium_ro_VladS_1000_steps_multi_gpu_25_02_2024
10
+ results: []
 
 
 
 
 
 
 
 
 
 
 
11
  ---
12
 
13
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
14
  should probably proofread and complete it, then remove this comment. -->
15
 
16
+ # Whisper_medium_ro_VladS_1000_steps_multi_gpu_25_02_2024
17
 
18
+ This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
+ - Loss: 0.1247
21
+ - Wer: 11.7262
22
 
23
  ## Model description
24
 
 
38
 
39
  The following hyperparameters were used during training:
40
  - learning_rate: 1e-05
41
+ - train_batch_size: 10
42
+ - eval_batch_size: 10
43
  - seed: 42
44
  - distributed_type: multi-GPU
45
  - num_devices: 3
46
+ - total_train_batch_size: 30
47
+ - total_eval_batch_size: 30
48
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
49
  - lr_scheduler_type: linear
50
+ - lr_scheduler_warmup_steps: 100
51
+ - training_steps: 1000
52
  - mixed_precision_training: Native AMP
53
 
54
  ### Training results
55
 
56
  | Training Loss | Epoch | Step | Validation Loss | Wer |
57
  |:-------------:|:-----:|:----:|:---------------:|:-------:|
58
+ | 0.1447 | 0.61 | 250 | 0.1532 | 13.8768 |
59
+ | 0.0599 | 1.23 | 500 | 0.1305 | 12.5141 |
60
+ | 0.0595 | 1.84 | 750 | 0.1256 | 12.3255 |
61
+ | 0.032 | 2.46 | 1000 | 0.1247 | 11.7262 |
 
 
 
 
62
 
63
 
64
  ### Framework versions
runs/Feb25_02-35-41_ubuntu-llama/events.out.tfevents.1708821359.ubuntu-llama.190234.0 CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:23f0e95684a4b52f91c28b9fe61e1c29bd3efcb11cd6573cddbccda45f47342b
3
- size 12879
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:197d474c749403d0a77b39c43c2c810061718fb78382575721f9c68ddbf888cf
3
+ size 13233