blackhole33 commited on
Commit
453cfa6
1 Parent(s): 54b5d79

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +94 -3
README.md CHANGED
@@ -1,3 +1,94 @@
1
- ---
2
- license: apache-2.0
3
- ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ datasets:
4
+ - mozilla-foundation/common_voice_17_0
5
+ language:
6
+ - uz
7
+ metrics:
8
+ - wer
9
+ library_name: transformers
10
+ pipeline_tag: automatic-speech-recognition
11
+ ---
12
+ language:
13
+ - uz
14
+ license: apache-2.0
15
+ base_model: openai/whisper-medium
16
+ tags:
17
+ - generated_from_trainer
18
+ datasets:
19
+ - mozilla-foundation/common_voice_17_0
20
+ metrics:
21
+ - wer
22
+ model-index:
23
+ - name: Whisper Mediu, UZB - AISHA
24
+ results:
25
+ - task:
26
+ name: Automatic Speech Recognition
27
+ type: automatic-speech-recognition
28
+ dataset:
29
+ name: Common Voice 17.0
30
+ type: mozilla-foundation/common_voice_17_0
31
+ config: uz
32
+ split: None
33
+ args: 'config: uz, split: test'
34
+ metrics:
35
+ - name: Wer
36
+ type: wer
37
+ value: 31.77905998468049
38
+ ---
39
+
40
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
41
+ should probably proofread and complete it, then remove this comment. -->
42
+
43
+ # Whisper Medium UZB - AISHA
44
+
45
+ This model is a fine-tuned version of [openai/whisper-medium](https://huggingface.co/openai/whisper-medium) on the Common Voice 17.0 dataset.
46
+ It achieves the following results on the evaluation set:
47
+ - Loss: 0.2859
48
+ - Wer: 31.7790
49
+
50
+ ## Model description
51
+
52
+ More information needed
53
+
54
+ ## Intended uses & limitations
55
+
56
+ More information needed
57
+
58
+ ## Training and evaluation data
59
+
60
+ More information needed
61
+
62
+ ## Training procedure
63
+
64
+ ### Training hyperparameters
65
+
66
+ The following hyperparameters were used during training:
67
+ - learning_rate: 1e-05
68
+ - train_batch_size: 16
69
+ - eval_batch_size: 8
70
+ - seed: 42
71
+ - gradient_accumulation_steps: 2
72
+ - total_train_batch_size: 32
73
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
74
+ - lr_scheduler_type: linear
75
+ - lr_scheduler_warmup_steps: 500
76
+ - training_steps: 4000
77
+ - mixed_precision_training: Native AMP
78
+
79
+ ### Training results
80
+
81
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
82
+ |:-------------:|:------:|:----:|:---------------:|:-------:|
83
+ | 0.5187 | 0.5392 | 1000 | 0.4935 | 44.1403 |
84
+ | 0.3423 | 1.0785 | 2000 | 0.4008 | 37.6948 |
85
+ | 0.3018 | 1.6177 | 3000 | 0.3739 | 36.3575 |
86
+ | 0.2401 | 2.1569 | 4000 | 0.2821 | 31.7791 |
87
+
88
+
89
+ ### Framework versions
90
+
91
+ - Transformers 4.41.2
92
+ - Pytorch 2.3.1+cu121
93
+ - Datasets 2.20.0
94
+ - Tokenizers 0.19.1