AMoustakis commited on
Commit
8b30b3a
1 Parent(s): d7e7c16

End of training

Browse files
README.md ADDED
@@ -0,0 +1,80 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ language:
3
+ - el
4
+ license: apache-2.0
5
+ base_model: openai/whisper-base
6
+ tags:
7
+ - generated_from_trainer
8
+ datasets:
9
+ - AMoustakis/test-dataset
10
+ metrics:
11
+ - wer
12
+ model-index:
13
+ - name: Whisper Base Greek
14
+ results:
15
+ - task:
16
+ name: Automatic Speech Recognition
17
+ type: automatic-speech-recognition
18
+ dataset:
19
+ name: Test Dataset for greek language
20
+ type: AMoustakis/test-dataset
21
+ args: 'split: train'
22
+ metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 61.77777777777778
26
+ ---
27
+
28
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
+ should probably proofread and complete it, then remove this comment. -->
30
+
31
+ # Whisper Base Greek
32
+
33
+ This model is a fine-tuned version of [openai/whisper-base](https://huggingface.co/openai/whisper-base) on the Test Dataset for greek language dataset.
34
+ It achieves the following results on the evaluation set:
35
+ - Loss: 0.2675
36
+ - Wer: 61.7778
37
+
38
+ ## Model description
39
+
40
+ More information needed
41
+
42
+ ## Intended uses & limitations
43
+
44
+ More information needed
45
+
46
+ ## Training and evaluation data
47
+
48
+ More information needed
49
+
50
+ ## Training procedure
51
+
52
+ ### Training hyperparameters
53
+
54
+ The following hyperparameters were used during training:
55
+ - learning_rate: 0.001
56
+ - train_batch_size: 3
57
+ - eval_batch_size: 1
58
+ - seed: 42
59
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
+ - lr_scheduler_type: linear
61
+ - num_epochs: 5
62
+ - mixed_precision_training: Native AMP
63
+
64
+ ### Training results
65
+
66
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
67
+ |:-------------:|:-----:|:----:|:---------------:|:-------:|
68
+ | 0.9014 | 1.0 | 4 | 0.5733 | 74.5185 |
69
+ | 0.4829 | 2.0 | 8 | 0.4158 | 64.4444 |
70
+ | 0.5963 | 3.0 | 12 | 0.3257 | 65.0370 |
71
+ | 0.3399 | 4.0 | 16 | 0.2857 | 61.7778 |
72
+ | 0.4436 | 5.0 | 20 | 0.2675 | 61.7778 |
73
+
74
+
75
+ ### Framework versions
76
+
77
+ - Transformers 4.40.2
78
+ - Pytorch 2.2.1+cu121
79
+ - Datasets 2.19.1
80
+ - Tokenizers 0.19.1
preprocessor_config.json ADDED
@@ -0,0 +1,14 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "chunk_length": 30,
3
+ "feature_extractor_type": "WhisperFeatureExtractor",
4
+ "feature_size": 80,
5
+ "hop_length": 160,
6
+ "n_fft": 400,
7
+ "n_samples": 480000,
8
+ "nb_max_frames": 3000,
9
+ "padding_side": "right",
10
+ "padding_value": 0.0,
11
+ "processor_class": "WhisperProcessor",
12
+ "return_attention_mask": false,
13
+ "sampling_rate": 16000
14
+ }
pytorch_model.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:103eb0a9d2b29bbcf54565c9178193b7d251f12831d55320653e11e679049eb9
3
+ size 295254590
runs/May15_12-27-08_4bed3e405a5e/events.out.tfevents.1715776033.4bed3e405a5e.5402.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d6a1923766c28d774259592ef65f0732d913bf06c607b23d43e8c48939e82d18
3
+ size 7210
runs/May15_12-28-12_4bed3e405a5e/events.out.tfevents.1715776097.4bed3e405a5e.5402.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:a2d19915626fa479dabd795f7db9016459cc02a927172c0f444b9959114f3c5a
3
+ size 7210
runs/May15_12-31-46_4bed3e405a5e/events.out.tfevents.1715776310.4bed3e405a5e.5402.2 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:294d9fce20e37dc25b50e310e813f5fab73db144e40a5be6d9197b04569d9170
3
+ size 12431
training_args.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aa93e0663f16d1818061b3bc752a2bf6b6339da00bac6c8c7f4537c41e1f66e9
3
+ size 5176