DewiBrynJones commited on
Commit
b04062d
·
verified ·
1 Parent(s): 6621ea7

Model save

Browse files
Files changed (2) hide show
  1. README.md +14 -13
  2. generation_config.json +1 -1
README.md CHANGED
@@ -1,4 +1,5 @@
1
  ---
 
2
  license: apache-2.0
3
  base_model: openai/whisper-large-v3
4
  tags:
@@ -15,10 +16,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # whisper-large-v3-ft-cv-cy-en
17
 
18
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the DewiBrynJones/commonvoice_18_0_cy_en train main dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 0.3036
21
- - Wer: 0.1826
22
 
23
  ## Model description
24
 
@@ -43,7 +44,7 @@ The following hyperparameters were used during training:
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
  - total_train_batch_size: 32
46
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: linear
48
  - lr_scheduler_warmup_steps: 500
49
  - training_steps: 5000
@@ -53,16 +54,16 @@ The following hyperparameters were used during training:
53
 
54
  | Training Loss | Epoch | Step | Validation Loss | Wer |
55
  |:-------------:|:------:|:----:|:---------------:|:------:|
56
- | 0.294 | 0.7075 | 1000 | 0.3054 | 0.2114 |
57
- | 0.1572 | 1.4149 | 2000 | 0.2768 | 0.1898 |
58
- | 0.0714 | 2.1224 | 3000 | 0.2789 | 0.1807 |
59
- | 0.0772 | 2.8299 | 4000 | 0.2759 | 0.1810 |
60
- | 0.0337 | 3.5373 | 5000 | 0.3036 | 0.1826 |
61
 
62
 
63
  ### Framework versions
64
 
65
- - Transformers 4.44.0
66
- - Pytorch 2.4.0+cu121
67
- - Datasets 2.20.0
68
- - Tokenizers 0.19.1
 
1
  ---
2
+ library_name: transformers
3
  license: apache-2.0
4
  base_model: openai/whisper-large-v3
5
  tags:
 
16
 
17
  # whisper-large-v3-ft-cv-cy-en
18
 
19
+ This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
+ - Loss: 0.2744
22
+ - Wer: 0.1474
23
 
24
  ## Model description
25
 
 
44
  - seed: 42
45
  - gradient_accumulation_steps: 2
46
  - total_train_batch_size: 32
47
+ - optimizer: Use adamw_torch with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
48
  - lr_scheduler_type: linear
49
  - lr_scheduler_warmup_steps: 500
50
  - training_steps: 5000
 
54
 
55
  | Training Loss | Epoch | Step | Validation Loss | Wer |
56
  |:-------------:|:------:|:----:|:---------------:|:------:|
57
+ | 0.4825 | 0.7075 | 1000 | 0.2708 | 0.1810 |
58
+ | 0.2262 | 1.4149 | 2000 | 0.2486 | 0.1594 |
59
+ | 0.0867 | 2.1224 | 3000 | 0.2506 | 0.1511 |
60
+ | 0.0973 | 2.8299 | 4000 | 0.2444 | 0.1490 |
61
+ | 0.0303 | 3.5373 | 5000 | 0.2744 | 0.1474 |
62
 
63
 
64
  ### Framework versions
65
 
66
+ - Transformers 4.46.1
67
+ - Pytorch 2.5.1+cu124
68
+ - Datasets 3.1.0
69
+ - Tokenizers 0.20.1
generation_config.json CHANGED
@@ -253,5 +253,5 @@
253
  "transcribe": 50360,
254
  "translate": 50359
255
  },
256
- "transformers_version": "4.44.0"
257
  }
 
253
  "transcribe": 50360,
254
  "translate": 50359
255
  },
256
+ "transformers_version": "4.46.1"
257
  }