fsicoli commited on
Commit
59c2a6c
1 Parent(s): f310f04

End of training

Browse files
README.md CHANGED
@@ -4,11 +4,24 @@ license: apache-2.0
4
  base_model: openai/whisper-large-v3
5
  tags:
6
  - generated_from_trainer
 
 
7
  metrics:
8
  - wer
9
  model-index:
10
  - name: whisper-large-v3-pt-cv19-fleurs
11
- results: []
 
 
 
 
 
 
 
 
 
 
 
12
  ---
13
 
14
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -16,10 +29,10 @@ should probably proofread and complete it, then remove this comment. -->
16
 
17
  # whisper-large-v3-pt-cv19-fleurs
18
 
19
- This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on an unknown dataset.
20
  It achieves the following results on the evaluation set:
21
- - Loss: 0.1848
22
- - Wer: 0.0756
23
 
24
  ## Model description
25
 
 
4
  base_model: openai/whisper-large-v3
5
  tags:
6
  - generated_from_trainer
7
+ datasets:
8
+ - fsicoli/cv19-fleurs
9
  metrics:
10
  - wer
11
  model-index:
12
  - name: whisper-large-v3-pt-cv19-fleurs
13
+ results:
14
+ - task:
15
+ name: Automatic Speech Recognition
16
+ type: automatic-speech-recognition
17
+ dataset:
18
+ name: fsicoli/cv19-fleurs default
19
+ type: fsicoli/cv19-fleurs
20
+ args: default
21
+ metrics:
22
+ - name: Wer
23
+ type: wer
24
+ value: 0.08493312855954438
25
  ---
26
 
27
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
29
 
30
  # whisper-large-v3-pt-cv19-fleurs
31
 
32
+ This model is a fine-tuned version of [openai/whisper-large-v3](https://huggingface.co/openai/whisper-large-v3) on the fsicoli/cv19-fleurs default dataset.
33
  It achieves the following results on the evaluation set:
34
+ - Loss: 0.1823
35
+ - Wer: 0.0849
36
 
37
  ## Model description
38
 
all_results.json CHANGED
@@ -1,9 +1,15 @@
1
  {
2
- "epoch": 22.88329519450801,
3
- "total_flos": 2.7176996653970227e+21,
4
- "train_loss": 0.0002151485295407474,
5
- "train_runtime": 70560.5564,
 
 
 
 
 
 
6
  "train_samples": 34956,
7
- "train_samples_per_second": 11.338,
8
- "train_steps_per_second": 0.709
9
  }
 
1
  {
2
+ "epoch": 22.8837528604119,
3
+ "eval_loss": 0.18233540654182434,
4
+ "eval_runtime": 8499.2837,
5
+ "eval_samples": 9522,
6
+ "eval_samples_per_second": 1.12,
7
+ "eval_steps_per_second": 0.14,
8
+ "eval_wer": 0.08493312855954438,
9
+ "total_flos": 2.7177540253699277e+21,
10
+ "train_loss": 1.0213549286778259e-09,
11
+ "train_runtime": 167.9133,
12
  "train_samples": 34956,
13
+ "train_samples_per_second": 4764.363,
14
+ "train_steps_per_second": 297.773
15
  }
eval_results.json ADDED
@@ -0,0 +1,9 @@
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "epoch": 22.8837528604119,
3
+ "eval_loss": 0.18233540654182434,
4
+ "eval_runtime": 8499.2837,
5
+ "eval_samples": 9522,
6
+ "eval_samples_per_second": 1.12,
7
+ "eval_steps_per_second": 0.14,
8
+ "eval_wer": 0.08493312855954438
9
+ }
runs/Oct02_12-22-25_DITEC2014063010/events.out.tfevents.1727891763.DITEC2014063010 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7f2fdd4fcd2d717ccacdd086e3dfdf811cde85f14d01144f205ed5ce1da81bc1
3
+ size 40
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 22.88329519450801,
3
- "total_flos": 2.7176996653970227e+21,
4
- "train_loss": 0.0002151485295407474,
5
- "train_runtime": 70560.5564,
6
  "train_samples": 34956,
7
- "train_samples_per_second": 11.338,
8
- "train_steps_per_second": 0.709
9
  }
 
1
  {
2
+ "epoch": 22.8837528604119,
3
+ "total_flos": 2.7177540253699277e+21,
4
+ "train_loss": 1.0213549286778259e-09,
5
+ "train_runtime": 167.9133,
6
  "train_samples": 34956,
7
+ "train_samples_per_second": 4764.363,
8
+ "train_steps_per_second": 297.773
9
  }
trainer_state.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
  "best_metric": 0.08493312855954438,
3
  "best_model_checkpoint": "d:\\\\whisper-large-v3-pt-cv19-fleurs\\checkpoint-25000",
4
- "epoch": 22.88329519450801,
5
  "eval_steps": 5000,
6
- "global_step": 50000,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
@@ -14099,13 +14099,13 @@
14099
  "step": 50000
14100
  },
14101
  {
14102
- "epoch": 22.88329519450801,
14103
- "step": 50000,
14104
- "total_flos": 2.7176996653970227e+21,
14105
- "train_loss": 0.0002151485295407474,
14106
- "train_runtime": 70560.5564,
14107
- "train_samples_per_second": 11.338,
14108
- "train_steps_per_second": 0.709
14109
  }
14110
  ],
14111
  "logging_steps": 25,
@@ -14125,7 +14125,7 @@
14125
  "attributes": {}
14126
  }
14127
  },
14128
- "total_flos": 2.7176996653970227e+21,
14129
  "train_batch_size": 8,
14130
  "trial_name": null,
14131
  "trial_params": null
 
1
  {
2
  "best_metric": 0.08493312855954438,
3
  "best_model_checkpoint": "d:\\\\whisper-large-v3-pt-cv19-fleurs\\checkpoint-25000",
4
+ "epoch": 22.8837528604119,
5
  "eval_steps": 5000,
6
+ "global_step": 50001,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
 
14099
  "step": 50000
14100
  },
14101
  {
14102
+ "epoch": 22.8837528604119,
14103
+ "step": 50001,
14104
+ "total_flos": 2.7177540253699277e+21,
14105
+ "train_loss": 1.0213549286778259e-09,
14106
+ "train_runtime": 167.9133,
14107
+ "train_samples_per_second": 4764.363,
14108
+ "train_steps_per_second": 297.773
14109
  }
14110
  ],
14111
  "logging_steps": 25,
 
14125
  "attributes": {}
14126
  }
14127
  },
14128
+ "total_flos": 2.7177540253699277e+21,
14129
  "train_batch_size": 8,
14130
  "trial_name": null,
14131
  "trial_params": null