End of training
Browse files- README.md +4 -4
- all_results.json +11 -11
- eval_results.json +6 -6
- train_results.json +6 -6
- trainer_state.json +0 -0
README.md
CHANGED
@@ -4,7 +4,7 @@ base_model: openai/whisper-small
|
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
-
- common_voice_17_0
|
8 |
metrics:
|
9 |
- wer
|
10 |
model-index:
|
@@ -14,8 +14,8 @@ model-index:
|
|
14 |
name: Automatic Speech Recognition
|
15 |
type: automatic-speech-recognition
|
16 |
dataset:
|
17 |
-
name: common_voice_17_0
|
18 |
-
type: common_voice_17_0
|
19 |
config: id
|
20 |
split: None
|
21 |
args: id
|
@@ -30,7 +30,7 @@ should probably proofread and complete it, then remove this comment. -->
|
|
30 |
|
31 |
# whisper-small-id
|
32 |
|
33 |
-
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the common_voice_17_0 dataset.
|
34 |
It achieves the following results on the evaluation set:
|
35 |
- Loss: 0.0878
|
36 |
- Wer: 0.0590
|
|
|
4 |
tags:
|
5 |
- generated_from_trainer
|
6 |
datasets:
|
7 |
+
- mozilla-foundation/common_voice_17_0
|
8 |
metrics:
|
9 |
- wer
|
10 |
model-index:
|
|
|
14 |
name: Automatic Speech Recognition
|
15 |
type: automatic-speech-recognition
|
16 |
dataset:
|
17 |
+
name: mozilla-foundation/common_voice_17_0 id
|
18 |
+
type: mozilla-foundation/common_voice_17_0
|
19 |
config: id
|
20 |
split: None
|
21 |
args: id
|
|
|
30 |
|
31 |
# whisper-small-id
|
32 |
|
33 |
+
This model is a fine-tuned version of [openai/whisper-small](https://huggingface.co/openai/whisper-small) on the mozilla-foundation/common_voice_17_0 id dataset.
|
34 |
It achieves the following results on the evaluation set:
|
35 |
- Loss: 0.0878
|
36 |
- Wer: 0.0590
|
all_results.json
CHANGED
@@ -1,15 +1,15 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"eval_loss": 0.
|
4 |
-
"eval_runtime":
|
5 |
"eval_samples": 3641,
|
6 |
-
"eval_samples_per_second": 8.
|
7 |
-
"eval_steps_per_second": 0.
|
8 |
-
"eval_wer": 0.
|
9 |
-
"total_flos":
|
10 |
-
"train_loss": 0.
|
11 |
-
"train_runtime":
|
12 |
-
"train_samples":
|
13 |
-
"train_samples_per_second": 8.
|
14 |
"train_steps_per_second": 0.275
|
15 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 16.913319238900634,
|
3 |
+
"eval_loss": 0.08780684322118759,
|
4 |
+
"eval_runtime": 417.4654,
|
5 |
"eval_samples": 3641,
|
6 |
+
"eval_samples_per_second": 8.722,
|
7 |
+
"eval_steps_per_second": 0.546,
|
8 |
+
"eval_wer": 0.05902826117221217,
|
9 |
+
"total_flos": 1.8463463044153344e+20,
|
10 |
+
"train_loss": 0.02644415222366224,
|
11 |
+
"train_runtime": 72777.8888,
|
12 |
+
"train_samples": 37827,
|
13 |
+
"train_samples_per_second": 8.794,
|
14 |
"train_steps_per_second": 0.275
|
15 |
}
|
eval_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"eval_loss": 0.
|
4 |
-
"eval_runtime":
|
5 |
"eval_samples": 3641,
|
6 |
-
"eval_samples_per_second": 8.
|
7 |
-
"eval_steps_per_second": 0.
|
8 |
-
"eval_wer": 0.
|
9 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 16.913319238900634,
|
3 |
+
"eval_loss": 0.08780684322118759,
|
4 |
+
"eval_runtime": 417.4654,
|
5 |
"eval_samples": 3641,
|
6 |
+
"eval_samples_per_second": 8.722,
|
7 |
+
"eval_steps_per_second": 0.546,
|
8 |
+
"eval_wer": 0.05902826117221217
|
9 |
}
|
train_results.json
CHANGED
@@ -1,9 +1,9 @@
|
|
1 |
{
|
2 |
-
"epoch":
|
3 |
-
"total_flos":
|
4 |
-
"train_loss": 0.
|
5 |
-
"train_runtime":
|
6 |
-
"train_samples":
|
7 |
-
"train_samples_per_second": 8.
|
8 |
"train_steps_per_second": 0.275
|
9 |
}
|
|
|
1 |
{
|
2 |
+
"epoch": 16.913319238900634,
|
3 |
+
"total_flos": 1.8463463044153344e+20,
|
4 |
+
"train_loss": 0.02644415222366224,
|
5 |
+
"train_runtime": 72777.8888,
|
6 |
+
"train_samples": 37827,
|
7 |
+
"train_samples_per_second": 8.794,
|
8 |
"train_steps_per_second": 0.275
|
9 |
}
|
trainer_state.json
CHANGED
The diff for this file is too large to render.
See raw diff
|
|