Commit
•
de822a1
1
Parent(s):
cd6086b
End of training
Browse files- trainer_state.json +2517 -17
trainer_state.json
CHANGED
@@ -1,8 +1,8 @@
|
|
1 |
{
|
2 |
"best_metric": 18.48141795311607,
|
3 |
"best_model_checkpoint": "whisper-base-nl-3/checkpoint-29000",
|
4 |
-
"epoch":
|
5 |
-
"global_step":
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
@@ -8758,28 +8758,2528 @@
|
|
8758 |
"step": 35000
|
8759 |
},
|
8760 |
{
|
8761 |
-
"epoch": 38.
|
8762 |
-
"
|
8763 |
-
"
|
8764 |
-
"
|
8765 |
-
|
8766 |
-
|
8767 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8768 |
},
|
8769 |
{
|
8770 |
-
"epoch":
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
8771 |
"eval_cer": 9.956974355224656,
|
8772 |
"eval_loss": 0.7910374999046326,
|
8773 |
-
"eval_runtime":
|
8774 |
-
"eval_samples_per_second": 0.
|
8775 |
-
"eval_steps_per_second": 0.
|
8776 |
"eval_wer": 31.400474471700036,
|
8777 |
-
"step":
|
8778 |
}
|
8779 |
],
|
8780 |
-
"max_steps":
|
8781 |
-
"num_train_epochs":
|
8782 |
-
"total_flos":
|
8783 |
"trial_name": null,
|
8784 |
"trial_params": null
|
8785 |
}
|
|
|
1 |
{
|
2 |
"best_metric": 18.48141795311607,
|
3 |
"best_model_checkpoint": "whisper-base-nl-3/checkpoint-29000",
|
4 |
+
"epoch": 49.07306434023991,
|
5 |
+
"global_step": 45000,
|
6 |
"is_hyper_param_search": false,
|
7 |
"is_local_process_zero": true,
|
8 |
"is_world_process_zero": true,
|
|
|
8758 |
"step": 35000
|
8759 |
},
|
8760 |
{
|
8761 |
+
"epoch": 38.2,
|
8762 |
+
"learning_rate": 2.244943820224719e-06,
|
8763 |
+
"loss": 0.5112,
|
8764 |
+
"step": 35025
|
8765 |
+
},
|
8766 |
+
{
|
8767 |
+
"epoch": 38.22,
|
8768 |
+
"learning_rate": 2.2393258426966294e-06,
|
8769 |
+
"loss": 0.4629,
|
8770 |
+
"step": 35050
|
8771 |
+
},
|
8772 |
+
{
|
8773 |
+
"epoch": 38.25,
|
8774 |
+
"learning_rate": 2.2337078651685396e-06,
|
8775 |
+
"loss": 0.4797,
|
8776 |
+
"step": 35075
|
8777 |
+
},
|
8778 |
+
{
|
8779 |
+
"epoch": 38.28,
|
8780 |
+
"learning_rate": 2.22808988764045e-06,
|
8781 |
+
"loss": 0.4799,
|
8782 |
+
"step": 35100
|
8783 |
+
},
|
8784 |
+
{
|
8785 |
+
"epoch": 38.3,
|
8786 |
+
"learning_rate": 2.22247191011236e-06,
|
8787 |
+
"loss": 0.4919,
|
8788 |
+
"step": 35125
|
8789 |
+
},
|
8790 |
+
{
|
8791 |
+
"epoch": 38.33,
|
8792 |
+
"learning_rate": 2.21685393258427e-06,
|
8793 |
+
"loss": 0.4669,
|
8794 |
+
"step": 35150
|
8795 |
+
},
|
8796 |
+
{
|
8797 |
+
"epoch": 38.36,
|
8798 |
+
"learning_rate": 2.21123595505618e-06,
|
8799 |
+
"loss": 0.4348,
|
8800 |
+
"step": 35175
|
8801 |
+
},
|
8802 |
+
{
|
8803 |
+
"epoch": 38.39,
|
8804 |
+
"learning_rate": 2.20561797752809e-06,
|
8805 |
+
"loss": 0.5559,
|
8806 |
+
"step": 35200
|
8807 |
+
},
|
8808 |
+
{
|
8809 |
+
"epoch": 38.41,
|
8810 |
+
"learning_rate": 2.2e-06,
|
8811 |
+
"loss": 0.4651,
|
8812 |
+
"step": 35225
|
8813 |
+
},
|
8814 |
+
{
|
8815 |
+
"epoch": 38.44,
|
8816 |
+
"learning_rate": 2.1943820224719103e-06,
|
8817 |
+
"loss": 0.4023,
|
8818 |
+
"step": 35250
|
8819 |
+
},
|
8820 |
+
{
|
8821 |
+
"epoch": 38.47,
|
8822 |
+
"learning_rate": 2.1887640449438206e-06,
|
8823 |
+
"loss": 0.4687,
|
8824 |
+
"step": 35275
|
8825 |
+
},
|
8826 |
+
{
|
8827 |
+
"epoch": 38.5,
|
8828 |
+
"learning_rate": 2.183146067415731e-06,
|
8829 |
+
"loss": 0.4999,
|
8830 |
+
"step": 35300
|
8831 |
+
},
|
8832 |
+
{
|
8833 |
+
"epoch": 38.52,
|
8834 |
+
"learning_rate": 2.1775280898876406e-06,
|
8835 |
+
"loss": 0.431,
|
8836 |
+
"step": 35325
|
8837 |
+
},
|
8838 |
+
{
|
8839 |
+
"epoch": 38.55,
|
8840 |
+
"learning_rate": 2.171910112359551e-06,
|
8841 |
+
"loss": 0.4887,
|
8842 |
+
"step": 35350
|
8843 |
+
},
|
8844 |
+
{
|
8845 |
+
"epoch": 38.58,
|
8846 |
+
"learning_rate": 2.1662921348314606e-06,
|
8847 |
+
"loss": 0.4401,
|
8848 |
+
"step": 35375
|
8849 |
+
},
|
8850 |
+
{
|
8851 |
+
"epoch": 38.6,
|
8852 |
+
"learning_rate": 2.160674157303371e-06,
|
8853 |
+
"loss": 0.4975,
|
8854 |
+
"step": 35400
|
8855 |
+
},
|
8856 |
+
{
|
8857 |
+
"epoch": 38.63,
|
8858 |
+
"learning_rate": 2.155056179775281e-06,
|
8859 |
+
"loss": 0.4447,
|
8860 |
+
"step": 35425
|
8861 |
+
},
|
8862 |
+
{
|
8863 |
+
"epoch": 38.66,
|
8864 |
+
"learning_rate": 2.1494382022471913e-06,
|
8865 |
+
"loss": 0.5491,
|
8866 |
+
"step": 35450
|
8867 |
+
},
|
8868 |
+
{
|
8869 |
+
"epoch": 38.69,
|
8870 |
+
"learning_rate": 2.143820224719101e-06,
|
8871 |
+
"loss": 0.4185,
|
8872 |
+
"step": 35475
|
8873 |
+
},
|
8874 |
+
{
|
8875 |
+
"epoch": 38.71,
|
8876 |
+
"learning_rate": 2.1382022471910113e-06,
|
8877 |
+
"loss": 0.4195,
|
8878 |
+
"step": 35500
|
8879 |
+
},
|
8880 |
+
{
|
8881 |
+
"epoch": 38.74,
|
8882 |
+
"learning_rate": 2.1325842696629216e-06,
|
8883 |
+
"loss": 0.5616,
|
8884 |
+
"step": 35525
|
8885 |
+
},
|
8886 |
+
{
|
8887 |
+
"epoch": 38.77,
|
8888 |
+
"learning_rate": 2.126966292134832e-06,
|
8889 |
+
"loss": 0.4316,
|
8890 |
+
"step": 35550
|
8891 |
+
},
|
8892 |
+
{
|
8893 |
+
"epoch": 38.79,
|
8894 |
+
"learning_rate": 2.1213483146067416e-06,
|
8895 |
+
"loss": 0.4601,
|
8896 |
+
"step": 35575
|
8897 |
+
},
|
8898 |
+
{
|
8899 |
+
"epoch": 38.82,
|
8900 |
+
"learning_rate": 2.1159550561797753e-06,
|
8901 |
+
"loss": 0.4268,
|
8902 |
+
"step": 35600
|
8903 |
+
},
|
8904 |
+
{
|
8905 |
+
"epoch": 38.85,
|
8906 |
+
"learning_rate": 2.1103370786516855e-06,
|
8907 |
+
"loss": 0.4646,
|
8908 |
+
"step": 35625
|
8909 |
+
},
|
8910 |
+
{
|
8911 |
+
"epoch": 38.88,
|
8912 |
+
"learning_rate": 2.1047191011235957e-06,
|
8913 |
+
"loss": 0.4052,
|
8914 |
+
"step": 35650
|
8915 |
+
},
|
8916 |
+
{
|
8917 |
+
"epoch": 38.9,
|
8918 |
+
"learning_rate": 2.099101123595506e-06,
|
8919 |
+
"loss": 0.4437,
|
8920 |
+
"step": 35675
|
8921 |
+
},
|
8922 |
+
{
|
8923 |
+
"epoch": 38.93,
|
8924 |
+
"learning_rate": 2.0934831460674158e-06,
|
8925 |
+
"loss": 0.4204,
|
8926 |
+
"step": 35700
|
8927 |
+
},
|
8928 |
+
{
|
8929 |
+
"epoch": 38.96,
|
8930 |
+
"learning_rate": 2.087865168539326e-06,
|
8931 |
+
"loss": 0.4344,
|
8932 |
+
"step": 35725
|
8933 |
+
},
|
8934 |
+
{
|
8935 |
+
"epoch": 38.99,
|
8936 |
+
"learning_rate": 2.082247191011236e-06,
|
8937 |
+
"loss": 0.5028,
|
8938 |
+
"step": 35750
|
8939 |
+
},
|
8940 |
+
{
|
8941 |
+
"epoch": 39.01,
|
8942 |
+
"learning_rate": 2.076629213483146e-06,
|
8943 |
+
"loss": 0.3716,
|
8944 |
+
"step": 35775
|
8945 |
+
},
|
8946 |
+
{
|
8947 |
+
"epoch": 39.04,
|
8948 |
+
"learning_rate": 2.0710112359550563e-06,
|
8949 |
+
"loss": 0.4251,
|
8950 |
+
"step": 35800
|
8951 |
+
},
|
8952 |
+
{
|
8953 |
+
"epoch": 39.07,
|
8954 |
+
"learning_rate": 2.0653932584269665e-06,
|
8955 |
+
"loss": 0.4155,
|
8956 |
+
"step": 35825
|
8957 |
+
},
|
8958 |
+
{
|
8959 |
+
"epoch": 39.09,
|
8960 |
+
"learning_rate": 2.0597752808988767e-06,
|
8961 |
+
"loss": 0.4249,
|
8962 |
+
"step": 35850
|
8963 |
+
},
|
8964 |
+
{
|
8965 |
+
"epoch": 39.12,
|
8966 |
+
"learning_rate": 2.054157303370787e-06,
|
8967 |
+
"loss": 0.3721,
|
8968 |
+
"step": 35875
|
8969 |
+
},
|
8970 |
+
{
|
8971 |
+
"epoch": 39.15,
|
8972 |
+
"learning_rate": 2.0485393258426968e-06,
|
8973 |
+
"loss": 0.4026,
|
8974 |
+
"step": 35900
|
8975 |
+
},
|
8976 |
+
{
|
8977 |
+
"epoch": 39.18,
|
8978 |
+
"learning_rate": 2.042921348314607e-06,
|
8979 |
+
"loss": 0.4425,
|
8980 |
+
"step": 35925
|
8981 |
+
},
|
8982 |
+
{
|
8983 |
+
"epoch": 39.2,
|
8984 |
+
"learning_rate": 2.037303370786517e-06,
|
8985 |
+
"loss": 0.399,
|
8986 |
+
"step": 35950
|
8987 |
+
},
|
8988 |
+
{
|
8989 |
+
"epoch": 39.23,
|
8990 |
+
"learning_rate": 2.031685393258427e-06,
|
8991 |
+
"loss": 0.4291,
|
8992 |
+
"step": 35975
|
8993 |
+
},
|
8994 |
+
{
|
8995 |
+
"epoch": 39.26,
|
8996 |
+
"learning_rate": 2.0260674157303372e-06,
|
8997 |
+
"loss": 0.3674,
|
8998 |
+
"step": 36000
|
8999 |
+
},
|
9000 |
+
{
|
9001 |
+
"epoch": 39.26,
|
9002 |
+
"eval_cer": 7.010413037587582,
|
9003 |
+
"eval_loss": 0.47536423802375793,
|
9004 |
+
"eval_runtime": 3145.4396,
|
9005 |
+
"eval_samples_per_second": 0.689,
|
9006 |
+
"eval_steps_per_second": 0.172,
|
9007 |
+
"eval_wer": 22.664244720915118,
|
9008 |
+
"step": 36000
|
9009 |
+
},
|
9010 |
+
{
|
9011 |
+
"epoch": 39.29,
|
9012 |
+
"learning_rate": 2.0204494382022475e-06,
|
9013 |
+
"loss": 0.3957,
|
9014 |
+
"step": 36025
|
9015 |
+
},
|
9016 |
+
{
|
9017 |
+
"epoch": 39.31,
|
9018 |
+
"learning_rate": 2.0148314606741577e-06,
|
9019 |
+
"loss": 0.3944,
|
9020 |
+
"step": 36050
|
9021 |
+
},
|
9022 |
+
{
|
9023 |
+
"epoch": 39.34,
|
9024 |
+
"learning_rate": 2.0092134831460675e-06,
|
9025 |
+
"loss": 0.4103,
|
9026 |
+
"step": 36075
|
9027 |
+
},
|
9028 |
+
{
|
9029 |
+
"epoch": 39.37,
|
9030 |
+
"learning_rate": 2.0035955056179777e-06,
|
9031 |
+
"loss": 0.3942,
|
9032 |
+
"step": 36100
|
9033 |
+
},
|
9034 |
+
{
|
9035 |
+
"epoch": 39.39,
|
9036 |
+
"learning_rate": 1.997977528089888e-06,
|
9037 |
+
"loss": 0.426,
|
9038 |
+
"step": 36125
|
9039 |
+
},
|
9040 |
+
{
|
9041 |
+
"epoch": 39.42,
|
9042 |
+
"learning_rate": 1.9923595505617978e-06,
|
9043 |
+
"loss": 0.3685,
|
9044 |
+
"step": 36150
|
9045 |
+
},
|
9046 |
+
{
|
9047 |
+
"epoch": 39.45,
|
9048 |
+
"learning_rate": 1.986741573033708e-06,
|
9049 |
+
"loss": 0.4251,
|
9050 |
+
"step": 36175
|
9051 |
+
},
|
9052 |
+
{
|
9053 |
+
"epoch": 39.48,
|
9054 |
+
"learning_rate": 1.9811235955056182e-06,
|
9055 |
+
"loss": 0.3852,
|
9056 |
+
"step": 36200
|
9057 |
+
},
|
9058 |
+
{
|
9059 |
+
"epoch": 39.5,
|
9060 |
+
"learning_rate": 1.9755056179775285e-06,
|
9061 |
+
"loss": 0.3743,
|
9062 |
+
"step": 36225
|
9063 |
+
},
|
9064 |
+
{
|
9065 |
+
"epoch": 39.53,
|
9066 |
+
"learning_rate": 1.9698876404494383e-06,
|
9067 |
+
"loss": 0.4061,
|
9068 |
+
"step": 36250
|
9069 |
+
},
|
9070 |
+
{
|
9071 |
+
"epoch": 39.56,
|
9072 |
+
"learning_rate": 1.9642696629213485e-06,
|
9073 |
+
"loss": 0.4062,
|
9074 |
+
"step": 36275
|
9075 |
+
},
|
9076 |
+
{
|
9077 |
+
"epoch": 39.59,
|
9078 |
+
"learning_rate": 1.9586516853932587e-06,
|
9079 |
+
"loss": 0.3523,
|
9080 |
+
"step": 36300
|
9081 |
+
},
|
9082 |
+
{
|
9083 |
+
"epoch": 39.61,
|
9084 |
+
"learning_rate": 1.9530337078651685e-06,
|
9085 |
+
"loss": 0.4023,
|
9086 |
+
"step": 36325
|
9087 |
+
},
|
9088 |
+
{
|
9089 |
+
"epoch": 39.64,
|
9090 |
+
"learning_rate": 1.9474157303370787e-06,
|
9091 |
+
"loss": 0.3799,
|
9092 |
+
"step": 36350
|
9093 |
+
},
|
9094 |
+
{
|
9095 |
+
"epoch": 39.67,
|
9096 |
+
"learning_rate": 1.941797752808989e-06,
|
9097 |
+
"loss": 0.3439,
|
9098 |
+
"step": 36375
|
9099 |
+
},
|
9100 |
+
{
|
9101 |
+
"epoch": 39.69,
|
9102 |
+
"learning_rate": 1.9361797752808988e-06,
|
9103 |
+
"loss": 0.4106,
|
9104 |
+
"step": 36400
|
9105 |
+
},
|
9106 |
+
{
|
9107 |
+
"epoch": 39.72,
|
9108 |
+
"learning_rate": 1.930561797752809e-06,
|
9109 |
+
"loss": 0.407,
|
9110 |
+
"step": 36425
|
9111 |
+
},
|
9112 |
+
{
|
9113 |
+
"epoch": 39.75,
|
9114 |
+
"learning_rate": 1.9249438202247192e-06,
|
9115 |
+
"loss": 0.3814,
|
9116 |
+
"step": 36450
|
9117 |
+
},
|
9118 |
+
{
|
9119 |
+
"epoch": 39.78,
|
9120 |
+
"learning_rate": 1.9193258426966295e-06,
|
9121 |
+
"loss": 0.3561,
|
9122 |
+
"step": 36475
|
9123 |
+
},
|
9124 |
+
{
|
9125 |
+
"epoch": 39.8,
|
9126 |
+
"learning_rate": 1.9137078651685397e-06,
|
9127 |
+
"loss": 0.3886,
|
9128 |
+
"step": 36500
|
9129 |
+
},
|
9130 |
+
{
|
9131 |
+
"epoch": 39.83,
|
9132 |
+
"learning_rate": 1.9080898876404495e-06,
|
9133 |
+
"loss": 0.3777,
|
9134 |
+
"step": 36525
|
9135 |
+
},
|
9136 |
+
{
|
9137 |
+
"epoch": 39.86,
|
9138 |
+
"learning_rate": 1.9024719101123597e-06,
|
9139 |
+
"loss": 0.39,
|
9140 |
+
"step": 36550
|
9141 |
+
},
|
9142 |
+
{
|
9143 |
+
"epoch": 39.89,
|
9144 |
+
"learning_rate": 1.8968539325842697e-06,
|
9145 |
+
"loss": 0.3945,
|
9146 |
+
"step": 36575
|
9147 |
+
},
|
9148 |
+
{
|
9149 |
+
"epoch": 39.91,
|
9150 |
+
"learning_rate": 1.8912359550561798e-06,
|
9151 |
+
"loss": 0.3531,
|
9152 |
+
"step": 36600
|
9153 |
+
},
|
9154 |
+
{
|
9155 |
+
"epoch": 39.94,
|
9156 |
+
"learning_rate": 1.88561797752809e-06,
|
9157 |
+
"loss": 0.4433,
|
9158 |
+
"step": 36625
|
9159 |
+
},
|
9160 |
+
{
|
9161 |
+
"epoch": 39.97,
|
9162 |
+
"learning_rate": 1.8800000000000002e-06,
|
9163 |
+
"loss": 0.4188,
|
9164 |
+
"step": 36650
|
9165 |
+
},
|
9166 |
+
{
|
9167 |
+
"epoch": 39.99,
|
9168 |
+
"learning_rate": 1.8743820224719102e-06,
|
9169 |
+
"loss": 0.3897,
|
9170 |
+
"step": 36675
|
9171 |
+
},
|
9172 |
+
{
|
9173 |
+
"epoch": 40.02,
|
9174 |
+
"learning_rate": 1.8687640449438205e-06,
|
9175 |
+
"loss": 0.3282,
|
9176 |
+
"step": 36700
|
9177 |
+
},
|
9178 |
+
{
|
9179 |
+
"epoch": 40.05,
|
9180 |
+
"learning_rate": 1.8631460674157307e-06,
|
9181 |
+
"loss": 0.341,
|
9182 |
+
"step": 36725
|
9183 |
+
},
|
9184 |
+
{
|
9185 |
+
"epoch": 40.08,
|
9186 |
+
"learning_rate": 1.8575280898876405e-06,
|
9187 |
+
"loss": 0.3239,
|
9188 |
+
"step": 36750
|
9189 |
+
},
|
9190 |
+
{
|
9191 |
+
"epoch": 40.1,
|
9192 |
+
"learning_rate": 1.8519101123595507e-06,
|
9193 |
+
"loss": 0.3518,
|
9194 |
+
"step": 36775
|
9195 |
+
},
|
9196 |
+
{
|
9197 |
+
"epoch": 40.13,
|
9198 |
+
"learning_rate": 1.8462921348314607e-06,
|
9199 |
+
"loss": 0.3387,
|
9200 |
+
"step": 36800
|
9201 |
+
},
|
9202 |
+
{
|
9203 |
+
"epoch": 40.16,
|
9204 |
+
"learning_rate": 1.840674157303371e-06,
|
9205 |
+
"loss": 0.3201,
|
9206 |
+
"step": 36825
|
9207 |
+
},
|
9208 |
+
{
|
9209 |
+
"epoch": 40.19,
|
9210 |
+
"learning_rate": 1.8350561797752812e-06,
|
9211 |
+
"loss": 0.3128,
|
9212 |
+
"step": 36850
|
9213 |
+
},
|
9214 |
+
{
|
9215 |
+
"epoch": 40.21,
|
9216 |
+
"learning_rate": 1.8294382022471912e-06,
|
9217 |
+
"loss": 0.3749,
|
9218 |
+
"step": 36875
|
9219 |
+
},
|
9220 |
+
{
|
9221 |
+
"epoch": 40.24,
|
9222 |
+
"learning_rate": 1.8238202247191012e-06,
|
9223 |
+
"loss": 0.3756,
|
9224 |
+
"step": 36900
|
9225 |
+
},
|
9226 |
+
{
|
9227 |
+
"epoch": 40.27,
|
9228 |
+
"learning_rate": 1.8182022471910112e-06,
|
9229 |
+
"loss": 0.3545,
|
9230 |
+
"step": 36925
|
9231 |
+
},
|
9232 |
+
{
|
9233 |
+
"epoch": 40.29,
|
9234 |
+
"learning_rate": 1.8125842696629215e-06,
|
9235 |
+
"loss": 0.3746,
|
9236 |
+
"step": 36950
|
9237 |
+
},
|
9238 |
+
{
|
9239 |
+
"epoch": 40.32,
|
9240 |
+
"learning_rate": 1.8069662921348317e-06,
|
9241 |
+
"loss": 0.3308,
|
9242 |
+
"step": 36975
|
9243 |
+
},
|
9244 |
+
{
|
9245 |
+
"epoch": 40.35,
|
9246 |
+
"learning_rate": 1.8013483146067417e-06,
|
9247 |
+
"loss": 0.3481,
|
9248 |
+
"step": 37000
|
9249 |
+
},
|
9250 |
+
{
|
9251 |
+
"epoch": 40.35,
|
9252 |
+
"eval_cer": 7.031058465270519,
|
9253 |
+
"eval_loss": 0.46790996193885803,
|
9254 |
+
"eval_runtime": 3161.6657,
|
9255 |
+
"eval_samples_per_second": 0.685,
|
9256 |
+
"eval_steps_per_second": 0.171,
|
9257 |
+
"eval_wer": 22.631360543685062,
|
9258 |
+
"step": 37000
|
9259 |
+
},
|
9260 |
+
{
|
9261 |
+
"epoch": 40.38,
|
9262 |
+
"learning_rate": 1.795730337078652e-06,
|
9263 |
+
"loss": 0.3225,
|
9264 |
+
"step": 37025
|
9265 |
+
},
|
9266 |
+
{
|
9267 |
+
"epoch": 40.4,
|
9268 |
+
"learning_rate": 1.7901123595505622e-06,
|
9269 |
+
"loss": 0.299,
|
9270 |
+
"step": 37050
|
9271 |
+
},
|
9272 |
+
{
|
9273 |
+
"epoch": 40.43,
|
9274 |
+
"learning_rate": 1.784494382022472e-06,
|
9275 |
+
"loss": 0.3152,
|
9276 |
+
"step": 37075
|
9277 |
+
},
|
9278 |
+
{
|
9279 |
+
"epoch": 40.46,
|
9280 |
+
"learning_rate": 1.778876404494382e-06,
|
9281 |
+
"loss": 0.3268,
|
9282 |
+
"step": 37100
|
9283 |
+
},
|
9284 |
+
{
|
9285 |
+
"epoch": 40.49,
|
9286 |
+
"learning_rate": 1.7732584269662922e-06,
|
9287 |
+
"loss": 0.3411,
|
9288 |
+
"step": 37125
|
9289 |
+
},
|
9290 |
+
{
|
9291 |
+
"epoch": 40.51,
|
9292 |
+
"learning_rate": 1.7676404494382024e-06,
|
9293 |
+
"loss": 0.3846,
|
9294 |
+
"step": 37150
|
9295 |
+
},
|
9296 |
+
{
|
9297 |
+
"epoch": 40.54,
|
9298 |
+
"learning_rate": 1.7620224719101125e-06,
|
9299 |
+
"loss": 0.3168,
|
9300 |
+
"step": 37175
|
9301 |
+
},
|
9302 |
+
{
|
9303 |
+
"epoch": 40.57,
|
9304 |
+
"learning_rate": 1.7564044943820227e-06,
|
9305 |
+
"loss": 0.3377,
|
9306 |
+
"step": 37200
|
9307 |
+
},
|
9308 |
+
{
|
9309 |
+
"epoch": 40.59,
|
9310 |
+
"learning_rate": 1.750786516853933e-06,
|
9311 |
+
"loss": 0.3618,
|
9312 |
+
"step": 37225
|
9313 |
+
},
|
9314 |
+
{
|
9315 |
+
"epoch": 40.62,
|
9316 |
+
"learning_rate": 1.7451685393258427e-06,
|
9317 |
+
"loss": 0.3215,
|
9318 |
+
"step": 37250
|
9319 |
+
},
|
9320 |
+
{
|
9321 |
+
"epoch": 40.65,
|
9322 |
+
"learning_rate": 1.739550561797753e-06,
|
9323 |
+
"loss": 0.3859,
|
9324 |
+
"step": 37275
|
9325 |
+
},
|
9326 |
+
{
|
9327 |
+
"epoch": 40.68,
|
9328 |
+
"learning_rate": 1.733932584269663e-06,
|
9329 |
+
"loss": 0.3484,
|
9330 |
+
"step": 37300
|
9331 |
+
},
|
9332 |
+
{
|
9333 |
+
"epoch": 40.7,
|
9334 |
+
"learning_rate": 1.7283146067415732e-06,
|
9335 |
+
"loss": 0.375,
|
9336 |
+
"step": 37325
|
9337 |
+
},
|
9338 |
+
{
|
9339 |
+
"epoch": 40.73,
|
9340 |
+
"learning_rate": 1.7226966292134834e-06,
|
9341 |
+
"loss": 0.327,
|
9342 |
+
"step": 37350
|
9343 |
+
},
|
9344 |
+
{
|
9345 |
+
"epoch": 40.76,
|
9346 |
+
"learning_rate": 1.7170786516853934e-06,
|
9347 |
+
"loss": 0.2798,
|
9348 |
+
"step": 37375
|
9349 |
+
},
|
9350 |
+
{
|
9351 |
+
"epoch": 40.79,
|
9352 |
+
"learning_rate": 1.7114606741573035e-06,
|
9353 |
+
"loss": 0.3273,
|
9354 |
+
"step": 37400
|
9355 |
+
},
|
9356 |
+
{
|
9357 |
+
"epoch": 40.81,
|
9358 |
+
"learning_rate": 1.7058426966292135e-06,
|
9359 |
+
"loss": 0.3762,
|
9360 |
+
"step": 37425
|
9361 |
+
},
|
9362 |
+
{
|
9363 |
+
"epoch": 40.84,
|
9364 |
+
"learning_rate": 1.7002247191011237e-06,
|
9365 |
+
"loss": 0.2886,
|
9366 |
+
"step": 37450
|
9367 |
+
},
|
9368 |
+
{
|
9369 |
+
"epoch": 40.87,
|
9370 |
+
"learning_rate": 1.6948314606741576e-06,
|
9371 |
+
"loss": 0.3368,
|
9372 |
+
"step": 37475
|
9373 |
+
},
|
9374 |
+
{
|
9375 |
+
"epoch": 40.89,
|
9376 |
+
"learning_rate": 1.6892134831460674e-06,
|
9377 |
+
"loss": 0.3812,
|
9378 |
+
"step": 37500
|
9379 |
+
},
|
9380 |
+
{
|
9381 |
+
"epoch": 40.92,
|
9382 |
+
"learning_rate": 1.6835955056179776e-06,
|
9383 |
+
"loss": 0.3301,
|
9384 |
+
"step": 37525
|
9385 |
+
},
|
9386 |
+
{
|
9387 |
+
"epoch": 40.95,
|
9388 |
+
"learning_rate": 1.6779775280898876e-06,
|
9389 |
+
"loss": 0.3193,
|
9390 |
+
"step": 37550
|
9391 |
+
},
|
9392 |
+
{
|
9393 |
+
"epoch": 40.98,
|
9394 |
+
"learning_rate": 1.6723595505617979e-06,
|
9395 |
+
"loss": 0.3263,
|
9396 |
+
"step": 37575
|
9397 |
+
},
|
9398 |
+
{
|
9399 |
+
"epoch": 41.0,
|
9400 |
+
"learning_rate": 1.666741573033708e-06,
|
9401 |
+
"loss": 0.3422,
|
9402 |
+
"step": 37600
|
9403 |
+
},
|
9404 |
+
{
|
9405 |
+
"epoch": 41.03,
|
9406 |
+
"learning_rate": 1.6611235955056181e-06,
|
9407 |
+
"loss": 0.3138,
|
9408 |
+
"step": 37625
|
9409 |
+
},
|
9410 |
+
{
|
9411 |
+
"epoch": 41.06,
|
9412 |
+
"learning_rate": 1.6555056179775283e-06,
|
9413 |
+
"loss": 0.2919,
|
9414 |
+
"step": 37650
|
9415 |
+
},
|
9416 |
+
{
|
9417 |
+
"epoch": 41.09,
|
9418 |
+
"learning_rate": 1.6498876404494381e-06,
|
9419 |
+
"loss": 0.2828,
|
9420 |
+
"step": 37675
|
9421 |
+
},
|
9422 |
+
{
|
9423 |
+
"epoch": 41.11,
|
9424 |
+
"learning_rate": 1.6442696629213484e-06,
|
9425 |
+
"loss": 0.2607,
|
9426 |
+
"step": 37700
|
9427 |
+
},
|
9428 |
+
{
|
9429 |
+
"epoch": 41.14,
|
9430 |
+
"learning_rate": 1.6386516853932586e-06,
|
9431 |
+
"loss": 0.2524,
|
9432 |
+
"step": 37725
|
9433 |
+
},
|
9434 |
+
{
|
9435 |
+
"epoch": 41.17,
|
9436 |
+
"learning_rate": 1.6330337078651686e-06,
|
9437 |
+
"loss": 0.3627,
|
9438 |
+
"step": 37750
|
9439 |
+
},
|
9440 |
+
{
|
9441 |
+
"epoch": 41.19,
|
9442 |
+
"learning_rate": 1.6274157303370788e-06,
|
9443 |
+
"loss": 0.2641,
|
9444 |
+
"step": 37775
|
9445 |
+
},
|
9446 |
+
{
|
9447 |
+
"epoch": 41.22,
|
9448 |
+
"learning_rate": 1.621797752808989e-06,
|
9449 |
+
"loss": 0.2859,
|
9450 |
+
"step": 37800
|
9451 |
+
},
|
9452 |
+
{
|
9453 |
+
"epoch": 41.25,
|
9454 |
+
"learning_rate": 1.616179775280899e-06,
|
9455 |
+
"loss": 0.277,
|
9456 |
+
"step": 37825
|
9457 |
+
},
|
9458 |
+
{
|
9459 |
+
"epoch": 41.28,
|
9460 |
+
"learning_rate": 1.610561797752809e-06,
|
9461 |
+
"loss": 0.2923,
|
9462 |
+
"step": 37850
|
9463 |
+
},
|
9464 |
+
{
|
9465 |
+
"epoch": 41.3,
|
9466 |
+
"learning_rate": 1.6049438202247191e-06,
|
9467 |
+
"loss": 0.3439,
|
9468 |
+
"step": 37875
|
9469 |
+
},
|
9470 |
+
{
|
9471 |
+
"epoch": 41.33,
|
9472 |
+
"learning_rate": 1.5993258426966293e-06,
|
9473 |
+
"loss": 0.2374,
|
9474 |
+
"step": 37900
|
9475 |
+
},
|
9476 |
+
{
|
9477 |
+
"epoch": 41.36,
|
9478 |
+
"learning_rate": 1.5937078651685396e-06,
|
9479 |
+
"loss": 0.3332,
|
9480 |
+
"step": 37925
|
9481 |
+
},
|
9482 |
+
{
|
9483 |
+
"epoch": 41.38,
|
9484 |
+
"learning_rate": 1.5880898876404496e-06,
|
9485 |
+
"loss": 0.2999,
|
9486 |
+
"step": 37950
|
9487 |
+
},
|
9488 |
+
{
|
9489 |
+
"epoch": 41.41,
|
9490 |
+
"learning_rate": 1.5824719101123598e-06,
|
9491 |
+
"loss": 0.3052,
|
9492 |
+
"step": 37975
|
9493 |
+
},
|
9494 |
+
{
|
9495 |
+
"epoch": 41.44,
|
9496 |
+
"learning_rate": 1.5768539325842696e-06,
|
9497 |
+
"loss": 0.2992,
|
9498 |
+
"step": 38000
|
9499 |
+
},
|
9500 |
+
{
|
9501 |
+
"epoch": 41.44,
|
9502 |
+
"eval_cer": 6.946469560180704,
|
9503 |
+
"eval_loss": 0.4622255265712738,
|
9504 |
+
"eval_runtime": 3167.1342,
|
9505 |
+
"eval_samples_per_second": 0.684,
|
9506 |
+
"eval_steps_per_second": 0.171,
|
9507 |
+
"eval_wer": 22.25945615834514,
|
9508 |
+
"step": 38000
|
9509 |
+
},
|
9510 |
+
{
|
9511 |
+
"epoch": 41.47,
|
9512 |
+
"learning_rate": 1.5712359550561799e-06,
|
9513 |
+
"loss": 0.2935,
|
9514 |
+
"step": 38025
|
9515 |
+
},
|
9516 |
+
{
|
9517 |
+
"epoch": 41.49,
|
9518 |
+
"learning_rate": 1.5656179775280899e-06,
|
9519 |
+
"loss": 0.3086,
|
9520 |
+
"step": 38050
|
9521 |
+
},
|
9522 |
+
{
|
9523 |
+
"epoch": 41.52,
|
9524 |
+
"learning_rate": 1.56e-06,
|
9525 |
+
"loss": 0.301,
|
9526 |
+
"step": 38075
|
9527 |
+
},
|
9528 |
+
{
|
9529 |
+
"epoch": 41.55,
|
9530 |
+
"learning_rate": 1.5543820224719103e-06,
|
9531 |
+
"loss": 0.2848,
|
9532 |
+
"step": 38100
|
9533 |
+
},
|
9534 |
+
{
|
9535 |
+
"epoch": 41.58,
|
9536 |
+
"learning_rate": 1.5487640449438203e-06,
|
9537 |
+
"loss": 0.2642,
|
9538 |
+
"step": 38125
|
9539 |
+
},
|
9540 |
+
{
|
9541 |
+
"epoch": 41.6,
|
9542 |
+
"learning_rate": 1.5431460674157306e-06,
|
9543 |
+
"loss": 0.3192,
|
9544 |
+
"step": 38150
|
9545 |
+
},
|
9546 |
+
{
|
9547 |
+
"epoch": 41.63,
|
9548 |
+
"learning_rate": 1.5375280898876404e-06,
|
9549 |
+
"loss": 0.2712,
|
9550 |
+
"step": 38175
|
9551 |
+
},
|
9552 |
+
{
|
9553 |
+
"epoch": 41.66,
|
9554 |
+
"learning_rate": 1.5319101123595506e-06,
|
9555 |
+
"loss": 0.2649,
|
9556 |
+
"step": 38200
|
9557 |
+
},
|
9558 |
+
{
|
9559 |
+
"epoch": 41.68,
|
9560 |
+
"learning_rate": 1.5262921348314608e-06,
|
9561 |
+
"loss": 0.2872,
|
9562 |
+
"step": 38225
|
9563 |
+
},
|
9564 |
+
{
|
9565 |
+
"epoch": 41.71,
|
9566 |
+
"learning_rate": 1.5206741573033708e-06,
|
9567 |
+
"loss": 0.3132,
|
9568 |
+
"step": 38250
|
9569 |
+
},
|
9570 |
+
{
|
9571 |
+
"epoch": 41.74,
|
9572 |
+
"learning_rate": 1.515056179775281e-06,
|
9573 |
+
"loss": 0.2831,
|
9574 |
+
"step": 38275
|
9575 |
+
},
|
9576 |
+
{
|
9577 |
+
"epoch": 41.77,
|
9578 |
+
"learning_rate": 1.5094382022471913e-06,
|
9579 |
+
"loss": 0.2998,
|
9580 |
+
"step": 38300
|
9581 |
+
},
|
9582 |
+
{
|
9583 |
+
"epoch": 41.79,
|
9584 |
+
"learning_rate": 1.5038202247191013e-06,
|
9585 |
+
"loss": 0.3153,
|
9586 |
+
"step": 38325
|
9587 |
+
},
|
9588 |
+
{
|
9589 |
+
"epoch": 41.82,
|
9590 |
+
"learning_rate": 1.4982022471910113e-06,
|
9591 |
+
"loss": 0.2693,
|
9592 |
+
"step": 38350
|
9593 |
+
},
|
9594 |
+
{
|
9595 |
+
"epoch": 41.85,
|
9596 |
+
"learning_rate": 1.4925842696629214e-06,
|
9597 |
+
"loss": 0.3361,
|
9598 |
+
"step": 38375
|
9599 |
+
},
|
9600 |
+
{
|
9601 |
+
"epoch": 41.88,
|
9602 |
+
"learning_rate": 1.4869662921348316e-06,
|
9603 |
+
"loss": 0.3173,
|
9604 |
+
"step": 38400
|
9605 |
+
},
|
9606 |
+
{
|
9607 |
+
"epoch": 41.9,
|
9608 |
+
"learning_rate": 1.4813483146067418e-06,
|
9609 |
+
"loss": 0.2723,
|
9610 |
+
"step": 38425
|
9611 |
+
},
|
9612 |
+
{
|
9613 |
+
"epoch": 41.93,
|
9614 |
+
"learning_rate": 1.4757303370786518e-06,
|
9615 |
+
"loss": 0.3004,
|
9616 |
+
"step": 38450
|
9617 |
+
},
|
9618 |
+
{
|
9619 |
+
"epoch": 41.96,
|
9620 |
+
"learning_rate": 1.470112359550562e-06,
|
9621 |
+
"loss": 0.3207,
|
9622 |
+
"step": 38475
|
9623 |
+
},
|
9624 |
+
{
|
9625 |
+
"epoch": 41.98,
|
9626 |
+
"learning_rate": 1.4644943820224719e-06,
|
9627 |
+
"loss": 0.2761,
|
9628 |
+
"step": 38500
|
9629 |
+
},
|
9630 |
+
{
|
9631 |
+
"epoch": 42.01,
|
9632 |
+
"learning_rate": 1.458876404494382e-06,
|
9633 |
+
"loss": 0.3081,
|
9634 |
+
"step": 38525
|
9635 |
+
},
|
9636 |
+
{
|
9637 |
+
"epoch": 42.04,
|
9638 |
+
"learning_rate": 1.4532584269662923e-06,
|
9639 |
+
"loss": 0.2535,
|
9640 |
+
"step": 38550
|
9641 |
+
},
|
9642 |
+
{
|
9643 |
+
"epoch": 42.07,
|
9644 |
+
"learning_rate": 1.4476404494382023e-06,
|
9645 |
+
"loss": 0.2461,
|
9646 |
+
"step": 38575
|
9647 |
+
},
|
9648 |
+
{
|
9649 |
+
"epoch": 42.09,
|
9650 |
+
"learning_rate": 1.4420224719101126e-06,
|
9651 |
+
"loss": 0.2822,
|
9652 |
+
"step": 38600
|
9653 |
+
},
|
9654 |
+
{
|
9655 |
+
"epoch": 42.12,
|
9656 |
+
"learning_rate": 1.4364044943820226e-06,
|
9657 |
+
"loss": 0.2379,
|
9658 |
+
"step": 38625
|
9659 |
+
},
|
9660 |
+
{
|
9661 |
+
"epoch": 42.15,
|
9662 |
+
"learning_rate": 1.4307865168539328e-06,
|
9663 |
+
"loss": 0.2468,
|
9664 |
+
"step": 38650
|
9665 |
+
},
|
9666 |
+
{
|
9667 |
+
"epoch": 42.18,
|
9668 |
+
"learning_rate": 1.4251685393258426e-06,
|
9669 |
+
"loss": 0.2701,
|
9670 |
+
"step": 38675
|
9671 |
+
},
|
9672 |
+
{
|
9673 |
+
"epoch": 42.2,
|
9674 |
+
"learning_rate": 1.4195505617977528e-06,
|
9675 |
+
"loss": 0.258,
|
9676 |
+
"step": 38700
|
9677 |
+
},
|
9678 |
+
{
|
9679 |
+
"epoch": 42.23,
|
9680 |
+
"learning_rate": 1.413932584269663e-06,
|
9681 |
+
"loss": 0.2449,
|
9682 |
+
"step": 38725
|
9683 |
+
},
|
9684 |
+
{
|
9685 |
+
"epoch": 42.26,
|
9686 |
+
"learning_rate": 1.408314606741573e-06,
|
9687 |
+
"loss": 0.2615,
|
9688 |
+
"step": 38750
|
9689 |
+
},
|
9690 |
+
{
|
9691 |
+
"epoch": 42.28,
|
9692 |
+
"learning_rate": 1.4026966292134833e-06,
|
9693 |
+
"loss": 0.2735,
|
9694 |
+
"step": 38775
|
9695 |
+
},
|
9696 |
+
{
|
9697 |
+
"epoch": 42.31,
|
9698 |
+
"learning_rate": 1.3970786516853935e-06,
|
9699 |
+
"loss": 0.2429,
|
9700 |
+
"step": 38800
|
9701 |
+
},
|
9702 |
+
{
|
9703 |
+
"epoch": 42.34,
|
9704 |
+
"learning_rate": 1.3914606741573033e-06,
|
9705 |
+
"loss": 0.2835,
|
9706 |
+
"step": 38825
|
9707 |
+
},
|
9708 |
+
{
|
9709 |
+
"epoch": 42.37,
|
9710 |
+
"learning_rate": 1.3858426966292136e-06,
|
9711 |
+
"loss": 0.2713,
|
9712 |
+
"step": 38850
|
9713 |
+
},
|
9714 |
+
{
|
9715 |
+
"epoch": 42.39,
|
9716 |
+
"learning_rate": 1.3802247191011236e-06,
|
9717 |
+
"loss": 0.2523,
|
9718 |
+
"step": 38875
|
9719 |
+
},
|
9720 |
+
{
|
9721 |
+
"epoch": 42.42,
|
9722 |
+
"learning_rate": 1.3746067415730338e-06,
|
9723 |
+
"loss": 0.3313,
|
9724 |
+
"step": 38900
|
9725 |
+
},
|
9726 |
+
{
|
9727 |
+
"epoch": 42.45,
|
9728 |
+
"learning_rate": 1.368988764044944e-06,
|
9729 |
+
"loss": 0.243,
|
9730 |
+
"step": 38925
|
9731 |
+
},
|
9732 |
+
{
|
9733 |
+
"epoch": 42.48,
|
9734 |
+
"learning_rate": 1.363370786516854e-06,
|
9735 |
+
"loss": 0.2526,
|
9736 |
+
"step": 38950
|
9737 |
+
},
|
9738 |
+
{
|
9739 |
+
"epoch": 42.5,
|
9740 |
+
"learning_rate": 1.3577528089887643e-06,
|
9741 |
+
"loss": 0.2498,
|
9742 |
+
"step": 38975
|
9743 |
+
},
|
9744 |
+
{
|
9745 |
+
"epoch": 42.53,
|
9746 |
+
"learning_rate": 1.352134831460674e-06,
|
9747 |
+
"loss": 0.2505,
|
9748 |
+
"step": 39000
|
9749 |
+
},
|
9750 |
+
{
|
9751 |
+
"epoch": 42.53,
|
9752 |
+
"eval_cer": 6.9198025494235775,
|
9753 |
+
"eval_loss": 0.4640973210334778,
|
9754 |
+
"eval_runtime": 3164.4611,
|
9755 |
+
"eval_samples_per_second": 0.684,
|
9756 |
+
"eval_steps_per_second": 0.171,
|
9757 |
+
"eval_wer": 22.19368780388503,
|
9758 |
+
"step": 39000
|
9759 |
+
},
|
9760 |
+
{
|
9761 |
+
"epoch": 42.56,
|
9762 |
+
"learning_rate": 1.3465168539325843e-06,
|
9763 |
+
"loss": 0.2524,
|
9764 |
+
"step": 39025
|
9765 |
+
},
|
9766 |
+
{
|
9767 |
+
"epoch": 42.58,
|
9768 |
+
"learning_rate": 1.3408988764044945e-06,
|
9769 |
+
"loss": 0.2544,
|
9770 |
+
"step": 39050
|
9771 |
+
},
|
9772 |
+
{
|
9773 |
+
"epoch": 42.61,
|
9774 |
+
"learning_rate": 1.3352808988764046e-06,
|
9775 |
+
"loss": 0.2662,
|
9776 |
+
"step": 39075
|
9777 |
+
},
|
9778 |
+
{
|
9779 |
+
"epoch": 42.64,
|
9780 |
+
"learning_rate": 1.3296629213483148e-06,
|
9781 |
+
"loss": 0.2585,
|
9782 |
+
"step": 39100
|
9783 |
+
},
|
9784 |
+
{
|
9785 |
+
"epoch": 42.67,
|
9786 |
+
"learning_rate": 1.324044943820225e-06,
|
9787 |
+
"loss": 0.2696,
|
9788 |
+
"step": 39125
|
9789 |
+
},
|
9790 |
+
{
|
9791 |
+
"epoch": 42.69,
|
9792 |
+
"learning_rate": 1.318426966292135e-06,
|
9793 |
+
"loss": 0.2485,
|
9794 |
+
"step": 39150
|
9795 |
+
},
|
9796 |
+
{
|
9797 |
+
"epoch": 42.72,
|
9798 |
+
"learning_rate": 1.312808988764045e-06,
|
9799 |
+
"loss": 0.2799,
|
9800 |
+
"step": 39175
|
9801 |
+
},
|
9802 |
+
{
|
9803 |
+
"epoch": 42.75,
|
9804 |
+
"learning_rate": 1.307191011235955e-06,
|
9805 |
+
"loss": 0.2432,
|
9806 |
+
"step": 39200
|
9807 |
+
},
|
9808 |
+
{
|
9809 |
+
"epoch": 42.78,
|
9810 |
+
"learning_rate": 1.3015730337078653e-06,
|
9811 |
+
"loss": 0.2511,
|
9812 |
+
"step": 39225
|
9813 |
+
},
|
9814 |
+
{
|
9815 |
+
"epoch": 42.8,
|
9816 |
+
"learning_rate": 1.2959550561797753e-06,
|
9817 |
+
"loss": 0.3178,
|
9818 |
+
"step": 39250
|
9819 |
+
},
|
9820 |
+
{
|
9821 |
+
"epoch": 42.83,
|
9822 |
+
"learning_rate": 1.2903370786516855e-06,
|
9823 |
+
"loss": 0.2375,
|
9824 |
+
"step": 39275
|
9825 |
+
},
|
9826 |
+
{
|
9827 |
+
"epoch": 42.86,
|
9828 |
+
"learning_rate": 1.2847191011235958e-06,
|
9829 |
+
"loss": 0.2477,
|
9830 |
+
"step": 39300
|
9831 |
+
},
|
9832 |
+
{
|
9833 |
+
"epoch": 42.88,
|
9834 |
+
"learning_rate": 1.2791011235955056e-06,
|
9835 |
+
"loss": 0.2577,
|
9836 |
+
"step": 39325
|
9837 |
+
},
|
9838 |
+
{
|
9839 |
+
"epoch": 42.91,
|
9840 |
+
"learning_rate": 1.2734831460674158e-06,
|
9841 |
+
"loss": 0.2973,
|
9842 |
+
"step": 39350
|
9843 |
+
},
|
9844 |
+
{
|
9845 |
+
"epoch": 42.94,
|
9846 |
+
"learning_rate": 1.2678651685393258e-06,
|
9847 |
+
"loss": 0.3231,
|
9848 |
+
"step": 39375
|
9849 |
+
},
|
9850 |
+
{
|
9851 |
+
"epoch": 42.97,
|
9852 |
+
"learning_rate": 1.262247191011236e-06,
|
9853 |
+
"loss": 0.2593,
|
9854 |
+
"step": 39400
|
9855 |
+
},
|
9856 |
+
{
|
9857 |
+
"epoch": 42.99,
|
9858 |
+
"learning_rate": 1.2566292134831463e-06,
|
9859 |
+
"loss": 0.2557,
|
9860 |
+
"step": 39425
|
9861 |
+
},
|
9862 |
+
{
|
9863 |
+
"epoch": 43.02,
|
9864 |
+
"learning_rate": 1.2510112359550563e-06,
|
9865 |
+
"loss": 0.248,
|
9866 |
+
"step": 39450
|
9867 |
+
},
|
9868 |
+
{
|
9869 |
+
"epoch": 43.05,
|
9870 |
+
"learning_rate": 1.2453932584269663e-06,
|
9871 |
+
"loss": 0.2346,
|
9872 |
+
"step": 39475
|
9873 |
+
},
|
9874 |
+
{
|
9875 |
+
"epoch": 43.08,
|
9876 |
+
"learning_rate": 1.2397752808988765e-06,
|
9877 |
+
"loss": 0.2541,
|
9878 |
+
"step": 39500
|
9879 |
+
},
|
9880 |
+
{
|
9881 |
+
"epoch": 43.1,
|
9882 |
+
"learning_rate": 1.2341573033707868e-06,
|
9883 |
+
"loss": 0.2447,
|
9884 |
+
"step": 39525
|
9885 |
+
},
|
9886 |
+
{
|
9887 |
+
"epoch": 43.13,
|
9888 |
+
"learning_rate": 1.2285393258426968e-06,
|
9889 |
+
"loss": 0.2534,
|
9890 |
+
"step": 39550
|
9891 |
+
},
|
9892 |
+
{
|
9893 |
+
"epoch": 43.16,
|
9894 |
+
"learning_rate": 1.2229213483146068e-06,
|
9895 |
+
"loss": 0.2688,
|
9896 |
+
"step": 39575
|
9897 |
+
},
|
9898 |
+
{
|
9899 |
+
"epoch": 43.18,
|
9900 |
+
"learning_rate": 1.217303370786517e-06,
|
9901 |
+
"loss": 0.2358,
|
9902 |
+
"step": 39600
|
9903 |
+
},
|
9904 |
+
{
|
9905 |
+
"epoch": 43.21,
|
9906 |
+
"learning_rate": 1.211685393258427e-06,
|
9907 |
+
"loss": 0.2607,
|
9908 |
+
"step": 39625
|
9909 |
+
},
|
9910 |
+
{
|
9911 |
+
"epoch": 43.24,
|
9912 |
+
"learning_rate": 1.2060674157303373e-06,
|
9913 |
+
"loss": 0.2362,
|
9914 |
+
"step": 39650
|
9915 |
+
},
|
9916 |
+
{
|
9917 |
+
"epoch": 43.27,
|
9918 |
+
"learning_rate": 1.2004494382022473e-06,
|
9919 |
+
"loss": 0.2014,
|
9920 |
+
"step": 39675
|
9921 |
+
},
|
9922 |
+
{
|
9923 |
+
"epoch": 43.29,
|
9924 |
+
"learning_rate": 1.1948314606741573e-06,
|
9925 |
+
"loss": 0.243,
|
9926 |
+
"step": 39700
|
9927 |
+
},
|
9928 |
+
{
|
9929 |
+
"epoch": 43.32,
|
9930 |
+
"learning_rate": 1.1892134831460675e-06,
|
9931 |
+
"loss": 0.2331,
|
9932 |
+
"step": 39725
|
9933 |
+
},
|
9934 |
+
{
|
9935 |
+
"epoch": 43.35,
|
9936 |
+
"learning_rate": 1.1835955056179778e-06,
|
9937 |
+
"loss": 0.2281,
|
9938 |
+
"step": 39750
|
9939 |
+
},
|
9940 |
+
{
|
9941 |
+
"epoch": 43.38,
|
9942 |
+
"learning_rate": 1.1779775280898878e-06,
|
9943 |
+
"loss": 0.2671,
|
9944 |
+
"step": 39775
|
9945 |
+
},
|
9946 |
+
{
|
9947 |
+
"epoch": 43.4,
|
9948 |
+
"learning_rate": 1.1723595505617978e-06,
|
9949 |
+
"loss": 0.2365,
|
9950 |
+
"step": 39800
|
9951 |
+
},
|
9952 |
+
{
|
9953 |
+
"epoch": 43.43,
|
9954 |
+
"learning_rate": 1.166741573033708e-06,
|
9955 |
+
"loss": 0.2356,
|
9956 |
+
"step": 39825
|
9957 |
+
},
|
9958 |
+
{
|
9959 |
+
"epoch": 43.46,
|
9960 |
+
"learning_rate": 1.161123595505618e-06,
|
9961 |
+
"loss": 0.2366,
|
9962 |
+
"step": 39850
|
9963 |
+
},
|
9964 |
+
{
|
9965 |
+
"epoch": 43.48,
|
9966 |
+
"learning_rate": 1.155505617977528e-06,
|
9967 |
+
"loss": 0.2345,
|
9968 |
+
"step": 39875
|
9969 |
+
},
|
9970 |
+
{
|
9971 |
+
"epoch": 43.51,
|
9972 |
+
"learning_rate": 1.1498876404494383e-06,
|
9973 |
+
"loss": 0.2239,
|
9974 |
+
"step": 39900
|
9975 |
+
},
|
9976 |
+
{
|
9977 |
+
"epoch": 43.54,
|
9978 |
+
"learning_rate": 1.1442696629213485e-06,
|
9979 |
+
"loss": 0.2391,
|
9980 |
+
"step": 39925
|
9981 |
+
},
|
9982 |
+
{
|
9983 |
+
"epoch": 43.57,
|
9984 |
+
"learning_rate": 1.1386516853932585e-06,
|
9985 |
+
"loss": 0.2517,
|
9986 |
+
"step": 39950
|
9987 |
+
},
|
9988 |
+
{
|
9989 |
+
"epoch": 43.59,
|
9990 |
+
"learning_rate": 1.1330337078651685e-06,
|
9991 |
+
"loss": 0.2091,
|
9992 |
+
"step": 39975
|
9993 |
+
},
|
9994 |
+
{
|
9995 |
+
"epoch": 43.62,
|
9996 |
+
"learning_rate": 1.1274157303370788e-06,
|
9997 |
+
"loss": 0.2477,
|
9998 |
+
"step": 40000
|
9999 |
+
},
|
10000 |
+
{
|
10001 |
+
"epoch": 43.62,
|
10002 |
+
"eval_cer": 7.2008097595524525,
|
10003 |
+
"eval_loss": 0.46780213713645935,
|
10004 |
+
"eval_runtime": 3174.0364,
|
10005 |
+
"eval_samples_per_second": 0.682,
|
10006 |
+
"eval_steps_per_second": 0.171,
|
10007 |
+
"eval_wer": 22.827882650464684,
|
10008 |
+
"step": 40000
|
10009 |
+
},
|
10010 |
+
{
|
10011 |
+
"epoch": 43.65,
|
10012 |
+
"learning_rate": 1.121797752808989e-06,
|
10013 |
+
"loss": 0.2545,
|
10014 |
+
"step": 40025
|
10015 |
+
},
|
10016 |
+
{
|
10017 |
+
"epoch": 43.68,
|
10018 |
+
"learning_rate": 1.116179775280899e-06,
|
10019 |
+
"loss": 0.2254,
|
10020 |
+
"step": 40050
|
10021 |
+
},
|
10022 |
+
{
|
10023 |
+
"epoch": 43.7,
|
10024 |
+
"learning_rate": 1.110561797752809e-06,
|
10025 |
+
"loss": 0.2051,
|
10026 |
+
"step": 40075
|
10027 |
+
},
|
10028 |
+
{
|
10029 |
+
"epoch": 43.73,
|
10030 |
+
"learning_rate": 1.1049438202247193e-06,
|
10031 |
+
"loss": 0.2168,
|
10032 |
+
"step": 40100
|
10033 |
+
},
|
10034 |
+
{
|
10035 |
+
"epoch": 43.76,
|
10036 |
+
"learning_rate": 1.0993258426966293e-06,
|
10037 |
+
"loss": 0.2304,
|
10038 |
+
"step": 40125
|
10039 |
+
},
|
10040 |
+
{
|
10041 |
+
"epoch": 43.78,
|
10042 |
+
"learning_rate": 1.0937078651685395e-06,
|
10043 |
+
"loss": 0.2346,
|
10044 |
+
"step": 40150
|
10045 |
+
},
|
10046 |
+
{
|
10047 |
+
"epoch": 43.81,
|
10048 |
+
"learning_rate": 1.0880898876404495e-06,
|
10049 |
+
"loss": 0.2231,
|
10050 |
+
"step": 40175
|
10051 |
+
},
|
10052 |
+
{
|
10053 |
+
"epoch": 43.84,
|
10054 |
+
"learning_rate": 1.0824719101123595e-06,
|
10055 |
+
"loss": 0.2233,
|
10056 |
+
"step": 40200
|
10057 |
+
},
|
10058 |
+
{
|
10059 |
+
"epoch": 43.87,
|
10060 |
+
"learning_rate": 1.0768539325842698e-06,
|
10061 |
+
"loss": 0.2394,
|
10062 |
+
"step": 40225
|
10063 |
+
},
|
10064 |
+
{
|
10065 |
+
"epoch": 43.89,
|
10066 |
+
"learning_rate": 1.07123595505618e-06,
|
10067 |
+
"loss": 0.2243,
|
10068 |
+
"step": 40250
|
10069 |
+
},
|
10070 |
+
{
|
10071 |
+
"epoch": 43.92,
|
10072 |
+
"learning_rate": 1.06561797752809e-06,
|
10073 |
+
"loss": 0.2352,
|
10074 |
+
"step": 40275
|
10075 |
+
},
|
10076 |
+
{
|
10077 |
+
"epoch": 43.95,
|
10078 |
+
"learning_rate": 1.06e-06,
|
10079 |
+
"loss": 0.2776,
|
10080 |
+
"step": 40300
|
10081 |
+
},
|
10082 |
+
{
|
10083 |
+
"epoch": 43.97,
|
10084 |
+
"learning_rate": 1.0543820224719102e-06,
|
10085 |
+
"loss": 0.2271,
|
10086 |
+
"step": 40325
|
10087 |
+
},
|
10088 |
+
{
|
10089 |
+
"epoch": 44.0,
|
10090 |
+
"learning_rate": 1.0487640449438203e-06,
|
10091 |
+
"loss": 0.2344,
|
10092 |
+
"step": 40350
|
10093 |
+
},
|
10094 |
+
{
|
10095 |
+
"epoch": 44.03,
|
10096 |
+
"learning_rate": 1.0431460674157303e-06,
|
10097 |
+
"loss": 0.1906,
|
10098 |
+
"step": 40375
|
10099 |
+
},
|
10100 |
+
{
|
10101 |
+
"epoch": 44.06,
|
10102 |
+
"learning_rate": 1.0375280898876405e-06,
|
10103 |
+
"loss": 0.2142,
|
10104 |
+
"step": 40400
|
10105 |
+
},
|
10106 |
+
{
|
10107 |
+
"epoch": 44.08,
|
10108 |
+
"learning_rate": 1.0319101123595507e-06,
|
10109 |
+
"loss": 0.1885,
|
10110 |
+
"step": 40425
|
10111 |
+
},
|
10112 |
+
{
|
10113 |
+
"epoch": 44.11,
|
10114 |
+
"learning_rate": 1.0262921348314607e-06,
|
10115 |
+
"loss": 0.2192,
|
10116 |
+
"step": 40450
|
10117 |
+
},
|
10118 |
+
{
|
10119 |
+
"epoch": 44.14,
|
10120 |
+
"learning_rate": 1.0206741573033708e-06,
|
10121 |
+
"loss": 0.2242,
|
10122 |
+
"step": 40475
|
10123 |
+
},
|
10124 |
+
{
|
10125 |
+
"epoch": 44.17,
|
10126 |
+
"learning_rate": 1.015056179775281e-06,
|
10127 |
+
"loss": 0.2627,
|
10128 |
+
"step": 40500
|
10129 |
+
},
|
10130 |
+
{
|
10131 |
+
"epoch": 44.19,
|
10132 |
+
"learning_rate": 1.0094382022471912e-06,
|
10133 |
+
"loss": 0.1955,
|
10134 |
+
"step": 40525
|
10135 |
+
},
|
10136 |
+
{
|
10137 |
+
"epoch": 44.22,
|
10138 |
+
"learning_rate": 1.0038202247191012e-06,
|
10139 |
+
"loss": 0.2384,
|
10140 |
+
"step": 40550
|
10141 |
+
},
|
10142 |
+
{
|
10143 |
+
"epoch": 44.25,
|
10144 |
+
"learning_rate": 9.982022471910113e-07,
|
10145 |
+
"loss": 0.2457,
|
10146 |
+
"step": 40575
|
10147 |
+
},
|
10148 |
+
{
|
10149 |
+
"epoch": 44.27,
|
10150 |
+
"learning_rate": 9.925842696629215e-07,
|
10151 |
+
"loss": 0.1971,
|
10152 |
+
"step": 40600
|
10153 |
+
},
|
10154 |
+
{
|
10155 |
+
"epoch": 44.3,
|
10156 |
+
"learning_rate": 9.869662921348315e-07,
|
10157 |
+
"loss": 0.2013,
|
10158 |
+
"step": 40625
|
10159 |
+
},
|
10160 |
+
{
|
10161 |
+
"epoch": 44.33,
|
10162 |
+
"learning_rate": 9.813483146067417e-07,
|
10163 |
+
"loss": 0.203,
|
10164 |
+
"step": 40650
|
10165 |
+
},
|
10166 |
+
{
|
10167 |
+
"epoch": 44.36,
|
10168 |
+
"learning_rate": 9.757303370786517e-07,
|
10169 |
+
"loss": 0.2259,
|
10170 |
+
"step": 40675
|
10171 |
+
},
|
10172 |
+
{
|
10173 |
+
"epoch": 44.38,
|
10174 |
+
"learning_rate": 9.701123595505618e-07,
|
10175 |
+
"loss": 0.201,
|
10176 |
+
"step": 40700
|
10177 |
+
},
|
10178 |
+
{
|
10179 |
+
"epoch": 44.41,
|
10180 |
+
"learning_rate": 9.64494382022472e-07,
|
10181 |
+
"loss": 0.2531,
|
10182 |
+
"step": 40725
|
10183 |
+
},
|
10184 |
+
{
|
10185 |
+
"epoch": 44.44,
|
10186 |
+
"learning_rate": 9.588764044943822e-07,
|
10187 |
+
"loss": 0.1835,
|
10188 |
+
"step": 40750
|
10189 |
+
},
|
10190 |
+
{
|
10191 |
+
"epoch": 44.47,
|
10192 |
+
"learning_rate": 9.532584269662921e-07,
|
10193 |
+
"loss": 0.2384,
|
10194 |
+
"step": 40775
|
10195 |
+
},
|
10196 |
+
{
|
10197 |
+
"epoch": 44.49,
|
10198 |
+
"learning_rate": 9.476404494382022e-07,
|
10199 |
+
"loss": 0.2023,
|
10200 |
+
"step": 40800
|
10201 |
+
},
|
10202 |
+
{
|
10203 |
+
"epoch": 44.52,
|
10204 |
+
"learning_rate": 9.420224719101125e-07,
|
10205 |
+
"loss": 0.2448,
|
10206 |
+
"step": 40825
|
10207 |
+
},
|
10208 |
+
{
|
10209 |
+
"epoch": 44.55,
|
10210 |
+
"learning_rate": 9.364044943820226e-07,
|
10211 |
+
"loss": 0.2104,
|
10212 |
+
"step": 40850
|
10213 |
+
},
|
10214 |
+
{
|
10215 |
+
"epoch": 44.57,
|
10216 |
+
"learning_rate": 9.307865168539326e-07,
|
10217 |
+
"loss": 0.2211,
|
10218 |
+
"step": 40875
|
10219 |
+
},
|
10220 |
+
{
|
10221 |
+
"epoch": 44.6,
|
10222 |
+
"learning_rate": 9.251685393258427e-07,
|
10223 |
+
"loss": 0.2498,
|
10224 |
+
"step": 40900
|
10225 |
+
},
|
10226 |
+
{
|
10227 |
+
"epoch": 44.63,
|
10228 |
+
"learning_rate": 9.19550561797753e-07,
|
10229 |
+
"loss": 0.2139,
|
10230 |
+
"step": 40925
|
10231 |
+
},
|
10232 |
+
{
|
10233 |
+
"epoch": 44.66,
|
10234 |
+
"learning_rate": 9.13932584269663e-07,
|
10235 |
+
"loss": 0.2349,
|
10236 |
+
"step": 40950
|
10237 |
+
},
|
10238 |
+
{
|
10239 |
+
"epoch": 44.68,
|
10240 |
+
"learning_rate": 9.083146067415731e-07,
|
10241 |
+
"loss": 0.2201,
|
10242 |
+
"step": 40975
|
10243 |
+
},
|
10244 |
+
{
|
10245 |
+
"epoch": 44.71,
|
10246 |
+
"learning_rate": 9.026966292134832e-07,
|
10247 |
+
"loss": 0.1994,
|
10248 |
+
"step": 41000
|
10249 |
+
},
|
10250 |
+
{
|
10251 |
+
"epoch": 44.71,
|
10252 |
+
"eval_cer": 7.11794130676955,
|
10253 |
+
"eval_loss": 0.46892717480659485,
|
10254 |
+
"eval_runtime": 3141.0329,
|
10255 |
+
"eval_samples_per_second": 0.69,
|
10256 |
+
"eval_steps_per_second": 0.173,
|
10257 |
+
"eval_wer": 22.380814431456063,
|
10258 |
+
"step": 41000
|
10259 |
+
},
|
10260 |
+
{
|
10261 |
+
"epoch": 44.74,
|
10262 |
+
"learning_rate": 8.970786516853932e-07,
|
10263 |
+
"loss": 0.1841,
|
10264 |
+
"step": 41025
|
10265 |
+
},
|
10266 |
+
{
|
10267 |
+
"epoch": 44.77,
|
10268 |
+
"learning_rate": 8.914606741573035e-07,
|
10269 |
+
"loss": 0.204,
|
10270 |
+
"step": 41050
|
10271 |
+
},
|
10272 |
+
{
|
10273 |
+
"epoch": 44.79,
|
10274 |
+
"learning_rate": 8.858426966292136e-07,
|
10275 |
+
"loss": 0.2144,
|
10276 |
+
"step": 41075
|
10277 |
+
},
|
10278 |
+
{
|
10279 |
+
"epoch": 44.82,
|
10280 |
+
"learning_rate": 8.802247191011237e-07,
|
10281 |
+
"loss": 0.2586,
|
10282 |
+
"step": 41100
|
10283 |
+
},
|
10284 |
+
{
|
10285 |
+
"epoch": 44.85,
|
10286 |
+
"learning_rate": 8.746067415730337e-07,
|
10287 |
+
"loss": 0.219,
|
10288 |
+
"step": 41125
|
10289 |
+
},
|
10290 |
+
{
|
10291 |
+
"epoch": 44.87,
|
10292 |
+
"learning_rate": 8.689887640449439e-07,
|
10293 |
+
"loss": 0.2091,
|
10294 |
+
"step": 41150
|
10295 |
+
},
|
10296 |
+
{
|
10297 |
+
"epoch": 44.9,
|
10298 |
+
"learning_rate": 8.633707865168541e-07,
|
10299 |
+
"loss": 0.1841,
|
10300 |
+
"step": 41175
|
10301 |
+
},
|
10302 |
+
{
|
10303 |
+
"epoch": 44.93,
|
10304 |
+
"learning_rate": 8.577528089887641e-07,
|
10305 |
+
"loss": 0.202,
|
10306 |
+
"step": 41200
|
10307 |
+
},
|
10308 |
+
{
|
10309 |
+
"epoch": 44.96,
|
10310 |
+
"learning_rate": 8.521348314606742e-07,
|
10311 |
+
"loss": 0.2057,
|
10312 |
+
"step": 41225
|
10313 |
+
},
|
10314 |
+
{
|
10315 |
+
"epoch": 44.98,
|
10316 |
+
"learning_rate": 8.465168539325843e-07,
|
10317 |
+
"loss": 0.2353,
|
10318 |
+
"step": 41250
|
10319 |
+
},
|
10320 |
+
{
|
10321 |
+
"epoch": 45.01,
|
10322 |
+
"learning_rate": 8.408988764044944e-07,
|
10323 |
+
"loss": 0.181,
|
10324 |
+
"step": 41275
|
10325 |
+
},
|
10326 |
+
{
|
10327 |
+
"epoch": 45.04,
|
10328 |
+
"learning_rate": 8.352808988764046e-07,
|
10329 |
+
"loss": 0.2289,
|
10330 |
+
"step": 41300
|
10331 |
+
},
|
10332 |
+
{
|
10333 |
+
"epoch": 45.07,
|
10334 |
+
"learning_rate": 8.296629213483147e-07,
|
10335 |
+
"loss": 0.2127,
|
10336 |
+
"step": 41325
|
10337 |
+
},
|
10338 |
+
{
|
10339 |
+
"epoch": 45.09,
|
10340 |
+
"learning_rate": 8.240449438202248e-07,
|
10341 |
+
"loss": 0.2119,
|
10342 |
+
"step": 41350
|
10343 |
+
},
|
10344 |
+
{
|
10345 |
+
"epoch": 45.12,
|
10346 |
+
"learning_rate": 8.184269662921348e-07,
|
10347 |
+
"loss": 0.2223,
|
10348 |
+
"step": 41375
|
10349 |
+
},
|
10350 |
+
{
|
10351 |
+
"epoch": 45.15,
|
10352 |
+
"learning_rate": 8.12808988764045e-07,
|
10353 |
+
"loss": 0.2207,
|
10354 |
+
"step": 41400
|
10355 |
+
},
|
10356 |
+
{
|
10357 |
+
"epoch": 45.17,
|
10358 |
+
"learning_rate": 8.071910112359552e-07,
|
10359 |
+
"loss": 0.213,
|
10360 |
+
"step": 41425
|
10361 |
+
},
|
10362 |
+
{
|
10363 |
+
"epoch": 45.2,
|
10364 |
+
"learning_rate": 8.015730337078652e-07,
|
10365 |
+
"loss": 0.2092,
|
10366 |
+
"step": 41450
|
10367 |
+
},
|
10368 |
+
{
|
10369 |
+
"epoch": 45.23,
|
10370 |
+
"learning_rate": 7.959550561797753e-07,
|
10371 |
+
"loss": 0.1926,
|
10372 |
+
"step": 41475
|
10373 |
+
},
|
10374 |
+
{
|
10375 |
+
"epoch": 45.26,
|
10376 |
+
"learning_rate": 7.903370786516855e-07,
|
10377 |
+
"loss": 0.2111,
|
10378 |
+
"step": 41500
|
10379 |
+
},
|
10380 |
+
{
|
10381 |
+
"epoch": 45.28,
|
10382 |
+
"learning_rate": 7.847191011235955e-07,
|
10383 |
+
"loss": 0.1974,
|
10384 |
+
"step": 41525
|
10385 |
+
},
|
10386 |
+
{
|
10387 |
+
"epoch": 45.31,
|
10388 |
+
"learning_rate": 7.791011235955057e-07,
|
10389 |
+
"loss": 0.1944,
|
10390 |
+
"step": 41550
|
10391 |
+
},
|
10392 |
+
{
|
10393 |
+
"epoch": 45.34,
|
10394 |
+
"learning_rate": 7.734831460674158e-07,
|
10395 |
+
"loss": 0.159,
|
10396 |
+
"step": 41575
|
10397 |
+
},
|
10398 |
+
{
|
10399 |
+
"epoch": 45.37,
|
10400 |
+
"learning_rate": 7.678651685393259e-07,
|
10401 |
+
"loss": 0.2071,
|
10402 |
+
"step": 41600
|
10403 |
+
},
|
10404 |
+
{
|
10405 |
+
"epoch": 45.39,
|
10406 |
+
"learning_rate": 7.62247191011236e-07,
|
10407 |
+
"loss": 0.2017,
|
10408 |
+
"step": 41625
|
10409 |
+
},
|
10410 |
+
{
|
10411 |
+
"epoch": 45.42,
|
10412 |
+
"learning_rate": 7.566292134831462e-07,
|
10413 |
+
"loss": 0.2139,
|
10414 |
+
"step": 41650
|
10415 |
+
},
|
10416 |
+
{
|
10417 |
+
"epoch": 45.45,
|
10418 |
+
"learning_rate": 7.510112359550563e-07,
|
10419 |
+
"loss": 0.1659,
|
10420 |
+
"step": 41675
|
10421 |
+
},
|
10422 |
+
{
|
10423 |
+
"epoch": 45.47,
|
10424 |
+
"learning_rate": 7.453932584269663e-07,
|
10425 |
+
"loss": 0.2089,
|
10426 |
+
"step": 41700
|
10427 |
+
},
|
10428 |
+
{
|
10429 |
+
"epoch": 45.5,
|
10430 |
+
"learning_rate": 7.397752808988764e-07,
|
10431 |
+
"loss": 0.1634,
|
10432 |
+
"step": 41725
|
10433 |
+
},
|
10434 |
+
{
|
10435 |
+
"epoch": 45.53,
|
10436 |
+
"learning_rate": 7.341573033707866e-07,
|
10437 |
+
"loss": 0.2303,
|
10438 |
+
"step": 41750
|
10439 |
+
},
|
10440 |
+
{
|
10441 |
+
"epoch": 45.56,
|
10442 |
+
"learning_rate": 7.285393258426966e-07,
|
10443 |
+
"loss": 0.2094,
|
10444 |
+
"step": 41775
|
10445 |
+
},
|
10446 |
+
{
|
10447 |
+
"epoch": 45.58,
|
10448 |
+
"learning_rate": 7.229213483146068e-07,
|
10449 |
+
"loss": 0.1817,
|
10450 |
+
"step": 41800
|
10451 |
+
},
|
10452 |
+
{
|
10453 |
+
"epoch": 45.61,
|
10454 |
+
"learning_rate": 7.173033707865169e-07,
|
10455 |
+
"loss": 0.1848,
|
10456 |
+
"step": 41825
|
10457 |
+
},
|
10458 |
+
{
|
10459 |
+
"epoch": 45.64,
|
10460 |
+
"learning_rate": 7.116853932584271e-07,
|
10461 |
+
"loss": 0.1971,
|
10462 |
+
"step": 41850
|
10463 |
+
},
|
10464 |
+
{
|
10465 |
+
"epoch": 45.67,
|
10466 |
+
"learning_rate": 7.060674157303371e-07,
|
10467 |
+
"loss": 0.2081,
|
10468 |
+
"step": 41875
|
10469 |
+
},
|
10470 |
+
{
|
10471 |
+
"epoch": 45.69,
|
10472 |
+
"learning_rate": 7.004494382022473e-07,
|
10473 |
+
"loss": 0.2245,
|
10474 |
+
"step": 41900
|
10475 |
+
},
|
10476 |
+
{
|
10477 |
+
"epoch": 45.72,
|
10478 |
+
"learning_rate": 6.948314606741574e-07,
|
10479 |
+
"loss": 0.1896,
|
10480 |
+
"step": 41925
|
10481 |
+
},
|
10482 |
+
{
|
10483 |
+
"epoch": 45.75,
|
10484 |
+
"learning_rate": 6.892134831460674e-07,
|
10485 |
+
"loss": 0.2193,
|
10486 |
+
"step": 41950
|
10487 |
+
},
|
10488 |
+
{
|
10489 |
+
"epoch": 45.77,
|
10490 |
+
"learning_rate": 6.835955056179776e-07,
|
10491 |
+
"loss": 0.1998,
|
10492 |
+
"step": 41975
|
10493 |
+
},
|
10494 |
+
{
|
10495 |
+
"epoch": 45.8,
|
10496 |
+
"learning_rate": 6.779775280898877e-07,
|
10497 |
+
"loss": 0.1865,
|
10498 |
+
"step": 42000
|
10499 |
+
},
|
10500 |
+
{
|
10501 |
+
"epoch": 45.8,
|
10502 |
+
"eval_cer": 7.135145829838664,
|
10503 |
+
"eval_loss": 0.47173011302948,
|
10504 |
+
"eval_runtime": 3187.2,
|
10505 |
+
"eval_samples_per_second": 0.68,
|
10506 |
+
"eval_steps_per_second": 0.17,
|
10507 |
+
"eval_wer": 22.566375145825667,
|
10508 |
+
"step": 42000
|
10509 |
+
},
|
10510 |
+
{
|
10511 |
+
"epoch": 45.83,
|
10512 |
+
"learning_rate": 6.723595505617977e-07,
|
10513 |
+
"loss": 0.2185,
|
10514 |
+
"step": 42025
|
10515 |
+
},
|
10516 |
+
{
|
10517 |
+
"epoch": 45.86,
|
10518 |
+
"learning_rate": 6.667415730337079e-07,
|
10519 |
+
"loss": 0.2142,
|
10520 |
+
"step": 42050
|
10521 |
+
},
|
10522 |
+
{
|
10523 |
+
"epoch": 45.88,
|
10524 |
+
"learning_rate": 6.61123595505618e-07,
|
10525 |
+
"loss": 0.203,
|
10526 |
+
"step": 42075
|
10527 |
+
},
|
10528 |
+
{
|
10529 |
+
"epoch": 45.91,
|
10530 |
+
"learning_rate": 6.555056179775282e-07,
|
10531 |
+
"loss": 0.1912,
|
10532 |
+
"step": 42100
|
10533 |
+
},
|
10534 |
+
{
|
10535 |
+
"epoch": 45.94,
|
10536 |
+
"learning_rate": 6.498876404494382e-07,
|
10537 |
+
"loss": 0.1853,
|
10538 |
+
"step": 42125
|
10539 |
+
},
|
10540 |
+
{
|
10541 |
+
"epoch": 45.97,
|
10542 |
+
"learning_rate": 6.442696629213484e-07,
|
10543 |
+
"loss": 0.2056,
|
10544 |
+
"step": 42150
|
10545 |
+
},
|
10546 |
+
{
|
10547 |
+
"epoch": 45.99,
|
10548 |
+
"learning_rate": 6.386516853932585e-07,
|
10549 |
+
"loss": 0.1564,
|
10550 |
+
"step": 42175
|
10551 |
+
},
|
10552 |
+
{
|
10553 |
+
"epoch": 46.02,
|
10554 |
+
"learning_rate": 6.330337078651686e-07,
|
10555 |
+
"loss": 0.1701,
|
10556 |
+
"step": 42200
|
10557 |
+
},
|
10558 |
+
{
|
10559 |
+
"epoch": 46.05,
|
10560 |
+
"learning_rate": 6.274157303370787e-07,
|
10561 |
+
"loss": 0.2156,
|
10562 |
+
"step": 42225
|
10563 |
},
|
10564 |
{
|
10565 |
+
"epoch": 46.07,
|
10566 |
+
"learning_rate": 6.217977528089888e-07,
|
10567 |
+
"loss": 0.1844,
|
10568 |
+
"step": 42250
|
10569 |
+
},
|
10570 |
+
{
|
10571 |
+
"epoch": 46.1,
|
10572 |
+
"learning_rate": 6.161797752808989e-07,
|
10573 |
+
"loss": 0.1641,
|
10574 |
+
"step": 42275
|
10575 |
+
},
|
10576 |
+
{
|
10577 |
+
"epoch": 46.13,
|
10578 |
+
"learning_rate": 6.10561797752809e-07,
|
10579 |
+
"loss": 0.1754,
|
10580 |
+
"step": 42300
|
10581 |
+
},
|
10582 |
+
{
|
10583 |
+
"epoch": 46.16,
|
10584 |
+
"learning_rate": 6.049438202247192e-07,
|
10585 |
+
"loss": 0.1857,
|
10586 |
+
"step": 42325
|
10587 |
+
},
|
10588 |
+
{
|
10589 |
+
"epoch": 46.18,
|
10590 |
+
"learning_rate": 5.993258426966293e-07,
|
10591 |
+
"loss": 0.159,
|
10592 |
+
"step": 42350
|
10593 |
+
},
|
10594 |
+
{
|
10595 |
+
"epoch": 46.21,
|
10596 |
+
"learning_rate": 5.937078651685393e-07,
|
10597 |
+
"loss": 0.2097,
|
10598 |
+
"step": 42375
|
10599 |
+
},
|
10600 |
+
{
|
10601 |
+
"epoch": 46.24,
|
10602 |
+
"learning_rate": 5.880898876404495e-07,
|
10603 |
+
"loss": 0.1941,
|
10604 |
+
"step": 42400
|
10605 |
+
},
|
10606 |
+
{
|
10607 |
+
"epoch": 46.26,
|
10608 |
+
"learning_rate": 5.824719101123595e-07,
|
10609 |
+
"loss": 0.1817,
|
10610 |
+
"step": 42425
|
10611 |
+
},
|
10612 |
+
{
|
10613 |
+
"epoch": 46.29,
|
10614 |
+
"learning_rate": 5.768539325842698e-07,
|
10615 |
+
"loss": 0.2022,
|
10616 |
+
"step": 42450
|
10617 |
+
},
|
10618 |
+
{
|
10619 |
+
"epoch": 46.32,
|
10620 |
+
"learning_rate": 5.712359550561798e-07,
|
10621 |
+
"loss": 0.1958,
|
10622 |
+
"step": 42475
|
10623 |
+
},
|
10624 |
+
{
|
10625 |
+
"epoch": 46.35,
|
10626 |
+
"learning_rate": 5.656179775280899e-07,
|
10627 |
+
"loss": 0.176,
|
10628 |
+
"step": 42500
|
10629 |
+
},
|
10630 |
+
{
|
10631 |
+
"epoch": 46.37,
|
10632 |
+
"learning_rate": 5.6e-07,
|
10633 |
+
"loss": 0.2181,
|
10634 |
+
"step": 42525
|
10635 |
+
},
|
10636 |
+
{
|
10637 |
+
"epoch": 46.4,
|
10638 |
+
"learning_rate": 5.543820224719102e-07,
|
10639 |
+
"loss": 0.2047,
|
10640 |
+
"step": 42550
|
10641 |
+
},
|
10642 |
+
{
|
10643 |
+
"epoch": 46.43,
|
10644 |
+
"learning_rate": 5.487640449438203e-07,
|
10645 |
+
"loss": 0.1942,
|
10646 |
+
"step": 42575
|
10647 |
+
},
|
10648 |
+
{
|
10649 |
+
"epoch": 46.46,
|
10650 |
+
"learning_rate": 5.431460674157304e-07,
|
10651 |
+
"loss": 0.1655,
|
10652 |
+
"step": 42600
|
10653 |
+
},
|
10654 |
+
{
|
10655 |
+
"epoch": 46.48,
|
10656 |
+
"learning_rate": 5.375280898876404e-07,
|
10657 |
+
"loss": 0.1958,
|
10658 |
+
"step": 42625
|
10659 |
+
},
|
10660 |
+
{
|
10661 |
+
"epoch": 46.51,
|
10662 |
+
"learning_rate": 5.319101123595506e-07,
|
10663 |
+
"loss": 0.2192,
|
10664 |
+
"step": 42650
|
10665 |
+
},
|
10666 |
+
{
|
10667 |
+
"epoch": 46.54,
|
10668 |
+
"learning_rate": 5.262921348314607e-07,
|
10669 |
+
"loss": 0.1425,
|
10670 |
+
"step": 42675
|
10671 |
+
},
|
10672 |
+
{
|
10673 |
+
"epoch": 46.56,
|
10674 |
+
"learning_rate": 5.206741573033709e-07,
|
10675 |
+
"loss": 0.1848,
|
10676 |
+
"step": 42700
|
10677 |
+
},
|
10678 |
+
{
|
10679 |
+
"epoch": 46.59,
|
10680 |
+
"learning_rate": 5.150561797752809e-07,
|
10681 |
+
"loss": 0.1621,
|
10682 |
+
"step": 42725
|
10683 |
+
},
|
10684 |
+
{
|
10685 |
+
"epoch": 46.62,
|
10686 |
+
"learning_rate": 5.09438202247191e-07,
|
10687 |
+
"loss": 0.177,
|
10688 |
+
"step": 42750
|
10689 |
+
},
|
10690 |
+
{
|
10691 |
+
"epoch": 46.65,
|
10692 |
+
"learning_rate": 5.038202247191012e-07,
|
10693 |
+
"loss": 0.1937,
|
10694 |
+
"step": 42775
|
10695 |
+
},
|
10696 |
+
{
|
10697 |
+
"epoch": 46.67,
|
10698 |
+
"learning_rate": 4.982022471910113e-07,
|
10699 |
+
"loss": 0.175,
|
10700 |
+
"step": 42800
|
10701 |
+
},
|
10702 |
+
{
|
10703 |
+
"epoch": 46.7,
|
10704 |
+
"learning_rate": 4.925842696629214e-07,
|
10705 |
+
"loss": 0.1738,
|
10706 |
+
"step": 42825
|
10707 |
+
},
|
10708 |
+
{
|
10709 |
+
"epoch": 46.73,
|
10710 |
+
"learning_rate": 4.869662921348315e-07,
|
10711 |
+
"loss": 0.1732,
|
10712 |
+
"step": 42850
|
10713 |
+
},
|
10714 |
+
{
|
10715 |
+
"epoch": 46.76,
|
10716 |
+
"learning_rate": 4.813483146067415e-07,
|
10717 |
+
"loss": 0.1871,
|
10718 |
+
"step": 42875
|
10719 |
+
},
|
10720 |
+
{
|
10721 |
+
"epoch": 46.78,
|
10722 |
+
"learning_rate": 4.7573033707865176e-07,
|
10723 |
+
"loss": 0.1949,
|
10724 |
+
"step": 42900
|
10725 |
+
},
|
10726 |
+
{
|
10727 |
+
"epoch": 46.81,
|
10728 |
+
"learning_rate": 4.7011235955056183e-07,
|
10729 |
+
"loss": 0.1999,
|
10730 |
+
"step": 42925
|
10731 |
+
},
|
10732 |
+
{
|
10733 |
+
"epoch": 46.84,
|
10734 |
+
"learning_rate": 4.6449438202247196e-07,
|
10735 |
+
"loss": 0.1805,
|
10736 |
+
"step": 42950
|
10737 |
+
},
|
10738 |
+
{
|
10739 |
+
"epoch": 46.86,
|
10740 |
+
"learning_rate": 4.588764044943821e-07,
|
10741 |
+
"loss": 0.1668,
|
10742 |
+
"step": 42975
|
10743 |
+
},
|
10744 |
+
{
|
10745 |
+
"epoch": 46.89,
|
10746 |
+
"learning_rate": 4.5325842696629215e-07,
|
10747 |
+
"loss": 0.2307,
|
10748 |
+
"step": 43000
|
10749 |
+
},
|
10750 |
+
{
|
10751 |
+
"epoch": 46.89,
|
10752 |
+
"eval_cer": 7.136436169068848,
|
10753 |
+
"eval_loss": 0.47538912296295166,
|
10754 |
+
"eval_runtime": 3233.0256,
|
10755 |
+
"eval_samples_per_second": 0.67,
|
10756 |
+
"eval_steps_per_second": 0.168,
|
10757 |
+
"eval_wer": 22.372201908848194,
|
10758 |
+
"step": 43000
|
10759 |
+
},
|
10760 |
+
{
|
10761 |
+
"epoch": 46.92,
|
10762 |
+
"learning_rate": 4.476404494382023e-07,
|
10763 |
+
"loss": 0.1901,
|
10764 |
+
"step": 43025
|
10765 |
+
},
|
10766 |
+
{
|
10767 |
+
"epoch": 46.95,
|
10768 |
+
"learning_rate": 4.420224719101124e-07,
|
10769 |
+
"loss": 0.197,
|
10770 |
+
"step": 43050
|
10771 |
+
},
|
10772 |
+
{
|
10773 |
+
"epoch": 46.97,
|
10774 |
+
"learning_rate": 4.364044943820225e-07,
|
10775 |
+
"loss": 0.2022,
|
10776 |
+
"step": 43075
|
10777 |
+
},
|
10778 |
+
{
|
10779 |
+
"epoch": 47.0,
|
10780 |
+
"learning_rate": 4.3078651685393264e-07,
|
10781 |
+
"loss": 0.1763,
|
10782 |
+
"step": 43100
|
10783 |
+
},
|
10784 |
+
{
|
10785 |
+
"epoch": 47.03,
|
10786 |
+
"learning_rate": 4.251685393258427e-07,
|
10787 |
+
"loss": 0.2063,
|
10788 |
+
"step": 43125
|
10789 |
+
},
|
10790 |
+
{
|
10791 |
+
"epoch": 47.06,
|
10792 |
+
"learning_rate": 4.195505617977529e-07,
|
10793 |
+
"loss": 0.1824,
|
10794 |
+
"step": 43150
|
10795 |
+
},
|
10796 |
+
{
|
10797 |
+
"epoch": 47.08,
|
10798 |
+
"learning_rate": 4.1393258426966295e-07,
|
10799 |
+
"loss": 0.2059,
|
10800 |
+
"step": 43175
|
10801 |
+
},
|
10802 |
+
{
|
10803 |
+
"epoch": 47.11,
|
10804 |
+
"learning_rate": 4.0831460674157307e-07,
|
10805 |
+
"loss": 0.194,
|
10806 |
+
"step": 43200
|
10807 |
+
},
|
10808 |
+
{
|
10809 |
+
"epoch": 47.14,
|
10810 |
+
"learning_rate": 4.026966292134832e-07,
|
10811 |
+
"loss": 0.155,
|
10812 |
+
"step": 43225
|
10813 |
+
},
|
10814 |
+
{
|
10815 |
+
"epoch": 47.16,
|
10816 |
+
"learning_rate": 3.9707865168539326e-07,
|
10817 |
+
"loss": 0.1933,
|
10818 |
+
"step": 43250
|
10819 |
+
},
|
10820 |
+
{
|
10821 |
+
"epoch": 47.19,
|
10822 |
+
"learning_rate": 3.9146067415730344e-07,
|
10823 |
+
"loss": 0.1939,
|
10824 |
+
"step": 43275
|
10825 |
+
},
|
10826 |
+
{
|
10827 |
+
"epoch": 47.22,
|
10828 |
+
"learning_rate": 3.858426966292135e-07,
|
10829 |
+
"loss": 0.1826,
|
10830 |
+
"step": 43300
|
10831 |
+
},
|
10832 |
+
{
|
10833 |
+
"epoch": 47.25,
|
10834 |
+
"learning_rate": 3.802247191011237e-07,
|
10835 |
+
"loss": 0.191,
|
10836 |
+
"step": 43325
|
10837 |
+
},
|
10838 |
+
{
|
10839 |
+
"epoch": 47.27,
|
10840 |
+
"learning_rate": 3.7460674157303375e-07,
|
10841 |
+
"loss": 0.1771,
|
10842 |
+
"step": 43350
|
10843 |
+
},
|
10844 |
+
{
|
10845 |
+
"epoch": 47.3,
|
10846 |
+
"learning_rate": 3.689887640449438e-07,
|
10847 |
+
"loss": 0.1942,
|
10848 |
+
"step": 43375
|
10849 |
+
},
|
10850 |
+
{
|
10851 |
+
"epoch": 47.33,
|
10852 |
+
"learning_rate": 3.63370786516854e-07,
|
10853 |
+
"loss": 0.1873,
|
10854 |
+
"step": 43400
|
10855 |
+
},
|
10856 |
+
{
|
10857 |
+
"epoch": 47.36,
|
10858 |
+
"learning_rate": 3.5775280898876406e-07,
|
10859 |
+
"loss": 0.1559,
|
10860 |
+
"step": 43425
|
10861 |
+
},
|
10862 |
+
{
|
10863 |
+
"epoch": 47.38,
|
10864 |
+
"learning_rate": 3.5213483146067424e-07,
|
10865 |
+
"loss": 0.1719,
|
10866 |
+
"step": 43450
|
10867 |
+
},
|
10868 |
+
{
|
10869 |
+
"epoch": 47.41,
|
10870 |
+
"learning_rate": 3.465168539325843e-07,
|
10871 |
+
"loss": 0.1971,
|
10872 |
+
"step": 43475
|
10873 |
+
},
|
10874 |
+
{
|
10875 |
+
"epoch": 47.44,
|
10876 |
+
"learning_rate": 3.408988764044944e-07,
|
10877 |
+
"loss": 0.1812,
|
10878 |
+
"step": 43500
|
10879 |
+
},
|
10880 |
+
{
|
10881 |
+
"epoch": 47.46,
|
10882 |
+
"learning_rate": 3.3528089887640455e-07,
|
10883 |
+
"loss": 0.1667,
|
10884 |
+
"step": 43525
|
10885 |
+
},
|
10886 |
+
{
|
10887 |
+
"epoch": 47.49,
|
10888 |
+
"learning_rate": 3.296629213483146e-07,
|
10889 |
+
"loss": 0.1625,
|
10890 |
+
"step": 43550
|
10891 |
+
},
|
10892 |
+
{
|
10893 |
+
"epoch": 47.52,
|
10894 |
+
"learning_rate": 3.240449438202248e-07,
|
10895 |
+
"loss": 0.1383,
|
10896 |
+
"step": 43575
|
10897 |
+
},
|
10898 |
+
{
|
10899 |
+
"epoch": 47.55,
|
10900 |
+
"learning_rate": 3.1842696629213487e-07,
|
10901 |
+
"loss": 0.1887,
|
10902 |
+
"step": 43600
|
10903 |
+
},
|
10904 |
+
{
|
10905 |
+
"epoch": 47.57,
|
10906 |
+
"learning_rate": 3.1280898876404494e-07,
|
10907 |
+
"loss": 0.2228,
|
10908 |
+
"step": 43625
|
10909 |
+
},
|
10910 |
+
{
|
10911 |
+
"epoch": 47.6,
|
10912 |
+
"learning_rate": 3.0719101123595506e-07,
|
10913 |
+
"loss": 0.18,
|
10914 |
+
"step": 43650
|
10915 |
+
},
|
10916 |
+
{
|
10917 |
+
"epoch": 47.63,
|
10918 |
+
"learning_rate": 3.015730337078652e-07,
|
10919 |
+
"loss": 0.1823,
|
10920 |
+
"step": 43675
|
10921 |
+
},
|
10922 |
+
{
|
10923 |
+
"epoch": 47.66,
|
10924 |
+
"learning_rate": 2.959550561797753e-07,
|
10925 |
+
"loss": 0.1498,
|
10926 |
+
"step": 43700
|
10927 |
+
},
|
10928 |
+
{
|
10929 |
+
"epoch": 47.68,
|
10930 |
+
"learning_rate": 2.903370786516854e-07,
|
10931 |
+
"loss": 0.1725,
|
10932 |
+
"step": 43725
|
10933 |
+
},
|
10934 |
+
{
|
10935 |
+
"epoch": 47.71,
|
10936 |
+
"learning_rate": 2.8471910112359555e-07,
|
10937 |
+
"loss": 0.1457,
|
10938 |
+
"step": 43750
|
10939 |
+
},
|
10940 |
+
{
|
10941 |
+
"epoch": 47.74,
|
10942 |
+
"learning_rate": 2.791011235955056e-07,
|
10943 |
+
"loss": 0.1575,
|
10944 |
+
"step": 43775
|
10945 |
+
},
|
10946 |
+
{
|
10947 |
+
"epoch": 47.76,
|
10948 |
+
"learning_rate": 2.7348314606741574e-07,
|
10949 |
+
"loss": 0.1722,
|
10950 |
+
"step": 43800
|
10951 |
+
},
|
10952 |
+
{
|
10953 |
+
"epoch": 47.79,
|
10954 |
+
"learning_rate": 2.6786516853932586e-07,
|
10955 |
+
"loss": 0.1841,
|
10956 |
+
"step": 43825
|
10957 |
+
},
|
10958 |
+
{
|
10959 |
+
"epoch": 47.82,
|
10960 |
+
"learning_rate": 2.62247191011236e-07,
|
10961 |
+
"loss": 0.1811,
|
10962 |
+
"step": 43850
|
10963 |
+
},
|
10964 |
+
{
|
10965 |
+
"epoch": 47.85,
|
10966 |
+
"learning_rate": 2.566292134831461e-07,
|
10967 |
+
"loss": 0.1808,
|
10968 |
+
"step": 43875
|
10969 |
+
},
|
10970 |
+
{
|
10971 |
+
"epoch": 47.87,
|
10972 |
+
"learning_rate": 2.5101123595505617e-07,
|
10973 |
+
"loss": 0.1564,
|
10974 |
+
"step": 43900
|
10975 |
+
},
|
10976 |
+
{
|
10977 |
+
"epoch": 47.9,
|
10978 |
+
"learning_rate": 2.453932584269663e-07,
|
10979 |
+
"loss": 0.1803,
|
10980 |
+
"step": 43925
|
10981 |
+
},
|
10982 |
+
{
|
10983 |
+
"epoch": 47.93,
|
10984 |
+
"learning_rate": 2.397752808988764e-07,
|
10985 |
+
"loss": 0.2052,
|
10986 |
+
"step": 43950
|
10987 |
+
},
|
10988 |
+
{
|
10989 |
+
"epoch": 47.96,
|
10990 |
+
"learning_rate": 2.3415730337078654e-07,
|
10991 |
+
"loss": 0.2026,
|
10992 |
+
"step": 43975
|
10993 |
+
},
|
10994 |
+
{
|
10995 |
+
"epoch": 47.98,
|
10996 |
+
"learning_rate": 2.2853932584269666e-07,
|
10997 |
+
"loss": 0.1705,
|
10998 |
+
"step": 44000
|
10999 |
+
},
|
11000 |
+
{
|
11001 |
+
"epoch": 47.98,
|
11002 |
+
"eval_cer": 7.082958776529017,
|
11003 |
+
"eval_loss": 0.47585317492485046,
|
11004 |
+
"eval_runtime": 3126.347,
|
11005 |
+
"eval_samples_per_second": 0.693,
|
11006 |
+
"eval_steps_per_second": 0.173,
|
11007 |
+
"eval_wer": 22.386295127661075,
|
11008 |
+
"step": 44000
|
11009 |
+
},
|
11010 |
+
{
|
11011 |
+
"epoch": 48.01,
|
11012 |
+
"learning_rate": 2.2292134831460676e-07,
|
11013 |
+
"loss": 0.1972,
|
11014 |
+
"step": 44025
|
11015 |
+
},
|
11016 |
+
{
|
11017 |
+
"epoch": 48.04,
|
11018 |
+
"learning_rate": 2.1752808988764047e-07,
|
11019 |
+
"loss": 0.156,
|
11020 |
+
"step": 44050
|
11021 |
+
},
|
11022 |
+
{
|
11023 |
+
"epoch": 48.06,
|
11024 |
+
"learning_rate": 2.119101123595506e-07,
|
11025 |
+
"loss": 0.1913,
|
11026 |
+
"step": 44075
|
11027 |
+
},
|
11028 |
+
{
|
11029 |
+
"epoch": 48.09,
|
11030 |
+
"learning_rate": 2.062921348314607e-07,
|
11031 |
+
"loss": 0.1576,
|
11032 |
+
"step": 44100
|
11033 |
+
},
|
11034 |
+
{
|
11035 |
+
"epoch": 48.12,
|
11036 |
+
"learning_rate": 2.0067415730337078e-07,
|
11037 |
+
"loss": 0.2132,
|
11038 |
+
"step": 44125
|
11039 |
+
},
|
11040 |
+
{
|
11041 |
+
"epoch": 48.15,
|
11042 |
+
"learning_rate": 1.950561797752809e-07,
|
11043 |
+
"loss": 0.1657,
|
11044 |
+
"step": 44150
|
11045 |
+
},
|
11046 |
+
{
|
11047 |
+
"epoch": 48.17,
|
11048 |
+
"learning_rate": 1.8943820224719102e-07,
|
11049 |
+
"loss": 0.1648,
|
11050 |
+
"step": 44175
|
11051 |
+
},
|
11052 |
+
{
|
11053 |
+
"epoch": 48.2,
|
11054 |
+
"learning_rate": 1.8382022471910115e-07,
|
11055 |
+
"loss": 0.1761,
|
11056 |
+
"step": 44200
|
11057 |
+
},
|
11058 |
+
{
|
11059 |
+
"epoch": 48.23,
|
11060 |
+
"learning_rate": 1.7820224719101127e-07,
|
11061 |
+
"loss": 0.1714,
|
11062 |
+
"step": 44225
|
11063 |
+
},
|
11064 |
+
{
|
11065 |
+
"epoch": 48.26,
|
11066 |
+
"learning_rate": 1.7258426966292134e-07,
|
11067 |
+
"loss": 0.1919,
|
11068 |
+
"step": 44250
|
11069 |
+
},
|
11070 |
+
{
|
11071 |
+
"epoch": 48.28,
|
11072 |
+
"learning_rate": 1.6696629213483146e-07,
|
11073 |
+
"loss": 0.1635,
|
11074 |
+
"step": 44275
|
11075 |
+
},
|
11076 |
+
{
|
11077 |
+
"epoch": 48.31,
|
11078 |
+
"learning_rate": 1.6134831460674158e-07,
|
11079 |
+
"loss": 0.1621,
|
11080 |
+
"step": 44300
|
11081 |
+
},
|
11082 |
+
{
|
11083 |
+
"epoch": 48.34,
|
11084 |
+
"learning_rate": 1.557303370786517e-07,
|
11085 |
+
"loss": 0.1812,
|
11086 |
+
"step": 44325
|
11087 |
+
},
|
11088 |
+
{
|
11089 |
+
"epoch": 48.36,
|
11090 |
+
"learning_rate": 1.501123595505618e-07,
|
11091 |
+
"loss": 0.1473,
|
11092 |
+
"step": 44350
|
11093 |
+
},
|
11094 |
+
{
|
11095 |
+
"epoch": 48.39,
|
11096 |
+
"learning_rate": 1.4449438202247192e-07,
|
11097 |
+
"loss": 0.1599,
|
11098 |
+
"step": 44375
|
11099 |
+
},
|
11100 |
+
{
|
11101 |
+
"epoch": 48.42,
|
11102 |
+
"learning_rate": 1.3887640449438202e-07,
|
11103 |
+
"loss": 0.1755,
|
11104 |
+
"step": 44400
|
11105 |
+
},
|
11106 |
+
{
|
11107 |
+
"epoch": 48.45,
|
11108 |
+
"learning_rate": 1.3325842696629214e-07,
|
11109 |
+
"loss": 0.1338,
|
11110 |
+
"step": 44425
|
11111 |
+
},
|
11112 |
+
{
|
11113 |
+
"epoch": 48.47,
|
11114 |
+
"learning_rate": 1.2764044943820226e-07,
|
11115 |
+
"loss": 0.1643,
|
11116 |
+
"step": 44450
|
11117 |
+
},
|
11118 |
+
{
|
11119 |
+
"epoch": 48.5,
|
11120 |
+
"learning_rate": 1.2202247191011236e-07,
|
11121 |
+
"loss": 0.1683,
|
11122 |
+
"step": 44475
|
11123 |
+
},
|
11124 |
+
{
|
11125 |
+
"epoch": 48.53,
|
11126 |
+
"learning_rate": 1.1640449438202248e-07,
|
11127 |
+
"loss": 0.2074,
|
11128 |
+
"step": 44500
|
11129 |
+
},
|
11130 |
+
{
|
11131 |
+
"epoch": 48.56,
|
11132 |
+
"learning_rate": 1.1078651685393259e-07,
|
11133 |
+
"loss": 0.1888,
|
11134 |
+
"step": 44525
|
11135 |
+
},
|
11136 |
+
{
|
11137 |
+
"epoch": 48.58,
|
11138 |
+
"learning_rate": 1.0516853932584271e-07,
|
11139 |
+
"loss": 0.1798,
|
11140 |
+
"step": 44550
|
11141 |
+
},
|
11142 |
+
{
|
11143 |
+
"epoch": 48.61,
|
11144 |
+
"learning_rate": 9.955056179775281e-08,
|
11145 |
+
"loss": 0.1636,
|
11146 |
+
"step": 44575
|
11147 |
+
},
|
11148 |
+
{
|
11149 |
+
"epoch": 48.64,
|
11150 |
+
"learning_rate": 9.393258426966293e-08,
|
11151 |
+
"loss": 0.1752,
|
11152 |
+
"step": 44600
|
11153 |
+
},
|
11154 |
+
{
|
11155 |
+
"epoch": 48.66,
|
11156 |
+
"learning_rate": 8.831460674157305e-08,
|
11157 |
+
"loss": 0.1478,
|
11158 |
+
"step": 44625
|
11159 |
+
},
|
11160 |
+
{
|
11161 |
+
"epoch": 48.69,
|
11162 |
+
"learning_rate": 8.269662921348315e-08,
|
11163 |
+
"loss": 0.1745,
|
11164 |
+
"step": 44650
|
11165 |
+
},
|
11166 |
+
{
|
11167 |
+
"epoch": 48.72,
|
11168 |
+
"learning_rate": 7.707865168539327e-08,
|
11169 |
+
"loss": 0.2115,
|
11170 |
+
"step": 44675
|
11171 |
+
},
|
11172 |
+
{
|
11173 |
+
"epoch": 48.75,
|
11174 |
+
"learning_rate": 7.146067415730338e-08,
|
11175 |
+
"loss": 0.1916,
|
11176 |
+
"step": 44700
|
11177 |
+
},
|
11178 |
+
{
|
11179 |
+
"epoch": 48.77,
|
11180 |
+
"learning_rate": 6.584269662921349e-08,
|
11181 |
+
"loss": 0.1798,
|
11182 |
+
"step": 44725
|
11183 |
+
},
|
11184 |
+
{
|
11185 |
+
"epoch": 48.8,
|
11186 |
+
"learning_rate": 6.02247191011236e-08,
|
11187 |
+
"loss": 0.1878,
|
11188 |
+
"step": 44750
|
11189 |
+
},
|
11190 |
+
{
|
11191 |
+
"epoch": 48.83,
|
11192 |
+
"learning_rate": 5.460674157303371e-08,
|
11193 |
+
"loss": 0.1734,
|
11194 |
+
"step": 44775
|
11195 |
+
},
|
11196 |
+
{
|
11197 |
+
"epoch": 48.85,
|
11198 |
+
"learning_rate": 4.8988764044943827e-08,
|
11199 |
+
"loss": 0.1588,
|
11200 |
+
"step": 44800
|
11201 |
+
},
|
11202 |
+
{
|
11203 |
+
"epoch": 48.88,
|
11204 |
+
"learning_rate": 4.3370786516853935e-08,
|
11205 |
+
"loss": 0.1622,
|
11206 |
+
"step": 44825
|
11207 |
+
},
|
11208 |
+
{
|
11209 |
+
"epoch": 48.91,
|
11210 |
+
"learning_rate": 3.775280898876405e-08,
|
11211 |
+
"loss": 0.1726,
|
11212 |
+
"step": 44850
|
11213 |
+
},
|
11214 |
+
{
|
11215 |
+
"epoch": 48.94,
|
11216 |
+
"learning_rate": 3.213483146067416e-08,
|
11217 |
+
"loss": 0.1523,
|
11218 |
+
"step": 44875
|
11219 |
+
},
|
11220 |
+
{
|
11221 |
+
"epoch": 48.96,
|
11222 |
+
"learning_rate": 2.6516853932584272e-08,
|
11223 |
+
"loss": 0.1753,
|
11224 |
+
"step": 44900
|
11225 |
+
},
|
11226 |
+
{
|
11227 |
+
"epoch": 48.99,
|
11228 |
+
"learning_rate": 2.0898876404494384e-08,
|
11229 |
+
"loss": 0.1672,
|
11230 |
+
"step": 44925
|
11231 |
+
},
|
11232 |
+
{
|
11233 |
+
"epoch": 49.02,
|
11234 |
+
"learning_rate": 1.5280898876404497e-08,
|
11235 |
+
"loss": 0.1675,
|
11236 |
+
"step": 44950
|
11237 |
+
},
|
11238 |
+
{
|
11239 |
+
"epoch": 49.05,
|
11240 |
+
"learning_rate": 9.662921348314607e-09,
|
11241 |
+
"loss": 0.1766,
|
11242 |
+
"step": 44975
|
11243 |
+
},
|
11244 |
+
{
|
11245 |
+
"epoch": 49.07,
|
11246 |
+
"learning_rate": 4.044943820224719e-09,
|
11247 |
+
"loss": 0.2007,
|
11248 |
+
"step": 45000
|
11249 |
+
},
|
11250 |
+
{
|
11251 |
+
"epoch": 49.07,
|
11252 |
+
"eval_cer": 7.1186581618974305,
|
11253 |
+
"eval_loss": 0.47667962312698364,
|
11254 |
+
"eval_runtime": 3154.6082,
|
11255 |
+
"eval_samples_per_second": 0.687,
|
11256 |
+
"eval_steps_per_second": 0.172,
|
11257 |
+
"eval_wer": 22.484947659351242,
|
11258 |
+
"step": 45000
|
11259 |
+
},
|
11260 |
+
{
|
11261 |
+
"epoch": 49.07,
|
11262 |
+
"step": 45000,
|
11263 |
+
"total_flos": 5.83519441453056e+18,
|
11264 |
+
"train_loss": 0.05880288976563348,
|
11265 |
+
"train_runtime": 37826.6777,
|
11266 |
+
"train_samples_per_second": 4.759,
|
11267 |
+
"train_steps_per_second": 1.19
|
11268 |
+
},
|
11269 |
+
{
|
11270 |
+
"epoch": 49.07,
|
11271 |
"eval_cer": 9.956974355224656,
|
11272 |
"eval_loss": 0.7910374999046326,
|
11273 |
+
"eval_runtime": 3295.3661,
|
11274 |
+
"eval_samples_per_second": 0.657,
|
11275 |
+
"eval_steps_per_second": 0.164,
|
11276 |
"eval_wer": 31.400474471700036,
|
11277 |
+
"step": 45000
|
11278 |
}
|
11279 |
],
|
11280 |
+
"max_steps": 45000,
|
11281 |
+
"num_train_epochs": 50,
|
11282 |
+
"total_flos": 5.83519441453056e+18,
|
11283 |
"trial_name": null,
|
11284 |
"trial_params": null
|
11285 |
}
|