DewiBrynJones commited on
Commit
e461203
1 Parent(s): c71568d

End of training

Browse files
README.md CHANGED
@@ -2,6 +2,8 @@
2
  license: apache-2.0
3
  base_model: facebook/wav2vec2-large-xlsr-53
4
  tags:
 
 
5
  - generated_from_trainer
6
  metrics:
7
  - wer
@@ -15,10 +17,10 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # wav2vec2-xlsr-53-ft-btb-cy
17
 
18
- This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on an unknown dataset.
19
  It achieves the following results on the evaluation set:
20
- - Loss: 1.9104
21
- - Wer: 0.8611
22
 
23
  ## Model description
24
 
 
2
  license: apache-2.0
3
  base_model: facebook/wav2vec2-large-xlsr-53
4
  tags:
5
+ - automatic-speech-recognition
6
+ - DewiBrynJones/banc-trawsgrifiadau-bangor-normalized
7
  - generated_from_trainer
8
  metrics:
9
  - wer
 
17
 
18
  # wav2vec2-xlsr-53-ft-btb-cy
19
 
20
+ This model is a fine-tuned version of [facebook/wav2vec2-large-xlsr-53](https://huggingface.co/facebook/wav2vec2-large-xlsr-53) on the DEWIBRYNJONES/BANC-TRAWSGRIFIADAU-BANGOR-NORMALIZED - DEFAULT dataset.
21
  It achieves the following results on the evaluation set:
22
+ - Loss: 0.6298
23
+ - Wer: 0.4414
24
 
25
  ## Model description
26
 
all_results.json CHANGED
@@ -1,14 +1,15 @@
1
  {
2
- "epoch": 5.66,
3
- "eval_loss": 0.5991869568824768,
4
- "eval_runtime": 213.631,
5
  "eval_samples": 5656,
6
- "eval_samples_per_second": 26.476,
7
- "eval_steps_per_second": 0.829,
8
- "eval_wer": 0.44248649327805,
9
- "train_loss": 1.8628226623535156,
10
- "train_runtime": 6582.3574,
 
11
  "train_samples": 22621,
12
- "train_samples_per_second": 48.615,
13
- "train_steps_per_second": 0.76
14
  }
 
1
  {
2
+ "epoch": 5.657708628005658,
3
+ "eval_loss": 0.6297643780708313,
4
+ "eval_runtime": 170.703,
5
  "eval_samples": 5656,
6
+ "eval_samples_per_second": 33.134,
7
+ "eval_steps_per_second": 1.037,
8
+ "eval_wer": 0.4413827414100239,
9
+ "total_flos": 1.5333464503060175e+19,
10
+ "train_loss": 2.1353194427490236,
11
+ "train_runtime": 5639.522,
12
  "train_samples": 22621,
13
+ "train_samples_per_second": 56.742,
14
+ "train_steps_per_second": 0.887
15
  }
eval_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 5.66,
3
- "eval_loss": 0.5991869568824768,
4
- "eval_runtime": 213.631,
5
  "eval_samples": 5656,
6
- "eval_samples_per_second": 26.476,
7
- "eval_steps_per_second": 0.829,
8
- "eval_wer": 0.44248649327805
9
  }
 
1
  {
2
+ "epoch": 5.657708628005658,
3
+ "eval_loss": 0.6297643780708313,
4
+ "eval_runtime": 170.703,
5
  "eval_samples": 5656,
6
+ "eval_samples_per_second": 33.134,
7
+ "eval_steps_per_second": 1.037,
8
+ "eval_wer": 0.4413827414100239
9
  }
runs/May10_07-39-17_6b3e2401f443/events.out.tfevents.1715329035.6b3e2401f443.30.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b278532640d54823aa4be70dd191911fe1c0b2762f913f5e618a8de06c8b1758
3
+ size 406
train_results.json CHANGED
@@ -1,8 +1,9 @@
1
  {
2
- "epoch": 5.66,
3
- "train_loss": 1.8628226623535156,
4
- "train_runtime": 6582.3574,
 
5
  "train_samples": 22621,
6
- "train_samples_per_second": 48.615,
7
- "train_steps_per_second": 0.76
8
  }
 
1
  {
2
+ "epoch": 5.657708628005658,
3
+ "total_flos": 1.5333464503060175e+19,
4
+ "train_loss": 2.1353194427490236,
5
+ "train_runtime": 5639.522,
6
  "train_samples": 22621,
7
+ "train_samples_per_second": 56.742,
8
+ "train_steps_per_second": 0.887
9
  }
trainer_state.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "best_metric": 0.5991869568824768,
3
  "best_model_checkpoint": "/models/hfhub/DewiBrynJones/wav2vec2-xlsr-53-ft-btb-cy/checkpoint-1000",
4
  "epoch": 5.657708628005658,
5
  "eval_steps": 500,
@@ -9,86 +9,93 @@
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 1.13,
13
- "learning_rate": 0.00014812499999999998,
14
- "loss": 4.6574,
 
15
  "step": 400
16
  },
17
  {
18
- "epoch": 1.41,
19
- "eval_loss": 1.0889060497283936,
20
- "eval_runtime": 233.5812,
21
- "eval_samples_per_second": 24.214,
22
- "eval_steps_per_second": 0.758,
23
- "eval_wer": 0.7293001633371027,
24
  "step": 500
25
  },
26
  {
27
- "epoch": 2.26,
28
- "learning_rate": 0.00029662499999999996,
29
- "loss": 1.0887,
 
30
  "step": 800
31
  },
32
  {
33
- "epoch": 2.83,
34
- "eval_loss": 0.5991869568824768,
35
- "eval_runtime": 234.4544,
36
- "eval_samples_per_second": 24.124,
37
- "eval_steps_per_second": 0.755,
38
- "eval_wer": 0.44248649327805,
39
  "step": 1000
40
  },
41
  {
42
- "epoch": 3.39,
43
- "learning_rate": 0.0002722142857142857,
44
- "loss": 0.7223,
 
45
  "step": 1200
46
  },
47
  {
48
- "epoch": 4.24,
49
- "eval_loss": 0.726735532283783,
50
- "eval_runtime": 213.9298,
51
- "eval_samples_per_second": 26.439,
52
- "eval_steps_per_second": 0.827,
53
- "eval_wer": 0.46312350797838925,
54
  "step": 1500
55
  },
56
  {
57
- "epoch": 4.53,
58
- "learning_rate": 0.0002439285714285714,
59
- "loss": 0.7162,
 
60
  "step": 1600
61
  },
62
  {
63
- "epoch": 5.66,
64
- "learning_rate": 0.00021564285714285712,
65
- "loss": 2.1295,
 
66
  "step": 2000
67
  },
68
  {
69
- "epoch": 5.66,
70
- "eval_loss": 3.0491700172424316,
71
- "eval_runtime": 217.4025,
72
- "eval_samples_per_second": 26.016,
73
- "eval_steps_per_second": 0.814,
74
- "eval_wer": 0.9983980399547682,
75
  "step": 2000
76
  },
77
  {
78
- "epoch": 5.66,
79
  "step": 2000,
80
  "total_flos": 1.5333464503060175e+19,
81
- "train_loss": 1.8628226623535156,
82
- "train_runtime": 6582.3574,
83
- "train_samples_per_second": 48.615,
84
- "train_steps_per_second": 0.76
85
  }
86
  ],
87
  "logging_steps": 400,
88
  "max_steps": 5000,
 
89
  "num_train_epochs": 15,
90
  "save_steps": 500,
91
  "total_flos": 1.5333464503060175e+19,
 
92
  "trial_name": null,
93
  "trial_params": null
94
  }
 
1
  {
2
+ "best_metric": 0.6297643780708313,
3
  "best_model_checkpoint": "/models/hfhub/DewiBrynJones/wav2vec2-xlsr-53-ft-btb-cy/checkpoint-1000",
4
  "epoch": 5.657708628005658,
5
  "eval_steps": 500,
 
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 1.1315417256011315,
13
+ "grad_norm": 1.0976253747940063,
14
+ "learning_rate": 0.000147375,
15
+ "loss": 4.9658,
16
  "step": 400
17
  },
18
  {
19
+ "epoch": 1.4144271570014144,
20
+ "eval_loss": 1.1673808097839355,
21
+ "eval_runtime": 174.6298,
22
+ "eval_samples_per_second": 32.389,
23
+ "eval_steps_per_second": 1.014,
24
+ "eval_wer": 0.802891945242413,
25
  "step": 500
26
  },
27
  {
28
+ "epoch": 2.263083451202263,
29
+ "grad_norm": 2.727048397064209,
30
+ "learning_rate": 0.00029587499999999997,
31
+ "loss": 1.2376,
32
  "step": 800
33
  },
34
  {
35
+ "epoch": 2.828854314002829,
36
+ "eval_loss": 0.6297643780708313,
37
+ "eval_runtime": 181.3206,
38
+ "eval_samples_per_second": 31.193,
39
+ "eval_steps_per_second": 0.976,
40
+ "eval_wer": 0.4413827414100239,
41
  "step": 1000
42
  },
43
  {
44
+ "epoch": 3.3946251768033946,
45
+ "grad_norm": 4.206020832061768,
46
+ "learning_rate": 0.0002723571428571428,
47
+ "loss": 0.8262,
48
  "step": 1200
49
  },
50
  {
51
+ "epoch": 4.243281471004243,
52
+ "eval_loss": 1.9106061458587646,
53
+ "eval_runtime": 170.5081,
54
+ "eval_samples_per_second": 33.171,
55
+ "eval_steps_per_second": 1.038,
56
+ "eval_wer": 0.8612444030748985,
57
  "step": 1500
58
  },
59
  {
60
+ "epoch": 4.526166902404526,
61
+ "grad_norm": 0.0,
62
+ "learning_rate": 0.0002462142857142857,
63
+ "loss": 1.653,
64
  "step": 1600
65
  },
66
  {
67
+ "epoch": 5.657708628005658,
68
+ "grad_norm": 0.0,
69
+ "learning_rate": 0.00021792857142857142,
70
+ "loss": 1.9941,
71
  "step": 2000
72
  },
73
  {
74
+ "epoch": 5.657708628005658,
75
+ "eval_loss": 1.9103525876998901,
76
+ "eval_runtime": 170.4356,
77
+ "eval_samples_per_second": 33.186,
78
+ "eval_steps_per_second": 1.039,
79
+ "eval_wer": 0.8610999662980854,
80
  "step": 2000
81
  },
82
  {
83
+ "epoch": 5.657708628005658,
84
  "step": 2000,
85
  "total_flos": 1.5333464503060175e+19,
86
+ "train_loss": 2.1353194427490236,
87
+ "train_runtime": 5639.522,
88
+ "train_samples_per_second": 56.742,
89
+ "train_steps_per_second": 0.887
90
  }
91
  ],
92
  "logging_steps": 400,
93
  "max_steps": 5000,
94
+ "num_input_tokens_seen": 0,
95
  "num_train_epochs": 15,
96
  "save_steps": 500,
97
  "total_flos": 1.5333464503060175e+19,
98
+ "train_batch_size": 32,
99
  "trial_name": null,
100
  "trial_params": null
101
  }