j0eBee commited on
Commit
0480aec
1 Parent(s): a13bd88

End of training

Browse files
README.md CHANGED
@@ -1,7 +1,12 @@
1
  ---
 
 
2
  license: apache-2.0
3
  base_model: facebook/wav2vec2-xls-r-300m
4
  tags:
 
 
 
5
  - generated_from_trainer
6
  datasets:
7
  - common_voice_17_0
@@ -14,11 +19,11 @@ model-index:
14
  name: Automatic Speech Recognition
15
  type: automatic-speech-recognition
16
  dataset:
17
- name: common_voice_17_0
18
  type: common_voice_17_0
19
  config: tw
20
  split: None
21
- args: tw
22
  metrics:
23
  - name: Wer
24
  type: wer
@@ -30,9 +35,9 @@ should probably proofread and complete it, then remove this comment. -->
30
 
31
  # wav2vec2-twi-adapter
32
 
33
- This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the common_voice_17_0 dataset.
34
  It achieves the following results on the evaluation set:
35
- - Loss: 2.4281
36
  - Wer: 1.0
37
  - Cer: 1.0
38
 
 
1
  ---
2
+ language:
3
+ - tw
4
  license: apache-2.0
5
  base_model: facebook/wav2vec2-xls-r-300m
6
  tags:
7
+ - automatic-speech-recognition
8
+ - mozilla-foundation/common_voice_17_0
9
+ - mms
10
  - generated_from_trainer
11
  datasets:
12
  - common_voice_17_0
 
19
  name: Automatic Speech Recognition
20
  type: automatic-speech-recognition
21
  dataset:
22
+ name: MOZILLA-FOUNDATION/COMMON_VOICE_17_0 - TW
23
  type: common_voice_17_0
24
  config: tw
25
  split: None
26
+ args: 'Config: tw, Training split: train, Eval split: validation+test'
27
  metrics:
28
  - name: Wer
29
  type: wer
 
35
 
36
  # wav2vec2-twi-adapter
37
 
38
+ This model is a fine-tuned version of [facebook/wav2vec2-xls-r-300m](https://huggingface.co/facebook/wav2vec2-xls-r-300m) on the MOZILLA-FOUNDATION/COMMON_VOICE_17_0 - TW dataset.
39
  It achieves the following results on the evaluation set:
40
+ - Loss: 2.4092
41
  - Wer: 1.0
42
  - Cer: 1.0
43
 
adapter.aka.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7664ed8cbab872e1072a86fa0bf31d93a62cb3824ce10e27599544797e4bc49a
3
- size 3590904
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dd8f507e47e198eacd2db4170d87b37cac423669c10ba32972457fca6b52835e
3
+ size 3615512
all_results.json CHANGED
@@ -1,16 +1,16 @@
1
  {
2
- "epoch": 4.0,
3
- "eval_cer": 5.130434782608695,
4
- "eval_loss": 35.69469451904297,
5
- "eval_runtime": 0.7985,
6
- "eval_samples": 3,
7
- "eval_samples_per_second": 3.757,
8
- "eval_steps_per_second": 1.252,
9
  "eval_wer": 1.0,
10
- "total_flos": 3.89806936817664e+16,
11
- "train_loss": 14.43960952758789,
12
- "train_runtime": 61.1436,
13
- "train_samples": 53,
14
- "train_samples_per_second": 3.467,
15
- "train_steps_per_second": 0.131
16
  }
 
1
  {
2
+ "epoch": 26.666666666666668,
3
+ "eval_cer": 1.0,
4
+ "eval_loss": 2.409198522567749,
5
+ "eval_runtime": 0.9391,
6
+ "eval_samples": 9,
7
+ "eval_samples_per_second": 9.584,
8
+ "eval_steps_per_second": 1.065,
9
  "eval_wer": 1.0,
10
+ "total_flos": 5.3095066035971686e+17,
11
+ "train_loss": 7.207664489746094,
12
+ "train_runtime": 305.0878,
13
+ "train_samples": 136,
14
+ "train_samples_per_second": 13.373,
15
+ "train_steps_per_second": 0.393
16
  }
eval_results.json CHANGED
@@ -1,10 +1,10 @@
1
  {
2
- "epoch": 4.0,
3
- "eval_cer": 5.130434782608695,
4
- "eval_loss": 35.69469451904297,
5
- "eval_runtime": 0.7985,
6
- "eval_samples": 3,
7
- "eval_samples_per_second": 3.757,
8
- "eval_steps_per_second": 1.252,
9
  "eval_wer": 1.0
10
  }
 
1
  {
2
+ "epoch": 26.666666666666668,
3
+ "eval_cer": 1.0,
4
+ "eval_loss": 2.409198522567749,
5
+ "eval_runtime": 0.9391,
6
+ "eval_samples": 9,
7
+ "eval_samples_per_second": 9.584,
8
+ "eval_steps_per_second": 1.065,
9
  "eval_wer": 1.0
10
  }
runs/May24_03-34-21_cc57b0143c57/events.out.tfevents.1716522040.cc57b0143c57.714.1 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6b2973e6c1914c5d9b1d7ae85bd25bd6e57bf7aa7bde90d445cea6acbcd244ab
3
+ size 398
train_results.json CHANGED
@@ -1,9 +1,9 @@
1
  {
2
- "epoch": 4.0,
3
- "total_flos": 3.89806936817664e+16,
4
- "train_loss": 14.43960952758789,
5
- "train_runtime": 61.1436,
6
- "train_samples": 53,
7
- "train_samples_per_second": 3.467,
8
- "train_steps_per_second": 0.131
9
  }
 
1
  {
2
+ "epoch": 26.666666666666668,
3
+ "total_flos": 5.3095066035971686e+17,
4
+ "train_loss": 7.207664489746094,
5
+ "train_runtime": 305.0878,
6
+ "train_samples": 136,
7
+ "train_samples_per_second": 13.373,
8
+ "train_steps_per_second": 0.393
9
  }
trainer_state.json CHANGED
@@ -1,27 +1,47 @@
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
- "epoch": 4.0,
5
  "eval_steps": 50,
6
- "global_step": 8,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
- "epoch": 4.0,
13
- "step": 8,
14
- "total_flos": 3.89806936817664e+16,
15
- "train_loss": 14.43960952758789,
16
- "train_runtime": 61.1436,
17
- "train_samples_per_second": 3.467,
18
- "train_steps_per_second": 0.131
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
19
  }
20
  ],
21
  "logging_steps": 500,
22
- "max_steps": 8,
23
  "num_input_tokens_seen": 0,
24
- "num_train_epochs": 4,
25
  "save_steps": 200,
26
  "stateful_callbacks": {
27
  "TrainerControl": {
@@ -35,7 +55,7 @@
35
  "attributes": {}
36
  }
37
  },
38
- "total_flos": 3.89806936817664e+16,
39
  "train_batch_size": 8,
40
  "trial_name": null,
41
  "trial_params": null
 
1
  {
2
  "best_metric": null,
3
  "best_model_checkpoint": null,
4
+ "epoch": 26.666666666666668,
5
  "eval_steps": 50,
6
+ "global_step": 120,
7
  "is_hyper_param_search": false,
8
  "is_local_process_zero": true,
9
  "is_world_process_zero": true,
10
  "log_history": [
11
  {
12
+ "epoch": 11.11111111111111,
13
+ "eval_cer": 1.0,
14
+ "eval_loss": 5.89304256439209,
15
+ "eval_runtime": 0.9784,
16
+ "eval_samples_per_second": 9.198,
17
+ "eval_steps_per_second": 1.022,
18
+ "eval_wer": 1.0,
19
+ "step": 50
20
+ },
21
+ {
22
+ "epoch": 22.22222222222222,
23
+ "eval_cer": 1.0,
24
+ "eval_loss": 2.4281415939331055,
25
+ "eval_runtime": 0.8402,
26
+ "eval_samples_per_second": 10.711,
27
+ "eval_steps_per_second": 1.19,
28
+ "eval_wer": 1.0,
29
+ "step": 100
30
+ },
31
+ {
32
+ "epoch": 26.666666666666668,
33
+ "step": 120,
34
+ "total_flos": 5.3095066035971686e+17,
35
+ "train_loss": 7.207664489746094,
36
+ "train_runtime": 305.0878,
37
+ "train_samples_per_second": 13.373,
38
+ "train_steps_per_second": 0.393
39
  }
40
  ],
41
  "logging_steps": 500,
42
+ "max_steps": 120,
43
  "num_input_tokens_seen": 0,
44
+ "num_train_epochs": 30,
45
  "save_steps": 200,
46
  "stateful_callbacks": {
47
  "TrainerControl": {
 
55
  "attributes": {}
56
  }
57
  },
58
+ "total_flos": 5.3095066035971686e+17,
59
  "train_batch_size": 8,
60
  "trial_name": null,
61
  "trial_params": null