Lasion commited on
Commit
6c2d6f2
1 Parent(s): d545ff6

End of training

Browse files
README.md CHANGED
@@ -1,11 +1,28 @@
1
  ---
2
  license: apache-2.0
 
3
  tags:
4
  - generated_from_trainer
5
- base_model: lasion/wav2vec2-base-timit-demo-google-colab
 
 
 
6
  model-index:
7
  - name: wav2vec2-base-timit-demo-google-colab
8
- results: []
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  ---
10
 
11
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
@@ -13,7 +30,10 @@ should probably proofread and complete it, then remove this comment. -->
13
 
14
  # wav2vec2-base-timit-demo-google-colab
15
 
16
- This model is a fine-tuned version of [lasion/wav2vec2-base-timit-demo-google-colab](https://huggingface.co/lasion/wav2vec2-base-timit-demo-google-colab) on the None dataset.
 
 
 
17
 
18
  ## Model description
19
 
@@ -39,16 +59,37 @@ The following hyperparameters were used during training:
39
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
40
  - lr_scheduler_type: linear
41
  - lr_scheduler_warmup_steps: 1000
42
- - num_epochs: 1
43
  - mixed_precision_training: Native AMP
44
 
45
  ### Training results
46
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
47
 
48
 
49
  ### Framework versions
50
 
51
- - Transformers 4.39.2
52
- - Pytorch 2.2.1+cu121
53
  - Datasets 1.18.3
54
  - Tokenizers 0.15.2
 
1
  ---
2
  license: apache-2.0
3
+ base_model: facebook/wav2vec2-base
4
  tags:
5
  - generated_from_trainer
6
+ datasets:
7
+ - timit_asr
8
+ metrics:
9
+ - wer
10
  model-index:
11
  - name: wav2vec2-base-timit-demo-google-colab
12
+ results:
13
+ - task:
14
+ name: Automatic Speech Recognition
15
+ type: automatic-speech-recognition
16
+ dataset:
17
+ name: timit_asr
18
+ type: timit_asr
19
+ config: clean
20
+ split: test
21
+ args: clean
22
+ metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 0.3367100820067535
26
  ---
27
 
28
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
 
30
 
31
  # wav2vec2-base-timit-demo-google-colab
32
 
33
+ This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the timit_asr dataset.
34
+ It achieves the following results on the evaluation set:
35
+ - Loss: 0.4634
36
+ - Wer: 0.3367
37
 
38
  ## Model description
39
 
 
59
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
60
  - lr_scheduler_type: linear
61
  - lr_scheduler_warmup_steps: 1000
62
+ - num_epochs: 20
63
  - mixed_precision_training: Native AMP
64
 
65
  ### Training results
66
 
67
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
68
+ |:-------------:|:-----:|:----:|:---------------:|:------:|
69
+ | 3.6019 | 1.0 | 500 | 2.4586 | 1.0 |
70
+ | 0.9594 | 2.01 | 1000 | 0.5023 | 0.5122 |
71
+ | 0.4324 | 3.01 | 1500 | 0.4808 | 0.4703 |
72
+ | 0.2991 | 4.02 | 2000 | 0.4098 | 0.4208 |
73
+ | 0.2257 | 5.02 | 2500 | 0.4883 | 0.4264 |
74
+ | 0.18 | 6.02 | 3000 | 0.4441 | 0.3914 |
75
+ | 0.1524 | 7.03 | 3500 | 0.4360 | 0.3869 |
76
+ | 0.1315 | 8.03 | 4000 | 0.4448 | 0.3783 |
77
+ | 0.1101 | 9.04 | 4500 | 0.4570 | 0.3704 |
78
+ | 0.1017 | 10.04 | 5000 | 0.4252 | 0.3680 |
79
+ | 0.0863 | 11.04 | 5500 | 0.4492 | 0.3606 |
80
+ | 0.0798 | 12.05 | 6000 | 0.4241 | 0.3604 |
81
+ | 0.0688 | 13.05 | 6500 | 0.4585 | 0.3535 |
82
+ | 0.0608 | 14.06 | 7000 | 0.4491 | 0.3488 |
83
+ | 0.0524 | 15.06 | 7500 | 0.4550 | 0.3456 |
84
+ | 0.0502 | 16.06 | 8000 | 0.4570 | 0.3453 |
85
+ | 0.0458 | 17.07 | 8500 | 0.4680 | 0.3421 |
86
+ | 0.0395 | 18.07 | 9000 | 0.4663 | 0.3390 |
87
+ | 0.0352 | 19.08 | 9500 | 0.4634 | 0.3367 |
88
 
89
 
90
  ### Framework versions
91
 
92
+ - Transformers 4.38.2
93
+ - Pytorch 2.1.2
94
  - Datasets 1.18.3
95
  - Tokenizers 0.15.2
config.json CHANGED
@@ -1,5 +1,5 @@
1
  {
2
- "_name_or_path": "lasion/wav2vec2-base-timit-demo-google-colab",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
@@ -111,7 +111,7 @@
111
  1
112
  ],
113
  "torch_dtype": "float32",
114
- "transformers_version": "4.39.2",
115
  "use_weighted_layer_sum": false,
116
  "vocab_size": 32,
117
  "xvector_output_dim": 512
 
1
  {
2
+ "_name_or_path": "facebook/wav2vec2-base",
3
  "activation_dropout": 0.0,
4
  "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
 
111
  1
112
  ],
113
  "torch_dtype": "float32",
114
+ "transformers_version": "4.38.2",
115
  "use_weighted_layer_sum": false,
116
  "vocab_size": 32,
117
  "xvector_output_dim": 512
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:47cc39eb1e68d64c87b876f5522d9c06d54f396355f0fb6ac81ba644a2758bda
3
  size 377611120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3398016c9678713fa6ab60f801bfd5ce6bcb7258ed3858651ea739b6937dd4f9
3
  size 377611120
runs/Apr02_09-05-54_b1a7da762c46/events.out.tfevents.1712048865.b1a7da762c46.25.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e844b6ec9bde84469770d4dbf46b62f8012d72db3aa50c268b733f8d4a88ca5a
3
+ size 16733
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:bd52b52b5640c3ccd1e37582eec25c7b4f22add78018650cbf8243781dd414cc
3
  size 4920
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:5642dbfd69a7f84e3646d86738d742370286f9da7336d58d085df4e596c5128f
3
  size 4920