jadorantes2 commited on
Commit
4f8eba6
1 Parent(s): e2e8ad9

End of training

Browse files
Files changed (5) hide show
  1. README.md +84 -53
  2. config.json +110 -110
  3. model.safetensors +1 -1
  4. preprocessor_config.json +9 -10
  5. training_args.bin +1 -1
README.md CHANGED
@@ -1,53 +1,84 @@
1
- ---
2
- license: apache-2.0
3
- base_model: facebook/wav2vec2-base-960h
4
- tags:
5
- - generated_from_trainer
6
- model-index:
7
- - name: my_awesome_asr_mind_model6e-5
8
- results: []
9
- ---
10
-
11
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
12
- should probably proofread and complete it, then remove this comment. -->
13
-
14
- [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jadorantes2-utep/%3Cmy-amazing-project6000%3E/runs/3kdv4p7m)
15
- # my_awesome_asr_mind_model6e-5
16
-
17
- This model is a fine-tuned version of [facebook/wav2vec2-base-960h](https://huggingface.co/facebook/wav2vec2-base-960h) on an unknown dataset.
18
-
19
- ## Model description
20
-
21
- More information needed
22
-
23
- ## Intended uses & limitations
24
-
25
- More information needed
26
-
27
- ## Training and evaluation data
28
-
29
- More information needed
30
-
31
- ## Training procedure
32
-
33
- ### Training hyperparameters
34
-
35
- The following hyperparameters were used during training:
36
- - learning_rate: 6e-05
37
- - train_batch_size: 8
38
- - eval_batch_size: 8
39
- - seed: 42
40
- - gradient_accumulation_steps: 2
41
- - total_train_batch_size: 16
42
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
43
- - lr_scheduler_type: linear
44
- - lr_scheduler_warmup_steps: 500
45
- - training_steps: 3000
46
- - mixed_precision_training: Native AMP
47
-
48
- ### Framework versions
49
-
50
- - Transformers 4.42.3
51
- - Pytorch 2.3.1
52
- - Datasets 2.20.0
53
- - Tokenizers 0.19.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ license: apache-2.0
3
+ base_model: facebook/wav2vec2-base-960h
4
+ tags:
5
+ - generated_from_trainer
6
+ datasets:
7
+ - ami
8
+ metrics:
9
+ - wer
10
+ model-index:
11
+ - name: my_awesome_asr_mind_model6e-5
12
+ results:
13
+ - task:
14
+ name: Automatic Speech Recognition
15
+ type: automatic-speech-recognition
16
+ dataset:
17
+ name: ami
18
+ type: ami
19
+ config: ihm
20
+ split: None
21
+ args: ihm
22
+ metrics:
23
+ - name: Wer
24
+ type: wer
25
+ value: 0.26252597552528284
26
+ ---
27
+
28
+ <!-- This model card has been generated automatically according to the information the Trainer had access to. You
29
+ should probably proofread and complete it, then remove this comment. -->
30
+
31
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jadorantes2-utep/%3Cmy-amazing-projecttokenizer6e-5%3E/runs/ujcw1zru)
32
+ [<img src="https://raw.githubusercontent.com/wandb/assets/main/wandb-github-badge-28.svg" alt="Visualize in Weights & Biases" width="200" height="32"/>](https://wandb.ai/jadorantes2-utep/%3Cmy-amazing-projecttokenizer6e-5%3E/runs/ujcw1zru)
33
+ # my_awesome_asr_mind_model6e-5
34
+
35
+ This model is a fine-tuned version of [facebook/wav2vec2-base-960h](https://huggingface.co/facebook/wav2vec2-base-960h) on the ami dataset.
36
+ It achieves the following results on the evaluation set:
37
+ - Loss: 0.9652
38
+ - Wer: 0.2625
39
+
40
+ ## Model description
41
+
42
+ More information needed
43
+
44
+ ## Intended uses & limitations
45
+
46
+ More information needed
47
+
48
+ ## Training and evaluation data
49
+
50
+ More information needed
51
+
52
+ ## Training procedure
53
+
54
+ ### Training hyperparameters
55
+
56
+ The following hyperparameters were used during training:
57
+ - learning_rate: 6e-05
58
+ - train_batch_size: 32
59
+ - eval_batch_size: 8
60
+ - seed: 42
61
+ - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
62
+ - lr_scheduler_type: linear
63
+ - lr_scheduler_warmup_steps: 1000
64
+ - training_steps: 3000
65
+ - mixed_precision_training: Native AMP
66
+
67
+ ### Training results
68
+
69
+ | Training Loss | Epoch | Step | Validation Loss | Wer |
70
+ |:-------------:|:-------:|:----:|:---------------:|:------:|
71
+ | 3.1601 | 15.1515 | 500 | 3.1815 | 1.0 |
72
+ | 3.0665 | 30.3030 | 1000 | 3.5100 | 1.0 |
73
+ | 2.1863 | 45.4545 | 1500 | 1.2838 | 0.3812 |
74
+ | 0.9609 | 60.6061 | 2000 | 0.9112 | 0.2863 |
75
+ | 0.6826 | 75.7576 | 2500 | 0.9450 | 0.2667 |
76
+ | 0.5687 | 90.9091 | 3000 | 0.9652 | 0.2625 |
77
+
78
+
79
+ ### Framework versions
80
+
81
+ - Transformers 4.42.4
82
+ - Pytorch 2.3.1+cu121
83
+ - Datasets 2.20.0
84
+ - Tokenizers 0.19.1
config.json CHANGED
@@ -1,110 +1,110 @@
1
- {
2
- "_name_or_path": "facebook/wav2vec2-base-960h",
3
- "activation_dropout": 0.1,
4
- "adapter_attn_dim": null,
5
- "adapter_kernel_size": 3,
6
- "adapter_stride": 2,
7
- "add_adapter": false,
8
- "apply_spec_augment": true,
9
- "architectures": [
10
- "Wav2Vec2ForCTC"
11
- ],
12
- "attention_dropout": 0.1,
13
- "bos_token_id": 1,
14
- "classifier_proj_size": 256,
15
- "codevector_dim": 256,
16
- "contrastive_logits_temperature": 0.1,
17
- "conv_bias": false,
18
- "conv_dim": [
19
- 512,
20
- 512,
21
- 512,
22
- 512,
23
- 512,
24
- 512,
25
- 512
26
- ],
27
- "conv_kernel": [
28
- 10,
29
- 3,
30
- 3,
31
- 3,
32
- 3,
33
- 2,
34
- 2
35
- ],
36
- "conv_stride": [
37
- 5,
38
- 2,
39
- 2,
40
- 2,
41
- 2,
42
- 2,
43
- 2
44
- ],
45
- "ctc_loss_reduction": "mean",
46
- "ctc_zero_infinity": false,
47
- "diversity_loss_weight": 0.1,
48
- "do_stable_layer_norm": false,
49
- "eos_token_id": 2,
50
- "feat_extract_activation": "gelu",
51
- "feat_extract_dropout": 0.0,
52
- "feat_extract_norm": "group",
53
- "feat_proj_dropout": 0.1,
54
- "feat_quantizer_dropout": 0.0,
55
- "final_dropout": 0.1,
56
- "gradient_checkpointing": false,
57
- "hidden_act": "gelu",
58
- "hidden_dropout": 0.1,
59
- "hidden_dropout_prob": 0.1,
60
- "hidden_size": 768,
61
- "initializer_range": 0.02,
62
- "intermediate_size": 3072,
63
- "layer_norm_eps": 1e-05,
64
- "layerdrop": 0.1,
65
- "mask_feature_length": 10,
66
- "mask_feature_min_masks": 0,
67
- "mask_feature_prob": 0.0,
68
- "mask_time_length": 10,
69
- "mask_time_min_masks": 2,
70
- "mask_time_prob": 0.05,
71
- "model_type": "wav2vec2",
72
- "num_adapter_layers": 3,
73
- "num_attention_heads": 12,
74
- "num_codevector_groups": 2,
75
- "num_codevectors_per_group": 320,
76
- "num_conv_pos_embedding_groups": 16,
77
- "num_conv_pos_embeddings": 128,
78
- "num_feat_extract_layers": 7,
79
- "num_hidden_layers": 12,
80
- "num_negatives": 100,
81
- "output_hidden_size": 768,
82
- "pad_token_id": 0,
83
- "proj_codevector_dim": 256,
84
- "tdnn_dilation": [
85
- 1,
86
- 2,
87
- 3,
88
- 1,
89
- 1
90
- ],
91
- "tdnn_dim": [
92
- 512,
93
- 512,
94
- 512,
95
- 512,
96
- 1500
97
- ],
98
- "tdnn_kernel": [
99
- 5,
100
- 3,
101
- 3,
102
- 1,
103
- 1
104
- ],
105
- "torch_dtype": "float32",
106
- "transformers_version": "4.42.3",
107
- "use_weighted_layer_sum": false,
108
- "vocab_size": 32,
109
- "xvector_output_dim": 512
110
- }
 
1
+ {
2
+ "_name_or_path": "facebook/wav2vec2-base-960h",
3
+ "activation_dropout": 0.1,
4
+ "adapter_attn_dim": null,
5
+ "adapter_kernel_size": 3,
6
+ "adapter_stride": 2,
7
+ "add_adapter": false,
8
+ "apply_spec_augment": true,
9
+ "architectures": [
10
+ "Wav2Vec2ForCTC"
11
+ ],
12
+ "attention_dropout": 0.1,
13
+ "bos_token_id": 1,
14
+ "classifier_proj_size": 256,
15
+ "codevector_dim": 256,
16
+ "contrastive_logits_temperature": 0.1,
17
+ "conv_bias": false,
18
+ "conv_dim": [
19
+ 512,
20
+ 512,
21
+ 512,
22
+ 512,
23
+ 512,
24
+ 512,
25
+ 512
26
+ ],
27
+ "conv_kernel": [
28
+ 10,
29
+ 3,
30
+ 3,
31
+ 3,
32
+ 3,
33
+ 2,
34
+ 2
35
+ ],
36
+ "conv_stride": [
37
+ 5,
38
+ 2,
39
+ 2,
40
+ 2,
41
+ 2,
42
+ 2,
43
+ 2
44
+ ],
45
+ "ctc_loss_reduction": "mean",
46
+ "ctc_zero_infinity": false,
47
+ "diversity_loss_weight": 0.1,
48
+ "do_stable_layer_norm": false,
49
+ "eos_token_id": 2,
50
+ "feat_extract_activation": "gelu",
51
+ "feat_extract_dropout": 0.0,
52
+ "feat_extract_norm": "group",
53
+ "feat_proj_dropout": 0.1,
54
+ "feat_quantizer_dropout": 0.0,
55
+ "final_dropout": 0.1,
56
+ "gradient_checkpointing": false,
57
+ "hidden_act": "gelu",
58
+ "hidden_dropout": 0.1,
59
+ "hidden_dropout_prob": 0.1,
60
+ "hidden_size": 768,
61
+ "initializer_range": 0.02,
62
+ "intermediate_size": 3072,
63
+ "layer_norm_eps": 1e-05,
64
+ "layerdrop": 0.1,
65
+ "mask_feature_length": 10,
66
+ "mask_feature_min_masks": 0,
67
+ "mask_feature_prob": 0.0,
68
+ "mask_time_length": 10,
69
+ "mask_time_min_masks": 2,
70
+ "mask_time_prob": 0.05,
71
+ "model_type": "wav2vec2",
72
+ "num_adapter_layers": 3,
73
+ "num_attention_heads": 12,
74
+ "num_codevector_groups": 2,
75
+ "num_codevectors_per_group": 320,
76
+ "num_conv_pos_embedding_groups": 16,
77
+ "num_conv_pos_embeddings": 128,
78
+ "num_feat_extract_layers": 7,
79
+ "num_hidden_layers": 12,
80
+ "num_negatives": 100,
81
+ "output_hidden_size": 768,
82
+ "pad_token_id": 29,
83
+ "proj_codevector_dim": 256,
84
+ "tdnn_dilation": [
85
+ 1,
86
+ 2,
87
+ 3,
88
+ 1,
89
+ 1
90
+ ],
91
+ "tdnn_dim": [
92
+ 512,
93
+ 512,
94
+ 512,
95
+ 512,
96
+ 1500
97
+ ],
98
+ "tdnn_kernel": [
99
+ 5,
100
+ 3,
101
+ 3,
102
+ 1,
103
+ 1
104
+ ],
105
+ "torch_dtype": "float32",
106
+ "transformers_version": "4.42.4",
107
+ "use_weighted_layer_sum": false,
108
+ "vocab_size": 32,
109
+ "xvector_output_dim": 512
110
+ }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fb39ca45e5dc918f0207a5d39fad8e8e13bbe7ecd076ad01ca8092f489cf9d2a
3
  size 377611120
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fb14fa2a545f16585142e65ac1def8e56c2613a3e69b9c3766ffa15076410380
3
  size 377611120
preprocessor_config.json CHANGED
@@ -1,10 +1,9 @@
1
- {
2
- "do_normalize": true,
3
- "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
- "feature_size": 1,
5
- "padding_side": "right",
6
- "padding_value": 0.0,
7
- "processor_class": "Wav2Vec2Processor",
8
- "return_attention_mask": false,
9
- "sampling_rate": 16000
10
- }
 
1
+ {
2
+ "do_normalize": true,
3
+ "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
+ "feature_size": 1,
5
+ "padding_side": "right",
6
+ "padding_value": 0.0,
7
+ "return_attention_mask": false,
8
+ "sampling_rate": 16000
9
+ }
 
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2d369f52336d929e645e2986fafb2cf70a22bf2f9d98f03ac2850f7e605106d4
3
  size 5112
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:73e2464a5cceea747b84cea9eea48be80bb3defeed401b1d0d62b6946d8c74e3
3
  size 5112