deryauysal commited on
Commit
ab64f81
1 Parent(s): c60fe8a

End of training

Browse files
Files changed (4) hide show
  1. README.md +6 -2
  2. config.json +78 -16
  3. pytorch_model.bin +2 -2
  4. training_args.bin +1 -1
README.md CHANGED
@@ -1,6 +1,6 @@
1
  ---
2
  license: cc-by-nc-4.0
3
- base_model: facebook/mms-1b-all
4
  tags:
5
  - generated_from_trainer
6
  datasets:
@@ -15,7 +15,7 @@ should probably proofread and complete it, then remove this comment. -->
15
 
16
  # wav2vec2-large-mms-1b-turkish-colab
17
 
18
- This model is a fine-tuned version of [facebook/mms-1b-all](https://huggingface.co/facebook/mms-1b-all) on the common_voice_6_1 dataset.
19
 
20
  ## Model description
21
 
@@ -43,6 +43,10 @@ The following hyperparameters were used during training:
43
  - lr_scheduler_warmup_steps: 100
44
  - num_epochs: 4
45
 
 
 
 
 
46
  ### Framework versions
47
 
48
  - Transformers 4.33.0
 
1
  ---
2
  license: cc-by-nc-4.0
3
+ base_model: facebook/mms-tts-tur
4
  tags:
5
  - generated_from_trainer
6
  datasets:
 
15
 
16
  # wav2vec2-large-mms-1b-turkish-colab
17
 
18
+ This model is a fine-tuned version of [facebook/mms-tts-tur](https://huggingface.co/facebook/mms-tts-tur) on the common_voice_6_1 dataset.
19
 
20
  ## Model description
21
 
 
43
  - lr_scheduler_warmup_steps: 100
44
  - num_epochs: 4
45
 
46
+ ### Training results
47
+
48
+
49
+
50
  ### Framework versions
51
 
52
  - Transformers 4.33.0
config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
- "_name_or_path": "facebook/mms-1b-all",
3
- "activation_dropout": 0.05,
4
- "adapter_attn_dim": 16,
5
  "adapter_kernel_size": 3,
6
  "adapter_stride": 2,
7
  "add_adapter": false,
@@ -12,9 +12,9 @@
12
  "attention_dropout": 0.0,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
- "codevector_dim": 1024,
16
  "contrastive_logits_temperature": 0.1,
17
- "conv_bias": true,
18
  "conv_dim": [
19
  512,
20
  512,
@@ -44,22 +44,33 @@
44
  ],
45
  "ctc_loss_reduction": "mean",
46
  "ctc_zero_infinity": false,
 
 
47
  "diversity_loss_weight": 0.1,
48
- "do_stable_layer_norm": true,
 
 
 
 
 
 
49
  "eos_token_id": 2,
50
  "feat_extract_activation": "gelu",
51
- "feat_extract_dropout": 0.0,
52
- "feat_extract_norm": "layer",
53
  "feat_proj_dropout": 0.0,
54
  "feat_quantizer_dropout": 0.0,
55
- "final_dropout": 0.05,
56
- "hidden_act": "gelu",
 
 
 
57
  "hidden_dropout": 0.0,
58
- "hidden_size": 1280,
59
  "initializer_range": 0.02,
60
- "intermediate_size": 5120,
61
  "layer_norm_eps": 1e-05,
62
  "layerdrop": 0.0,
 
63
  "mask_feature_length": 10,
64
  "mask_feature_min_masks": 0,
65
  "mask_feature_prob": 0.0,
@@ -67,18 +78,50 @@
67
  "mask_time_min_masks": 2,
68
  "mask_time_prob": 0.05,
69
  "model_type": "wav2vec2",
 
 
70
  "num_adapter_layers": 3,
71
- "num_attention_heads": 16,
72
  "num_codevector_groups": 2,
73
  "num_codevectors_per_group": 320,
74
  "num_conv_pos_embedding_groups": 16,
75
  "num_conv_pos_embeddings": 128,
76
  "num_feat_extract_layers": 7,
77
- "num_hidden_layers": 48,
78
  "num_negatives": 100,
79
- "output_hidden_size": 1280,
 
80
  "pad_token_id": 39,
81
- "proj_codevector_dim": 1024,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
82
  "tdnn_dilation": [
83
  1,
84
  2,
@@ -102,7 +145,26 @@
102
  ],
103
  "torch_dtype": "float32",
104
  "transformers_version": "4.33.0",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
105
  "use_weighted_layer_sum": false,
106
  "vocab_size": 42,
 
 
 
 
107
  "xvector_output_dim": 512
108
  }
 
1
  {
2
+ "_name_or_path": "facebook/mms-tts-tur",
3
+ "activation_dropout": 0.1,
4
+ "adapter_attn_dim": null,
5
  "adapter_kernel_size": 3,
6
  "adapter_stride": 2,
7
  "add_adapter": false,
 
12
  "attention_dropout": 0.0,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
+ "codevector_dim": 256,
16
  "contrastive_logits_temperature": 0.1,
17
+ "conv_bias": false,
18
  "conv_dim": [
19
  512,
20
  512,
 
44
  ],
45
  "ctc_loss_reduction": "mean",
46
  "ctc_zero_infinity": false,
47
+ "depth_separable_channels": 2,
48
+ "depth_separable_num_layers": 3,
49
  "diversity_loss_weight": 0.1,
50
+ "do_stable_layer_norm": false,
51
+ "duration_predictor_dropout": 0.5,
52
+ "duration_predictor_filter_channels": 256,
53
+ "duration_predictor_flow_bins": 10,
54
+ "duration_predictor_kernel_size": 3,
55
+ "duration_predictor_num_flows": 4,
56
+ "duration_predictor_tail_bound": 5.0,
57
  "eos_token_id": 2,
58
  "feat_extract_activation": "gelu",
59
+ "feat_extract_norm": "group",
 
60
  "feat_proj_dropout": 0.0,
61
  "feat_quantizer_dropout": 0.0,
62
+ "ffn_dim": 768,
63
+ "ffn_kernel_size": 3,
64
+ "final_dropout": 0.1,
65
+ "flow_size": 192,
66
+ "hidden_act": "relu",
67
  "hidden_dropout": 0.0,
68
+ "hidden_size": 192,
69
  "initializer_range": 0.02,
70
+ "intermediate_size": 3072,
71
  "layer_norm_eps": 1e-05,
72
  "layerdrop": 0.0,
73
+ "leaky_relu_slope": 0.1,
74
  "mask_feature_length": 10,
75
  "mask_feature_min_masks": 0,
76
  "mask_feature_prob": 0.0,
 
78
  "mask_time_min_masks": 2,
79
  "mask_time_prob": 0.05,
80
  "model_type": "wav2vec2",
81
+ "noise_scale": 0.667,
82
+ "noise_scale_duration": 0.8,
83
  "num_adapter_layers": 3,
84
+ "num_attention_heads": 2,
85
  "num_codevector_groups": 2,
86
  "num_codevectors_per_group": 320,
87
  "num_conv_pos_embedding_groups": 16,
88
  "num_conv_pos_embeddings": 128,
89
  "num_feat_extract_layers": 7,
90
+ "num_hidden_layers": 6,
91
  "num_negatives": 100,
92
+ "num_speakers": 1,
93
+ "output_hidden_size": 192,
94
  "pad_token_id": 39,
95
+ "posterior_encoder_num_wavenet_layers": 16,
96
+ "prior_encoder_num_flows": 4,
97
+ "prior_encoder_num_wavenet_layers": 4,
98
+ "proj_codevector_dim": 256,
99
+ "resblock_dilation_sizes": [
100
+ [
101
+ 1,
102
+ 3,
103
+ 5
104
+ ],
105
+ [
106
+ 1,
107
+ 3,
108
+ 5
109
+ ],
110
+ [
111
+ 1,
112
+ 3,
113
+ 5
114
+ ]
115
+ ],
116
+ "resblock_kernel_sizes": [
117
+ 3,
118
+ 7,
119
+ 11
120
+ ],
121
+ "sampling_rate": 16000,
122
+ "speaker_embedding_size": 0,
123
+ "speaking_rate": 1.0,
124
+ "spectrogram_bins": 513,
125
  "tdnn_dilation": [
126
  1,
127
  2,
 
145
  ],
146
  "torch_dtype": "float32",
147
  "transformers_version": "4.33.0",
148
+ "upsample_initial_channel": 512,
149
+ "upsample_kernel_sizes": [
150
+ 16,
151
+ 16,
152
+ 4,
153
+ 4
154
+ ],
155
+ "upsample_rates": [
156
+ 8,
157
+ 8,
158
+ 2,
159
+ 2
160
+ ],
161
+ "use_bias": true,
162
+ "use_stochastic_duration_prediction": true,
163
  "use_weighted_layer_sum": false,
164
  "vocab_size": 42,
165
+ "wavenet_dilation_rate": 1,
166
+ "wavenet_dropout": 0.0,
167
+ "wavenet_kernel_size": 5,
168
+ "window_size": 4,
169
  "xvector_output_dim": 512
170
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1353b7463bc0b05caa24de87dac233af90840d68312e5c420f112d40ca67adce
3
- size 3859190925
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:317edd91999dcf40c03783dde09dbb410e4d52b82a32e1489e557131d6a676ab
3
+ size 50421025
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:0929c7f20be76eff677fd3560f7c20f8fd68fc1efab7a616dc8ecb14f47fd590
3
  size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cf2466b3592147cc4d4a696658f59523a5155740c96e8bfba71b1c5812ba131f
3
  size 4091