kingabzpro commited on
Commit
81d3d72
1 Parent(s): 31d25b6

Training in progress, step 100

Browse files
config.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
- "_name_or_path": "manandey/wav2vec2-large-xlsr-punjabi",
3
- "activation_dropout": 0.055,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
6
  "add_adapter": false,
@@ -9,11 +9,12 @@
9
  "Wav2Vec2ForCTC"
10
  ],
11
  "attention_dropout": 0.1,
12
- "bos_token_id": 1,
 
13
  "classifier_proj_size": 256,
14
  "codevector_dim": 256,
15
  "contrastive_logits_temperature": 0.1,
16
- "conv_bias": true,
17
  "conv_dim": [
18
  512,
19
  512,
@@ -44,47 +45,42 @@
44
  "ctc_loss_reduction": "mean",
45
  "ctc_zero_infinity": false,
46
  "diversity_loss_weight": 0.1,
47
- "do_stable_layer_norm": true,
 
 
48
  "eos_token_id": 2,
49
  "feat_extract_activation": "gelu",
50
- "feat_extract_dropout": 0.0,
51
- "feat_extract_norm": "layer",
52
  "feat_proj_dropout": 0.0,
53
  "feat_quantizer_dropout": 0.0,
54
- "final_dropout": 0.0,
 
55
  "hidden_act": "gelu",
56
  "hidden_dropout": 0.1,
57
- "hidden_size": 1024,
58
  "initializer_range": 0.02,
59
- "intermediate_size": 4096,
60
  "layer_norm_eps": 1e-05,
61
  "layerdrop": 0.1,
62
- "mask_channel_length": 10,
63
- "mask_channel_min_space": 1,
64
- "mask_channel_other": 0.0,
65
- "mask_channel_prob": 0.0,
66
- "mask_channel_selection": "static",
67
  "mask_feature_length": 10,
68
  "mask_feature_min_masks": 0,
69
  "mask_feature_prob": 0.0,
70
  "mask_time_length": 10,
71
  "mask_time_min_masks": 2,
72
- "mask_time_min_space": 1,
73
- "mask_time_other": 0.0,
74
  "mask_time_prob": 0.05,
75
- "mask_time_selection": "static",
76
  "model_type": "wav2vec2",
77
  "num_adapter_layers": 3,
78
- "num_attention_heads": 16,
79
  "num_codevector_groups": 2,
80
  "num_codevectors_per_group": 320,
81
  "num_conv_pos_embedding_groups": 16,
82
  "num_conv_pos_embeddings": 128,
83
  "num_feat_extract_layers": 7,
84
- "num_hidden_layers": 24,
85
  "num_negatives": 100,
86
- "output_hidden_size": 1024,
87
- "pad_token_id": 62,
 
88
  "proj_codevector_dim": 256,
89
  "tdnn_dilation": [
90
  1,
@@ -109,7 +105,9 @@
109
  ],
110
  "torch_dtype": "float32",
111
  "transformers_version": "4.17.0.dev0",
 
112
  "use_weighted_layer_sum": false,
113
- "vocab_size": 63,
 
114
  "xvector_output_dim": 512
115
  }
 
1
  {
2
+ "_name_or_path": "Harveenchadha/vakyansh-wav2vec2-punjabi-pam-10",
3
+ "activation_dropout": 0.1,
4
  "adapter_kernel_size": 3,
5
  "adapter_stride": 2,
6
  "add_adapter": false,
 
9
  "Wav2Vec2ForCTC"
10
  ],
11
  "attention_dropout": 0.1,
12
+ "bos_token": "<s>",
13
+ "bos_token_id": 0,
14
  "classifier_proj_size": 256,
15
  "codevector_dim": 256,
16
  "contrastive_logits_temperature": 0.1,
17
+ "conv_bias": false,
18
  "conv_dim": [
19
  512,
20
  512,
 
45
  "ctc_loss_reduction": "mean",
46
  "ctc_zero_infinity": false,
47
  "diversity_loss_weight": 0.1,
48
+ "do_lower_case": false,
49
+ "do_stable_layer_norm": false,
50
+ "eos_token": "</s>",
51
  "eos_token_id": 2,
52
  "feat_extract_activation": "gelu",
53
+ "feat_extract_norm": "group",
 
54
  "feat_proj_dropout": 0.0,
55
  "feat_quantizer_dropout": 0.0,
56
+ "final_dropout": 0.1,
57
+ "gradient_checkpointing": false,
58
  "hidden_act": "gelu",
59
  "hidden_dropout": 0.1,
60
+ "hidden_size": 768,
61
  "initializer_range": 0.02,
62
+ "intermediate_size": 3072,
63
  "layer_norm_eps": 1e-05,
64
  "layerdrop": 0.1,
 
 
 
 
 
65
  "mask_feature_length": 10,
66
  "mask_feature_min_masks": 0,
67
  "mask_feature_prob": 0.0,
68
  "mask_time_length": 10,
69
  "mask_time_min_masks": 2,
 
 
70
  "mask_time_prob": 0.05,
 
71
  "model_type": "wav2vec2",
72
  "num_adapter_layers": 3,
73
+ "num_attention_heads": 12,
74
  "num_codevector_groups": 2,
75
  "num_codevectors_per_group": 320,
76
  "num_conv_pos_embedding_groups": 16,
77
  "num_conv_pos_embeddings": 128,
78
  "num_feat_extract_layers": 7,
79
+ "num_hidden_layers": 12,
80
  "num_negatives": 100,
81
+ "output_hidden_size": 768,
82
+ "pad_token": "[PAD]",
83
+ "pad_token_id": 1,
84
  "proj_codevector_dim": 256,
85
  "tdnn_dilation": [
86
  1,
 
105
  ],
106
  "torch_dtype": "float32",
107
  "transformers_version": "4.17.0.dev0",
108
+ "unk_token": "[UNK]",
109
  "use_weighted_layer_sum": false,
110
+ "vocab_size": 65,
111
+ "word_delimiter_token": "|",
112
  "xvector_output_dim": 512
113
  }
preprocessor_config.json CHANGED
@@ -3,7 +3,7 @@
3
  "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
  "feature_size": 1,
5
  "padding_side": "right",
6
- "padding_value": 0.0,
7
- "return_attention_mask": true,
8
  "sampling_rate": 16000
9
  }
 
3
  "feature_extractor_type": "Wav2Vec2FeatureExtractor",
4
  "feature_size": 1,
5
  "padding_side": "right",
6
+ "padding_value": 0,
7
+ "return_attention_mask": false,
8
  "sampling_rate": 16000
9
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:3e32a930f12c587f75cc2c1339df46b9e95f8b6591df4a2b83cefd5cc1b03185
3
- size 1262181937
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:27afaf324bc1b55344fde57ba5395291d6aa7973f7597f6e152cd5b904410e57
3
+ size 377771543
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:1633279170073f611c95b71ad76e953a8f57c1d8684a88916db343257bebd0e1
3
  size 3055
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f715df5ad229b667f9b088bdd89153372f7e5e32465198659a8d4d947b014ab8
3
  size 3055