dbaggi commited on
Commit
1f0c11b
1 Parent(s): 74a2a36

Training in progress, epoch 0

Browse files
Files changed (3) hide show
  1. config.json +12 -54
  2. pytorch_model.bin +2 -2
  3. training_args.bin +1 -1
config.json CHANGED
@@ -1,19 +1,13 @@
1
  {
2
- "_name_or_path": "facebook/wav2vec2-base",
3
- "activation_dropout": 0.0,
4
- "adapter_attn_dim": null,
5
- "adapter_kernel_size": 3,
6
- "adapter_stride": 2,
7
- "add_adapter": false,
8
  "apply_spec_augment": true,
9
  "architectures": [
10
- "Wav2Vec2ForSequenceClassification"
11
  ],
12
  "attention_dropout": 0.1,
13
  "bos_token_id": 1,
14
  "classifier_proj_size": 256,
15
- "codevector_dim": 256,
16
- "contrastive_logits_temperature": 0.1,
17
  "conv_bias": false,
18
  "conv_dim": [
19
  512,
@@ -44,17 +38,18 @@
44
  ],
45
  "ctc_loss_reduction": "sum",
46
  "ctc_zero_infinity": false,
47
- "diversity_loss_weight": 0.1,
48
  "do_stable_layer_norm": false,
49
  "eos_token_id": 2,
50
  "feat_extract_activation": "gelu",
 
51
  "feat_extract_norm": "group",
52
  "feat_proj_dropout": 0.1,
53
- "feat_quantizer_dropout": 0.0,
54
- "final_dropout": 0.0,
55
- "freeze_feat_extract_train": true,
56
  "hidden_act": "gelu",
57
  "hidden_dropout": 0.1,
 
58
  "hidden_size": 768,
59
  "id2label": {
60
  "0": "abroad",
@@ -91,60 +86,23 @@
91
  "pay_bill": "13"
92
  },
93
  "layer_norm_eps": 1e-05,
94
- "layerdrop": 0.0,
95
- "mask_channel_length": 10,
96
- "mask_channel_min_space": 1,
97
- "mask_channel_other": 0.0,
98
- "mask_channel_prob": 0.0,
99
- "mask_channel_selection": "static",
100
  "mask_feature_length": 10,
101
  "mask_feature_min_masks": 0,
102
  "mask_feature_prob": 0.0,
103
  "mask_time_length": 10,
104
  "mask_time_min_masks": 2,
105
- "mask_time_min_space": 1,
106
- "mask_time_other": 0.0,
107
  "mask_time_prob": 0.05,
108
- "mask_time_selection": "static",
109
- "model_type": "wav2vec2",
110
- "no_mask_channel_overlap": false,
111
- "no_mask_time_overlap": false,
112
- "num_adapter_layers": 3,
113
  "num_attention_heads": 12,
114
- "num_codevector_groups": 2,
115
- "num_codevectors_per_group": 320,
116
  "num_conv_pos_embedding_groups": 16,
117
  "num_conv_pos_embeddings": 128,
118
  "num_feat_extract_layers": 7,
119
  "num_hidden_layers": 12,
120
- "num_negatives": 100,
121
- "output_hidden_size": 768,
122
  "pad_token_id": 0,
123
- "proj_codevector_dim": 256,
124
- "tdnn_dilation": [
125
- 1,
126
- 2,
127
- 3,
128
- 1,
129
- 1
130
- ],
131
- "tdnn_dim": [
132
- 512,
133
- 512,
134
- 512,
135
- 512,
136
- 1500
137
- ],
138
- "tdnn_kernel": [
139
- 5,
140
- 3,
141
- 3,
142
- 1,
143
- 1
144
- ],
145
  "torch_dtype": "float32",
146
  "transformers_version": "4.33.3",
147
  "use_weighted_layer_sum": false,
148
- "vocab_size": 32,
149
- "xvector_output_dim": 512
150
  }
 
1
  {
2
+ "_name_or_path": "facebook/hubert-base-ls960",
3
+ "activation_dropout": 0.1,
 
 
 
 
4
  "apply_spec_augment": true,
5
  "architectures": [
6
+ "HubertForSequenceClassification"
7
  ],
8
  "attention_dropout": 0.1,
9
  "bos_token_id": 1,
10
  "classifier_proj_size": 256,
 
 
11
  "conv_bias": false,
12
  "conv_dim": [
13
  512,
 
38
  ],
39
  "ctc_loss_reduction": "sum",
40
  "ctc_zero_infinity": false,
 
41
  "do_stable_layer_norm": false,
42
  "eos_token_id": 2,
43
  "feat_extract_activation": "gelu",
44
+ "feat_extract_dropout": 0.0,
45
  "feat_extract_norm": "group",
46
  "feat_proj_dropout": 0.1,
47
+ "feat_proj_layer_norm": true,
48
+ "final_dropout": 0.1,
49
+ "gradient_checkpointing": false,
50
  "hidden_act": "gelu",
51
  "hidden_dropout": 0.1,
52
+ "hidden_dropout_prob": 0.1,
53
  "hidden_size": 768,
54
  "id2label": {
55
  "0": "abroad",
 
86
  "pay_bill": "13"
87
  },
88
  "layer_norm_eps": 1e-05,
89
+ "layerdrop": 0.1,
 
 
 
 
 
90
  "mask_feature_length": 10,
91
  "mask_feature_min_masks": 0,
92
  "mask_feature_prob": 0.0,
93
  "mask_time_length": 10,
94
  "mask_time_min_masks": 2,
 
 
95
  "mask_time_prob": 0.05,
96
+ "model_type": "hubert",
 
 
 
 
97
  "num_attention_heads": 12,
 
 
98
  "num_conv_pos_embedding_groups": 16,
99
  "num_conv_pos_embeddings": 128,
100
  "num_feat_extract_layers": 7,
101
  "num_hidden_layers": 12,
 
 
102
  "pad_token_id": 0,
103
+ "tokenizer_class": "Wav2Vec2CTCTokenizer",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
104
  "torch_dtype": "float32",
105
  "transformers_version": "4.33.3",
106
  "use_weighted_layer_sum": false,
107
+ "vocab_size": 32
 
108
  }
pytorch_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:65ed7245100c67ccec41c327c35c74f2f291401c465a14df153f28af8dff4414
3
- size 378362729
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7e85bc10b1f4b7feb0cb3155894377f4aabae3db7068c8463a3d56d3f222d296
3
+ size 378362345
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:c7cf62ab5eb433d7f8155c3af82bbcddd92ae0036f4f59ea2786720a31dfcef3
3
  size 4091
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:342901876858fa8da2f2a8890310b168f76096c46f6db020e8e4dd392b96374a
3
  size 4091