Add JPQD model
Browse files- README.md +70 -0
- all_results.json +12 -0
- compressed_graph.dot +0 -0
- config.json +146 -0
- eval_results.json +8 -0
- nncf_output.log +0 -0
- openvino_config.json +82 -0
- openvino_model.bin +3 -0
- openvino_model.xml +0 -0
- original_graph.dot +0 -0
- preprocessor_config.json +9 -0
- structured_sparsity.csv +73 -0
- train_results.json +7 -0
- trainer_state.json +0 -0
- training_args.bin +3 -0
README.md
ADDED
@@ -0,0 +1,70 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
tags:
|
4 |
+
- audio-classification
|
5 |
+
- generated_from_trainer
|
6 |
+
datasets:
|
7 |
+
- superb
|
8 |
+
metrics:
|
9 |
+
- accuracy
|
10 |
+
model-index:
|
11 |
+
- name: wav2vec2-base-superb-ks-jpqd-ov
|
12 |
+
results: []
|
13 |
+
---
|
14 |
+
|
15 |
+
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
16 |
+
should probably proofread and complete it, then remove this comment. -->
|
17 |
+
|
18 |
+
# wav2vec2-base-superb-ks-jpqd-ov
|
19 |
+
|
20 |
+
This model is a fine-tuned version of [facebook/wav2vec2-base](https://huggingface.co/facebook/wav2vec2-base) on the superb dataset.
|
21 |
+
|
22 |
+
It was compressed using [NNCF](https://github.com/openvinotoolkit/nncf) with [Optimum Intel](https://github.com/huggingface/optimum-intel#openvino) following the
|
23 |
+
JPQD image classification example.
|
24 |
+
|
25 |
+
It achieves the following results on the evaluation set:
|
26 |
+
- Loss: 0.0996
|
27 |
+
- Accuracy: 0.9734
|
28 |
+
|
29 |
+
### Training hyperparameters
|
30 |
+
|
31 |
+
The following hyperparameters were used during training:
|
32 |
+
- learning_rate: 0.0002
|
33 |
+
- train_batch_size: 32
|
34 |
+
- eval_batch_size: 64
|
35 |
+
- seed: 0
|
36 |
+
- gradient_accumulation_steps: 4
|
37 |
+
- total_train_batch_size: 128
|
38 |
+
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
39 |
+
- lr_scheduler_type: linear
|
40 |
+
- lr_scheduler_warmup_ratio: 0.1
|
41 |
+
- num_epochs: 15.0
|
42 |
+
- mixed_precision_training: Native AMP
|
43 |
+
|
44 |
+
### Training results
|
45 |
+
|
46 |
+
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
47 |
+
|:-------------:|:-----:|:----:|:---------------:|:--------:|
|
48 |
+
| 2.5323 | 1.0 | 399 | 2.7354 | 0.6209 |
|
49 |
+
| 2.716 | 2.0 | 798 | 2.7349 | 0.6209 |
|
50 |
+
| 2.6712 | 3.0 | 1197 | 2.7327 | 0.6209 |
|
51 |
+
| 13.204 | 4.0 | 1596 | 13.0685 | 0.8610 |
|
52 |
+
| 19.3024 | 5.0 | 1995 | 19.1352 | 0.9335 |
|
53 |
+
| 21.9148 | 6.0 | 2394 | 21.6996 | 0.9547 |
|
54 |
+
| 22.3788 | 7.0 | 2793 | 22.1079 | 0.9620 |
|
55 |
+
| 21.8158 | 8.0 | 3192 | 21.6228 | 0.9634 |
|
56 |
+
| 0.3194 | 9.0 | 3591 | 0.1420 | 0.9657 |
|
57 |
+
| 0.253 | 10.0 | 3990 | 0.1238 | 0.9693 |
|
58 |
+
| 0.2921 | 11.0 | 4389 | 0.1159 | 0.9687 |
|
59 |
+
| 0.2229 | 12.0 | 4788 | 0.1096 | 0.9704 |
|
60 |
+
| 0.2492 | 13.0 | 5187 | 0.1028 | 0.9712 |
|
61 |
+
| 0.2658 | 14.0 | 5586 | 0.1026 | 0.9734 |
|
62 |
+
| 0.213 | 15.0 | 5985 | 0.0996 | 0.9734 |
|
63 |
+
|
64 |
+
|
65 |
+
### Framework versions
|
66 |
+
|
67 |
+
- Transformers 4.26.1
|
68 |
+
- Pytorch 1.13.1+cu117
|
69 |
+
- Datasets 2.8.0
|
70 |
+
- Tokenizers 0.13.2
|
all_results.json
ADDED
@@ -0,0 +1,12 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 15.0,
|
3 |
+
"eval_accuracy": 0.9733745219182113,
|
4 |
+
"eval_loss": 0.09959075599908829,
|
5 |
+
"eval_runtime": 29.3442,
|
6 |
+
"eval_samples_per_second": 231.664,
|
7 |
+
"eval_steps_per_second": 3.646,
|
8 |
+
"train_loss": 6.759114137568271,
|
9 |
+
"train_runtime": 10160.2494,
|
10 |
+
"train_samples_per_second": 75.432,
|
11 |
+
"train_steps_per_second": 0.589
|
12 |
+
}
|
compressed_graph.dot
ADDED
The diff for this file is too large to render.
See raw diff
|
|
config.json
ADDED
@@ -0,0 +1,146 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"_name_or_path": "facebook/wav2vec2-base",
|
3 |
+
"activation_dropout": 0.0,
|
4 |
+
"adapter_kernel_size": 3,
|
5 |
+
"adapter_stride": 2,
|
6 |
+
"add_adapter": false,
|
7 |
+
"apply_spec_augment": true,
|
8 |
+
"architectures": [
|
9 |
+
"NNCFNetwork"
|
10 |
+
],
|
11 |
+
"attention_dropout": 0.1,
|
12 |
+
"bos_token_id": 1,
|
13 |
+
"classifier_proj_size": 256,
|
14 |
+
"codevector_dim": 256,
|
15 |
+
"contrastive_logits_temperature": 0.1,
|
16 |
+
"conv_bias": false,
|
17 |
+
"conv_dim": [
|
18 |
+
512,
|
19 |
+
512,
|
20 |
+
512,
|
21 |
+
512,
|
22 |
+
512,
|
23 |
+
512,
|
24 |
+
512
|
25 |
+
],
|
26 |
+
"conv_kernel": [
|
27 |
+
10,
|
28 |
+
3,
|
29 |
+
3,
|
30 |
+
3,
|
31 |
+
3,
|
32 |
+
2,
|
33 |
+
2
|
34 |
+
],
|
35 |
+
"conv_stride": [
|
36 |
+
5,
|
37 |
+
2,
|
38 |
+
2,
|
39 |
+
2,
|
40 |
+
2,
|
41 |
+
2,
|
42 |
+
2
|
43 |
+
],
|
44 |
+
"ctc_loss_reduction": "sum",
|
45 |
+
"ctc_zero_infinity": false,
|
46 |
+
"diversity_loss_weight": 0.1,
|
47 |
+
"do_stable_layer_norm": false,
|
48 |
+
"eos_token_id": 2,
|
49 |
+
"feat_extract_activation": "gelu",
|
50 |
+
"feat_extract_norm": "group",
|
51 |
+
"feat_proj_dropout": 0.1,
|
52 |
+
"feat_quantizer_dropout": 0.0,
|
53 |
+
"final_dropout": 0.0,
|
54 |
+
"finetuning_task": "audio-classification",
|
55 |
+
"freeze_feat_extract_train": true,
|
56 |
+
"hidden_act": "gelu",
|
57 |
+
"hidden_dropout": 0.1,
|
58 |
+
"hidden_size": 768,
|
59 |
+
"id2label": {
|
60 |
+
"0": "yes",
|
61 |
+
"1": "no",
|
62 |
+
"10": "_silence_",
|
63 |
+
"11": "_unknown_",
|
64 |
+
"2": "up",
|
65 |
+
"3": "down",
|
66 |
+
"4": "left",
|
67 |
+
"5": "right",
|
68 |
+
"6": "on",
|
69 |
+
"7": "off",
|
70 |
+
"8": "stop",
|
71 |
+
"9": "go"
|
72 |
+
},
|
73 |
+
"initializer_range": 0.02,
|
74 |
+
"intermediate_size": 3072,
|
75 |
+
"label2id": {
|
76 |
+
"_silence_": "10",
|
77 |
+
"_unknown_": "11",
|
78 |
+
"down": "3",
|
79 |
+
"go": "9",
|
80 |
+
"left": "4",
|
81 |
+
"no": "1",
|
82 |
+
"off": "7",
|
83 |
+
"on": "6",
|
84 |
+
"right": "5",
|
85 |
+
"stop": "8",
|
86 |
+
"up": "2",
|
87 |
+
"yes": "0"
|
88 |
+
},
|
89 |
+
"layer_norm_eps": 1e-05,
|
90 |
+
"layerdrop": 0.0,
|
91 |
+
"mask_channel_length": 10,
|
92 |
+
"mask_channel_min_space": 1,
|
93 |
+
"mask_channel_other": 0.0,
|
94 |
+
"mask_channel_prob": 0.0,
|
95 |
+
"mask_channel_selection": "static",
|
96 |
+
"mask_feature_length": 10,
|
97 |
+
"mask_feature_min_masks": 0,
|
98 |
+
"mask_feature_prob": 0.0,
|
99 |
+
"mask_time_length": 10,
|
100 |
+
"mask_time_min_masks": 2,
|
101 |
+
"mask_time_min_space": 1,
|
102 |
+
"mask_time_other": 0.0,
|
103 |
+
"mask_time_prob": 0.05,
|
104 |
+
"mask_time_selection": "static",
|
105 |
+
"model_type": "wav2vec2",
|
106 |
+
"no_mask_channel_overlap": false,
|
107 |
+
"no_mask_time_overlap": false,
|
108 |
+
"num_adapter_layers": 3,
|
109 |
+
"num_attention_heads": 12,
|
110 |
+
"num_codevector_groups": 2,
|
111 |
+
"num_codevectors_per_group": 320,
|
112 |
+
"num_conv_pos_embedding_groups": 16,
|
113 |
+
"num_conv_pos_embeddings": 128,
|
114 |
+
"num_feat_extract_layers": 7,
|
115 |
+
"num_hidden_layers": 12,
|
116 |
+
"num_negatives": 100,
|
117 |
+
"output_hidden_size": 768,
|
118 |
+
"pad_token_id": 0,
|
119 |
+
"proj_codevector_dim": 256,
|
120 |
+
"tdnn_dilation": [
|
121 |
+
1,
|
122 |
+
2,
|
123 |
+
3,
|
124 |
+
1,
|
125 |
+
1
|
126 |
+
],
|
127 |
+
"tdnn_dim": [
|
128 |
+
512,
|
129 |
+
512,
|
130 |
+
512,
|
131 |
+
512,
|
132 |
+
1500
|
133 |
+
],
|
134 |
+
"tdnn_kernel": [
|
135 |
+
5,
|
136 |
+
3,
|
137 |
+
3,
|
138 |
+
1,
|
139 |
+
1
|
140 |
+
],
|
141 |
+
"torch_dtype": "float32",
|
142 |
+
"transformers_version": "4.26.1",
|
143 |
+
"use_weighted_layer_sum": false,
|
144 |
+
"vocab_size": 32,
|
145 |
+
"xvector_output_dim": 512
|
146 |
+
}
|
eval_results.json
ADDED
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 15.0,
|
3 |
+
"eval_accuracy": 0.9733745219182113,
|
4 |
+
"eval_loss": 0.09959075599908829,
|
5 |
+
"eval_runtime": 29.3442,
|
6 |
+
"eval_samples_per_second": 231.664,
|
7 |
+
"eval_steps_per_second": 3.646
|
8 |
+
}
|
nncf_output.log
ADDED
The diff for this file is too large to render.
See raw diff
|
|
openvino_config.json
ADDED
@@ -0,0 +1,82 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"compression": [
|
3 |
+
{
|
4 |
+
"algorithm": "movement_sparsity",
|
5 |
+
"ignored_scopes": [
|
6 |
+
"{re}projector",
|
7 |
+
"{re}classifier",
|
8 |
+
"{re}feature_extractor",
|
9 |
+
"{re}feature_projection",
|
10 |
+
"{re}pos_conv_embed"
|
11 |
+
],
|
12 |
+
"params": {
|
13 |
+
"enable_structured_masking": true,
|
14 |
+
"importance_regularization_factor": 0.1,
|
15 |
+
"warmup_end_epoch": 8,
|
16 |
+
"warmup_start_epoch": 3
|
17 |
+
},
|
18 |
+
"sparse_structure_by_scopes": [
|
19 |
+
{
|
20 |
+
"mode": "block",
|
21 |
+
"sparse_factors": [
|
22 |
+
32,
|
23 |
+
32
|
24 |
+
],
|
25 |
+
"target_scopes": "{re}.*Wav2Vec2Attention.*"
|
26 |
+
},
|
27 |
+
{
|
28 |
+
"axis": 0,
|
29 |
+
"mode": "per_dim",
|
30 |
+
"target_scopes": "{re}.*intermediate_dense.*"
|
31 |
+
},
|
32 |
+
{
|
33 |
+
"axis": 1,
|
34 |
+
"mode": "per_dim",
|
35 |
+
"target_scopes": "{re}.*output_dense.*"
|
36 |
+
}
|
37 |
+
]
|
38 |
+
},
|
39 |
+
{
|
40 |
+
"algorithm": "quantization",
|
41 |
+
"export_to_onnx_standard_ops": false,
|
42 |
+
"ignored_scopes": [
|
43 |
+
"{re}.*feature_extractor.*",
|
44 |
+
"{re}.*__add___[0-1]",
|
45 |
+
"{re}.*layer_norm_0"
|
46 |
+
],
|
47 |
+
"initializer": {
|
48 |
+
"batchnorm_adaptation": {
|
49 |
+
"num_bn_adaptation_samples": 0
|
50 |
+
},
|
51 |
+
"range": {
|
52 |
+
"num_init_samples": 300,
|
53 |
+
"type": "mean_min_max"
|
54 |
+
}
|
55 |
+
},
|
56 |
+
"overflow_fix": "enable",
|
57 |
+
"preset": "mixed",
|
58 |
+
"quantize_inputs": false,
|
59 |
+
"scope_overrides": {
|
60 |
+
"activations": {
|
61 |
+
"{re}.*matmul_0": {
|
62 |
+
"mode": "symmetric"
|
63 |
+
}
|
64 |
+
}
|
65 |
+
}
|
66 |
+
}
|
67 |
+
],
|
68 |
+
"input_info": [
|
69 |
+
{
|
70 |
+
"keyword": "input_values",
|
71 |
+
"sample_size": [
|
72 |
+
32,
|
73 |
+
16000
|
74 |
+
],
|
75 |
+
"type": "float"
|
76 |
+
}
|
77 |
+
],
|
78 |
+
"log_dir": "wav2vec2-base-superb-ks-jpqd-ov",
|
79 |
+
"optimum_version": "1.7.1",
|
80 |
+
"save_onnx_model": false,
|
81 |
+
"transformers_version": "4.26.1"
|
82 |
+
}
|
openvino_model.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42d9bc39ca11a88469e0a7a72206079d24e261002c3abeb80947347483d51c78
|
3 |
+
size 378966316
|
openvino_model.xml
ADDED
The diff for this file is too large to render.
See raw diff
|
|
original_graph.dot
ADDED
The diff for this file is too large to render.
See raw diff
|
|
preprocessor_config.json
ADDED
@@ -0,0 +1,9 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"do_normalize": true,
|
3 |
+
"feature_extractor_type": "Wav2Vec2FeatureExtractor",
|
4 |
+
"feature_size": 1,
|
5 |
+
"padding_side": "right",
|
6 |
+
"padding_value": 0.0,
|
7 |
+
"return_attention_mask": false,
|
8 |
+
"sampling_rate": 16000
|
9 |
+
}
|
structured_sparsity.csv
ADDED
@@ -0,0 +1,73 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
,group_id,type,torch_module,weight_shape,pruned_weight_shape,bias_shape,pruned_bias_shape,head_or_channel_id_to_keep,module_node_name
|
2 |
+
0,0,MHSA,nncf_module.wav2vec2.encoder.layers.0.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[4],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[0]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
3 |
+
1,0,MHSA,nncf_module.wav2vec2.encoder.layers.0.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[4],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[0]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
4 |
+
2,0,MHSA,nncf_module.wav2vec2.encoder.layers.0.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[4],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[0]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
5 |
+
3,0,MHSA,nncf_module.wav2vec2.encoder.layers.0.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[4],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[0]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
6 |
+
4,1,FF,nncf_module.wav2vec2.encoder.layers.0.feed_forward.intermediate_dense,"(3072, 768)","(381, 768)","(3072,)","(381,)",[381 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[0]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
7 |
+
5,1,FF,nncf_module.wav2vec2.encoder.layers.0.feed_forward.output_dense,"(768, 3072)","(768, 381)","(768,)","(768,)",[381 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[0]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
8 |
+
6,2,MHSA,nncf_module.wav2vec2.encoder.layers.1.attention.q_proj,"(768, 768)","(192, 768)","(768,)","(192,)","[5, 10, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[1]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
9 |
+
7,2,MHSA,nncf_module.wav2vec2.encoder.layers.1.attention.k_proj,"(768, 768)","(192, 768)","(768,)","(192,)","[5, 10, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[1]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
10 |
+
8,2,MHSA,nncf_module.wav2vec2.encoder.layers.1.attention.v_proj,"(768, 768)","(192, 768)","(768,)","(192,)","[5, 10, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[1]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
11 |
+
9,2,MHSA,nncf_module.wav2vec2.encoder.layers.1.attention.out_proj,"(768, 768)","(768, 192)","(768,)","(768,)","[5, 10, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[1]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
12 |
+
10,3,FF,nncf_module.wav2vec2.encoder.layers.1.feed_forward.intermediate_dense,"(3072, 768)","(147, 768)","(3072,)","(147,)",[147 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[1]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
13 |
+
11,3,FF,nncf_module.wav2vec2.encoder.layers.1.feed_forward.output_dense,"(768, 3072)","(768, 147)","(768,)","(768,)",[147 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[1]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
14 |
+
12,4,MHSA,nncf_module.wav2vec2.encoder.layers.2.attention.q_proj,"(768, 768)","(384, 768)","(768,)","(384,)","[1, 3, 5, 6, 9, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[2]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
15 |
+
13,4,MHSA,nncf_module.wav2vec2.encoder.layers.2.attention.k_proj,"(768, 768)","(384, 768)","(768,)","(384,)","[1, 3, 5, 6, 9, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[2]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
16 |
+
14,4,MHSA,nncf_module.wav2vec2.encoder.layers.2.attention.v_proj,"(768, 768)","(384, 768)","(768,)","(384,)","[1, 3, 5, 6, 9, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[2]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
17 |
+
15,4,MHSA,nncf_module.wav2vec2.encoder.layers.2.attention.out_proj,"(768, 768)","(768, 384)","(768,)","(768,)","[1, 3, 5, 6, 9, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[2]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
18 |
+
16,5,FF,nncf_module.wav2vec2.encoder.layers.2.feed_forward.intermediate_dense,"(3072, 768)","(97, 768)","(3072,)","(97,)",[97 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[2]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
19 |
+
17,5,FF,nncf_module.wav2vec2.encoder.layers.2.feed_forward.output_dense,"(768, 3072)","(768, 97)","(768,)","(768,)",[97 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[2]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
20 |
+
18,6,MHSA,nncf_module.wav2vec2.encoder.layers.3.attention.q_proj,"(768, 768)","(128, 768)","(768,)","(128,)","[5, 8]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[3]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
21 |
+
19,6,MHSA,nncf_module.wav2vec2.encoder.layers.3.attention.k_proj,"(768, 768)","(128, 768)","(768,)","(128,)","[5, 8]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[3]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
22 |
+
20,6,MHSA,nncf_module.wav2vec2.encoder.layers.3.attention.v_proj,"(768, 768)","(128, 768)","(768,)","(128,)","[5, 8]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[3]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
23 |
+
21,6,MHSA,nncf_module.wav2vec2.encoder.layers.3.attention.out_proj,"(768, 768)","(768, 128)","(768,)","(768,)","[5, 8]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[3]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
24 |
+
22,7,FF,nncf_module.wav2vec2.encoder.layers.3.feed_forward.intermediate_dense,"(3072, 768)","(82, 768)","(3072,)","(82,)",[82 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[3]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
25 |
+
23,7,FF,nncf_module.wav2vec2.encoder.layers.3.feed_forward.output_dense,"(768, 3072)","(768, 82)","(768,)","(768,)",[82 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[3]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
26 |
+
24,8,MHSA,nncf_module.wav2vec2.encoder.layers.4.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[4]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
27 |
+
25,8,MHSA,nncf_module.wav2vec2.encoder.layers.4.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[4]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
28 |
+
26,8,MHSA,nncf_module.wav2vec2.encoder.layers.4.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[4]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
29 |
+
27,8,MHSA,nncf_module.wav2vec2.encoder.layers.4.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[4]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
30 |
+
28,9,FF,nncf_module.wav2vec2.encoder.layers.4.feed_forward.intermediate_dense,"(3072, 768)","(78, 768)","(3072,)","(78,)",[78 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[4]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
31 |
+
29,9,FF,nncf_module.wav2vec2.encoder.layers.4.feed_forward.output_dense,"(768, 3072)","(768, 78)","(768,)","(768,)",[78 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[4]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
32 |
+
30,10,MHSA,nncf_module.wav2vec2.encoder.layers.5.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[1],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[5]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
33 |
+
31,10,MHSA,nncf_module.wav2vec2.encoder.layers.5.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[1],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[5]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
34 |
+
32,10,MHSA,nncf_module.wav2vec2.encoder.layers.5.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[1],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[5]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
35 |
+
33,10,MHSA,nncf_module.wav2vec2.encoder.layers.5.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[1],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[5]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
36 |
+
34,11,FF,nncf_module.wav2vec2.encoder.layers.5.feed_forward.intermediate_dense,"(3072, 768)","(74, 768)","(3072,)","(74,)",[74 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[5]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
37 |
+
35,11,FF,nncf_module.wav2vec2.encoder.layers.5.feed_forward.output_dense,"(768, 3072)","(768, 74)","(768,)","(768,)",[74 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[5]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
38 |
+
36,12,MHSA,nncf_module.wav2vec2.encoder.layers.6.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[6]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
39 |
+
37,12,MHSA,nncf_module.wav2vec2.encoder.layers.6.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[6]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
40 |
+
38,12,MHSA,nncf_module.wav2vec2.encoder.layers.6.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[6]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
41 |
+
39,12,MHSA,nncf_module.wav2vec2.encoder.layers.6.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[6]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
42 |
+
40,13,FF,nncf_module.wav2vec2.encoder.layers.6.feed_forward.intermediate_dense,"(3072, 768)","(77, 768)","(3072,)","(77,)",[77 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[6]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
43 |
+
41,13,FF,nncf_module.wav2vec2.encoder.layers.6.feed_forward.output_dense,"(768, 3072)","(768, 77)","(768,)","(768,)",[77 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[6]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
44 |
+
42,14,MHSA,nncf_module.wav2vec2.encoder.layers.7.attention.q_proj,"(768, 768)","(576, 768)","(768,)","(576,)","[0, 1, 2, 3, 4, 6, 7, 8, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[7]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
45 |
+
43,14,MHSA,nncf_module.wav2vec2.encoder.layers.7.attention.k_proj,"(768, 768)","(576, 768)","(768,)","(576,)","[0, 1, 2, 3, 4, 6, 7, 8, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[7]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
46 |
+
44,14,MHSA,nncf_module.wav2vec2.encoder.layers.7.attention.v_proj,"(768, 768)","(576, 768)","(768,)","(576,)","[0, 1, 2, 3, 4, 6, 7, 8, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[7]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
47 |
+
45,14,MHSA,nncf_module.wav2vec2.encoder.layers.7.attention.out_proj,"(768, 768)","(768, 576)","(768,)","(768,)","[0, 1, 2, 3, 4, 6, 7, 8, 10]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[7]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
48 |
+
46,15,FF,nncf_module.wav2vec2.encoder.layers.7.feed_forward.intermediate_dense,"(3072, 768)","(72, 768)","(3072,)","(72,)",[72 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[7]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
49 |
+
47,15,FF,nncf_module.wav2vec2.encoder.layers.7.feed_forward.output_dense,"(768, 3072)","(768, 72)","(768,)","(768,)",[72 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[7]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
50 |
+
48,16,MHSA,nncf_module.wav2vec2.encoder.layers.8.attention.q_proj,"(768, 768)","(512, 768)","(768,)","(512,)","[1, 2, 3, 5, 7, 8, 9, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[8]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
51 |
+
49,16,MHSA,nncf_module.wav2vec2.encoder.layers.8.attention.k_proj,"(768, 768)","(512, 768)","(768,)","(512,)","[1, 2, 3, 5, 7, 8, 9, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[8]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
52 |
+
50,16,MHSA,nncf_module.wav2vec2.encoder.layers.8.attention.v_proj,"(768, 768)","(512, 768)","(768,)","(512,)","[1, 2, 3, 5, 7, 8, 9, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[8]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
53 |
+
51,16,MHSA,nncf_module.wav2vec2.encoder.layers.8.attention.out_proj,"(768, 768)","(768, 512)","(768,)","(768,)","[1, 2, 3, 5, 7, 8, 9, 11]",Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[8]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
54 |
+
52,17,FF,nncf_module.wav2vec2.encoder.layers.8.feed_forward.intermediate_dense,"(3072, 768)","(99, 768)","(3072,)","(99,)",[99 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[8]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
55 |
+
53,17,FF,nncf_module.wav2vec2.encoder.layers.8.feed_forward.output_dense,"(768, 3072)","(768, 99)","(768,)","(768,)",[99 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[8]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
56 |
+
54,18,MHSA,nncf_module.wav2vec2.encoder.layers.9.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[9]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
57 |
+
55,18,MHSA,nncf_module.wav2vec2.encoder.layers.9.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[9]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
58 |
+
56,18,MHSA,nncf_module.wav2vec2.encoder.layers.9.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[9]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
59 |
+
57,18,MHSA,nncf_module.wav2vec2.encoder.layers.9.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[9],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[9]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
60 |
+
58,19,FF,nncf_module.wav2vec2.encoder.layers.9.feed_forward.intermediate_dense,"(3072, 768)","(299, 768)","(3072,)","(299,)",[299 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[9]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
61 |
+
59,19,FF,nncf_module.wav2vec2.encoder.layers.9.feed_forward.output_dense,"(768, 3072)","(768, 299)","(768,)","(768,)",[299 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[9]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
62 |
+
60,20,MHSA,nncf_module.wav2vec2.encoder.layers.10.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[10]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
63 |
+
61,20,MHSA,nncf_module.wav2vec2.encoder.layers.10.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[10]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
64 |
+
62,20,MHSA,nncf_module.wav2vec2.encoder.layers.10.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[10]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
65 |
+
63,20,MHSA,nncf_module.wav2vec2.encoder.layers.10.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[3],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[10]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
66 |
+
64,21,FF,nncf_module.wav2vec2.encoder.layers.10.feed_forward.intermediate_dense,"(3072, 768)","(847, 768)","(3072,)","(847,)",[847 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[10]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
67 |
+
65,21,FF,nncf_module.wav2vec2.encoder.layers.10.feed_forward.output_dense,"(768, 3072)","(768, 847)","(768,)","(768,)",[847 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[10]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
68 |
+
66,22,MHSA,nncf_module.wav2vec2.encoder.layers.11.attention.q_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[6],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[11]/Wav2Vec2Attention[attention]/NNCFLinear[q_proj]/linear_0
|
69 |
+
67,22,MHSA,nncf_module.wav2vec2.encoder.layers.11.attention.k_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[6],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[11]/Wav2Vec2Attention[attention]/NNCFLinear[k_proj]/linear_0
|
70 |
+
68,22,MHSA,nncf_module.wav2vec2.encoder.layers.11.attention.v_proj,"(768, 768)","(64, 768)","(768,)","(64,)",[6],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[11]/Wav2Vec2Attention[attention]/NNCFLinear[v_proj]/linear_0
|
71 |
+
69,22,MHSA,nncf_module.wav2vec2.encoder.layers.11.attention.out_proj,"(768, 768)","(768, 64)","(768,)","(768,)",[6],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[11]/Wav2Vec2Attention[attention]/NNCFLinear[out_proj]/linear_0
|
72 |
+
70,23,FF,nncf_module.wav2vec2.encoder.layers.11.feed_forward.intermediate_dense,"(3072, 768)","(671, 768)","(3072,)","(671,)",[671 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[11]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[intermediate_dense]/linear_0
|
73 |
+
71,23,FF,nncf_module.wav2vec2.encoder.layers.11.feed_forward.output_dense,"(768, 3072)","(768, 671)","(768,)","(768,)",[671 items],Wav2Vec2ForSequenceClassification/Wav2Vec2Model[wav2vec2]/Wav2Vec2Encoder[encoder]/ModuleList[layers]/Wav2Vec2EncoderLayer[11]/Wav2Vec2FeedForward[feed_forward]/NNCFLinear[output_dense]/linear_0
|
train_results.json
ADDED
@@ -0,0 +1,7 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"epoch": 15.0,
|
3 |
+
"train_loss": 6.759114137568271,
|
4 |
+
"train_runtime": 10160.2494,
|
5 |
+
"train_samples_per_second": 75.432,
|
6 |
+
"train_steps_per_second": 0.589
|
7 |
+
}
|
trainer_state.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
training_args.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:87c5cf768f8f1c3cc470cb2683153a107f1d062da69d318f8cf30cb38f94441f
|
3 |
+
size 3579
|