End of training
Browse files- README.md +5 -11
- compressed_graph.dot +0 -0
- logs/events.out.tfevents.1700312106.0a848c80699a.1576.0 +3 -0
- logs/events.out.tfevents.1700312559.0a848c80699a.1576.1 +3 -0
- nncf_output.log +188 -0
- openvino_config.json +60 -0
- openvino_model.bin +3 -0
- openvino_model.xml +0 -0
- original_graph.dot +0 -0
- pytorch_model.bin +2 -2
- training_args.bin +1 -1
README.md
CHANGED
|
@@ -22,7 +22,7 @@ model-index:
|
|
| 22 |
metrics:
|
| 23 |
- name: Accuracy
|
| 24 |
type: accuracy
|
| 25 |
-
value: 0.
|
| 26 |
---
|
| 27 |
|
| 28 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
@@ -32,8 +32,8 @@ should probably proofread and complete it, then remove this comment. -->
|
|
| 32 |
|
| 33 |
This model is a fine-tuned version of [google/bert_uncased_L-6_H-768_A-12](https://huggingface.co/google/bert_uncased_L-6_H-768_A-12) on the glue dataset.
|
| 34 |
It achieves the following results on the evaluation set:
|
| 35 |
-
- Loss: 0.
|
| 36 |
-
- Accuracy: 0.
|
| 37 |
|
| 38 |
## Model description
|
| 39 |
|
|
@@ -58,20 +58,14 @@ The following hyperparameters were used during training:
|
|
| 58 |
- seed: 33
|
| 59 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 60 |
- lr_scheduler_type: linear
|
| 61 |
-
- num_epochs:
|
| 62 |
- mixed_precision_training: Native AMP
|
| 63 |
|
| 64 |
### Training results
|
| 65 |
|
| 66 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
| 67 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
|
| 68 |
-
| 0.
|
| 69 |
-
| 0.1196 | 2.0 | 1054 | 0.2980 | 0.9002 |
|
| 70 |
-
| 0.0789 | 3.0 | 1581 | 0.2825 | 0.9071 |
|
| 71 |
-
| 0.0529 | 4.0 | 2108 | 0.3194 | 0.9071 |
|
| 72 |
-
| 0.0364 | 5.0 | 2635 | 0.3683 | 0.9151 |
|
| 73 |
-
| 0.0236 | 6.0 | 3162 | 0.4103 | 0.9094 |
|
| 74 |
-
| 0.0154 | 7.0 | 3689 | 0.4751 | 0.9083 |
|
| 75 |
|
| 76 |
|
| 77 |
### Framework versions
|
|
|
|
| 22 |
metrics:
|
| 23 |
- name: Accuracy
|
| 24 |
type: accuracy
|
| 25 |
+
value: 0.908256880733945
|
| 26 |
---
|
| 27 |
|
| 28 |
<!-- This model card has been generated automatically according to the information the Trainer had access to. You
|
|
|
|
| 32 |
|
| 33 |
This model is a fine-tuned version of [google/bert_uncased_L-6_H-768_A-12](https://huggingface.co/google/bert_uncased_L-6_H-768_A-12) on the glue dataset.
|
| 34 |
It achieves the following results on the evaluation set:
|
| 35 |
+
- Loss: 0.2574
|
| 36 |
+
- Accuracy: 0.9083
|
| 37 |
|
| 38 |
## Model description
|
| 39 |
|
|
|
|
| 58 |
- seed: 33
|
| 59 |
- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
|
| 60 |
- lr_scheduler_type: linear
|
| 61 |
+
- num_epochs: 1.0
|
| 62 |
- mixed_precision_training: Native AMP
|
| 63 |
|
| 64 |
### Training results
|
| 65 |
|
| 66 |
| Training Loss | Epoch | Step | Validation Loss | Accuracy |
|
| 67 |
|:-------------:|:-----:|:----:|:---------------:|:--------:|
|
| 68 |
+
| 0.244 | 1.0 | 527 | 0.2574 | 0.9083 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 69 |
|
| 70 |
|
| 71 |
### Framework versions
|
compressed_graph.dot
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
logs/events.out.tfevents.1700312106.0a848c80699a.1576.0
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3c33d221ab0910be5d758fb04c7aebf629f244f1223c6b9112c2c2d89a6e89aa
|
| 3 |
+
size 5069
|
logs/events.out.tfevents.1700312559.0a848c80699a.1576.1
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:550e6d05743dca4c9e7b90c2cba8bda631811f82e98e62c4ea67e9d7cdc37b45
|
| 3 |
+
size 411
|
nncf_output.log
ADDED
|
@@ -0,0 +1,188 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
INFO:nncf:Not adding activation input quantizer for operation: 7 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[position_embeddings]/embedding_0
|
| 2 |
+
INFO:nncf:Not adding activation input quantizer for operation: 4 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[word_embeddings]/embedding_0
|
| 3 |
+
INFO:nncf:Not adding activation input quantizer for operation: 5 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[token_type_embeddings]/embedding_0
|
| 4 |
+
INFO:nncf:Not adding activation input quantizer for operation: 6 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/__add___0
|
| 5 |
+
INFO:nncf:Not adding activation input quantizer for operation: 8 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/__iadd___0
|
| 6 |
+
INFO:nncf:Not adding activation input quantizer for operation: 9 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 7 |
+
INFO:nncf:Not adding activation input quantizer for operation: 10 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/Dropout[dropout]/dropout_0
|
| 8 |
+
INFO:nncf:Not adding activation input quantizer for operation: 23 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 9 |
+
INFO:nncf:Not adding activation input quantizer for operation: 26 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 10 |
+
INFO:nncf:Not adding activation input quantizer for operation: 32 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 11 |
+
INFO:nncf:Not adding activation input quantizer for operation: 33 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 12 |
+
INFO:nncf:Not adding activation input quantizer for operation: 37 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertOutput[output]/__add___0
|
| 13 |
+
INFO:nncf:Not adding activation input quantizer for operation: 38 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 14 |
+
INFO:nncf:Not adding activation input quantizer for operation: 51 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 15 |
+
INFO:nncf:Not adding activation input quantizer for operation: 54 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 16 |
+
INFO:nncf:Not adding activation input quantizer for operation: 60 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 17 |
+
INFO:nncf:Not adding activation input quantizer for operation: 61 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 18 |
+
INFO:nncf:Not adding activation input quantizer for operation: 65 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertOutput[output]/__add___0
|
| 19 |
+
INFO:nncf:Not adding activation input quantizer for operation: 66 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 20 |
+
INFO:nncf:Not adding activation input quantizer for operation: 79 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 21 |
+
INFO:nncf:Not adding activation input quantizer for operation: 82 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 22 |
+
INFO:nncf:Not adding activation input quantizer for operation: 88 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 23 |
+
INFO:nncf:Not adding activation input quantizer for operation: 89 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 24 |
+
INFO:nncf:Not adding activation input quantizer for operation: 93 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertOutput[output]/__add___0
|
| 25 |
+
INFO:nncf:Not adding activation input quantizer for operation: 94 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 26 |
+
INFO:nncf:Not adding activation input quantizer for operation: 107 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 27 |
+
INFO:nncf:Not adding activation input quantizer for operation: 110 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 28 |
+
INFO:nncf:Not adding activation input quantizer for operation: 116 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 29 |
+
INFO:nncf:Not adding activation input quantizer for operation: 117 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 30 |
+
INFO:nncf:Not adding activation input quantizer for operation: 121 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertOutput[output]/__add___0
|
| 31 |
+
INFO:nncf:Not adding activation input quantizer for operation: 122 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 32 |
+
INFO:nncf:Not adding activation input quantizer for operation: 135 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 33 |
+
INFO:nncf:Not adding activation input quantizer for operation: 138 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 34 |
+
INFO:nncf:Not adding activation input quantizer for operation: 144 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 35 |
+
INFO:nncf:Not adding activation input quantizer for operation: 145 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 36 |
+
INFO:nncf:Not adding activation input quantizer for operation: 149 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertOutput[output]/__add___0
|
| 37 |
+
INFO:nncf:Not adding activation input quantizer for operation: 150 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 38 |
+
INFO:nncf:Not adding activation input quantizer for operation: 163 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 39 |
+
INFO:nncf:Not adding activation input quantizer for operation: 166 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 40 |
+
INFO:nncf:Not adding activation input quantizer for operation: 172 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 41 |
+
INFO:nncf:Not adding activation input quantizer for operation: 173 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 42 |
+
INFO:nncf:Not adding activation input quantizer for operation: 177 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertOutput[output]/__add___0
|
| 43 |
+
INFO:nncf:Not adding activation input quantizer for operation: 178 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 44 |
+
INFO:nncf:Collecting tensor statistics |█████ | 1 / 3
|
| 45 |
+
INFO:nncf:Collecting tensor statistics |██████████ | 2 / 3
|
| 46 |
+
INFO:nncf:Collecting tensor statistics |████████████████| 3 / 3
|
| 47 |
+
INFO:nncf:Compiling and loading torch extension: quantized_functions_cuda...
|
| 48 |
+
INFO:nncf:Not adding activation input quantizer for operation: 7 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[position_embeddings]/embedding_0
|
| 49 |
+
INFO:nncf:Not adding activation input quantizer for operation: 4 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[word_embeddings]/embedding_0
|
| 50 |
+
INFO:nncf:Not adding activation input quantizer for operation: 5 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFEmbedding[token_type_embeddings]/embedding_0
|
| 51 |
+
INFO:nncf:Not adding activation input quantizer for operation: 6 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/__add___0
|
| 52 |
+
INFO:nncf:Not adding activation input quantizer for operation: 8 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/__iadd___0
|
| 53 |
+
INFO:nncf:Not adding activation input quantizer for operation: 9 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 54 |
+
INFO:nncf:Not adding activation input quantizer for operation: 10 BertForSequenceClassification/BertModel[bert]/BertEmbeddings[embeddings]/Dropout[dropout]/dropout_0
|
| 55 |
+
INFO:nncf:Not adding activation input quantizer for operation: 23 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 56 |
+
INFO:nncf:Not adding activation input quantizer for operation: 26 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 57 |
+
INFO:nncf:Not adding activation input quantizer for operation: 32 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 58 |
+
INFO:nncf:Not adding activation input quantizer for operation: 33 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 59 |
+
INFO:nncf:Not adding activation input quantizer for operation: 37 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertOutput[output]/__add___0
|
| 60 |
+
INFO:nncf:Not adding activation input quantizer for operation: 38 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[0]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 61 |
+
INFO:nncf:Not adding activation input quantizer for operation: 51 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 62 |
+
INFO:nncf:Not adding activation input quantizer for operation: 54 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 63 |
+
INFO:nncf:Not adding activation input quantizer for operation: 60 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 64 |
+
INFO:nncf:Not adding activation input quantizer for operation: 61 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 65 |
+
INFO:nncf:Not adding activation input quantizer for operation: 65 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertOutput[output]/__add___0
|
| 66 |
+
INFO:nncf:Not adding activation input quantizer for operation: 66 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[1]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 67 |
+
INFO:nncf:Not adding activation input quantizer for operation: 79 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 68 |
+
INFO:nncf:Not adding activation input quantizer for operation: 82 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 69 |
+
INFO:nncf:Not adding activation input quantizer for operation: 88 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 70 |
+
INFO:nncf:Not adding activation input quantizer for operation: 89 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 71 |
+
INFO:nncf:Not adding activation input quantizer for operation: 93 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertOutput[output]/__add___0
|
| 72 |
+
INFO:nncf:Not adding activation input quantizer for operation: 94 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[2]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 73 |
+
INFO:nncf:Not adding activation input quantizer for operation: 107 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 74 |
+
INFO:nncf:Not adding activation input quantizer for operation: 110 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 75 |
+
INFO:nncf:Not adding activation input quantizer for operation: 116 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 76 |
+
INFO:nncf:Not adding activation input quantizer for operation: 117 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 77 |
+
INFO:nncf:Not adding activation input quantizer for operation: 121 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertOutput[output]/__add___0
|
| 78 |
+
INFO:nncf:Not adding activation input quantizer for operation: 122 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[3]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 79 |
+
INFO:nncf:Not adding activation input quantizer for operation: 135 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 80 |
+
INFO:nncf:Not adding activation input quantizer for operation: 138 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 81 |
+
INFO:nncf:Not adding activation input quantizer for operation: 144 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 82 |
+
INFO:nncf:Not adding activation input quantizer for operation: 145 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 83 |
+
INFO:nncf:Not adding activation input quantizer for operation: 149 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertOutput[output]/__add___0
|
| 84 |
+
INFO:nncf:Not adding activation input quantizer for operation: 150 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[4]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 85 |
+
INFO:nncf:Not adding activation input quantizer for operation: 163 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfAttention[self]/__add___0
|
| 86 |
+
INFO:nncf:Not adding activation input quantizer for operation: 166 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfAttention[self]/matmul_1
|
| 87 |
+
INFO:nncf:Not adding activation input quantizer for operation: 172 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfOutput[output]/__add___0
|
| 88 |
+
INFO:nncf:Not adding activation input quantizer for operation: 173 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertAttention[attention]/BertSelfOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 89 |
+
INFO:nncf:Not adding activation input quantizer for operation: 177 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertOutput[output]/__add___0
|
| 90 |
+
INFO:nncf:Not adding activation input quantizer for operation: 178 BertForSequenceClassification/BertModel[bert]/BertEncoder[encoder]/ModuleList[layer]/BertLayer[5]/BertOutput[output]/NNCFLayerNorm[LayerNorm]/layer_norm_0
|
| 91 |
+
INFO:nncf:Collecting tensor statistics |█████ | 1 / 3
|
| 92 |
+
INFO:nncf:Collecting tensor statistics |██████████ | 2 / 3
|
| 93 |
+
INFO:nncf:Collecting tensor statistics |████████████████| 3 / 3
|
| 94 |
+
INFO:nncf:Compiling and loading torch extension: quantized_functions_cuda...
|
| 95 |
+
INFO:nncf:Finished loading torch extension: quantized_functions_cuda
|
| 96 |
+
WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
|
| 97 |
+
NNCF relies on custom-wrapping the `forward` call in order to function properly.
|
| 98 |
+
Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
|
| 99 |
+
If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
|
| 100 |
+
model.nncf.set_original_unbound_forward(fn)
|
| 101 |
+
if `fn` has an unbound 0-th `self` argument, or
|
| 102 |
+
with model.nncf.temporary_bound_original_forward(fn): ...
|
| 103 |
+
if `fn` already had 0-th `self` argument bound or never had it in the first place.
|
| 104 |
+
WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
|
| 105 |
+
NNCF relies on custom-wrapping the `forward` call in order to function properly.
|
| 106 |
+
Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
|
| 107 |
+
If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
|
| 108 |
+
model.nncf.set_original_unbound_forward(fn)
|
| 109 |
+
if `fn` has an unbound 0-th `self` argument, or
|
| 110 |
+
with model.nncf.temporary_bound_original_forward(fn): ...
|
| 111 |
+
if `fn` already had 0-th `self` argument bound or never had it in the first place.
|
| 112 |
+
INFO:nncf:Statistics of the quantization algorithm:
|
| 113 |
+
Epoch 0 |+--------------------------------+-------+
|
| 114 |
+
Epoch 0 || Statistic's name | Value |
|
| 115 |
+
Epoch 0 |+================================+=======+
|
| 116 |
+
Epoch 0 || Ratio of enabled quantizations | 100 |
|
| 117 |
+
Epoch 0 |+--------------------------------+-------+
|
| 118 |
+
Epoch 0 |
|
| 119 |
+
Epoch 0 |Statistics of the quantization share:
|
| 120 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 121 |
+
Epoch 0 || Statistic's name | Value |
|
| 122 |
+
Epoch 0 |+==================================+====================+
|
| 123 |
+
Epoch 0 || Symmetric WQs / All placed WQs | 100.00 % (38 / 38) |
|
| 124 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 125 |
+
Epoch 0 || Asymmetric WQs / All placed WQs | 0.00 % (0 / 38) |
|
| 126 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 127 |
+
Epoch 0 || Signed WQs / All placed WQs | 100.00 % (38 / 38) |
|
| 128 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 129 |
+
Epoch 0 || Unsigned WQs / All placed WQs | 0.00 % (0 / 38) |
|
| 130 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 131 |
+
Epoch 0 || Per-tensor WQs / All placed WQs | 0.00 % (0 / 38) |
|
| 132 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 133 |
+
Epoch 0 || Per-channel WQs / All placed WQs | 100.00 % (38 / 38) |
|
| 134 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 135 |
+
Epoch 0 || Placed WQs / Potential WQs | 70.37 % (38 / 54) |
|
| 136 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 137 |
+
Epoch 0 || Symmetric AQs / All placed AQs | 27.27 % (12 / 44) |
|
| 138 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 139 |
+
Epoch 0 || Asymmetric AQs / All placed AQs | 72.73 % (32 / 44) |
|
| 140 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 141 |
+
Epoch 0 || Signed AQs / All placed AQs | 100.00 % (44 / 44) |
|
| 142 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 143 |
+
Epoch 0 || Unsigned AQs / All placed AQs | 0.00 % (0 / 44) |
|
| 144 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 145 |
+
Epoch 0 || Per-tensor AQs / All placed AQs | 100.00 % (44 / 44) |
|
| 146 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 147 |
+
Epoch 0 || Per-channel AQs / All placed AQs | 0.00 % (0 / 44) |
|
| 148 |
+
Epoch 0 |+----------------------------------+--------------------+
|
| 149 |
+
Epoch 0 |
|
| 150 |
+
Epoch 0 |Statistics of the bitwidth distribution:
|
| 151 |
+
Epoch 0 |+--------------+---------------------+--------------------+--------------------+
|
| 152 |
+
Epoch 0 || Num bits (N) | N-bits WQs / Placed | N-bits AQs / | N-bits Qs / Placed |
|
| 153 |
+
Epoch 0 || | WQs | Placed AQs | Qs |
|
| 154 |
+
Epoch 0 |+==============+=====================+====================+====================+
|
| 155 |
+
Epoch 0 || 8 | 100.00 % (38 / 38) | 100.00 % (44 / 44) | 100.00 % (82 / 82) |
|
| 156 |
+
Epoch 0 |+--------------+---------------------+--------------------+--------------------+
|
| 157 |
+
WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
|
| 158 |
+
NNCF relies on custom-wrapping the `forward` call in order to function properly.
|
| 159 |
+
Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
|
| 160 |
+
If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
|
| 161 |
+
model.nncf.set_original_unbound_forward(fn)
|
| 162 |
+
if `fn` has an unbound 0-th `self` argument, or
|
| 163 |
+
with model.nncf.temporary_bound_original_forward(fn): ...
|
| 164 |
+
if `fn` already had 0-th `self` argument bound or never had it in the first place.
|
| 165 |
+
WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
|
| 166 |
+
NNCF relies on custom-wrapping the `forward` call in order to function properly.
|
| 167 |
+
Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
|
| 168 |
+
If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
|
| 169 |
+
model.nncf.set_original_unbound_forward(fn)
|
| 170 |
+
if `fn` has an unbound 0-th `self` argument, or
|
| 171 |
+
with model.nncf.temporary_bound_original_forward(fn): ...
|
| 172 |
+
if `fn` already had 0-th `self` argument bound or never had it in the first place.
|
| 173 |
+
WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
|
| 174 |
+
NNCF relies on custom-wrapping the `forward` call in order to function properly.
|
| 175 |
+
Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
|
| 176 |
+
If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
|
| 177 |
+
model.nncf.set_original_unbound_forward(fn)
|
| 178 |
+
if `fn` has an unbound 0-th `self` argument, or
|
| 179 |
+
with model.nncf.temporary_bound_original_forward(fn): ...
|
| 180 |
+
if `fn` already had 0-th `self` argument bound or never had it in the first place.
|
| 181 |
+
WARNING:nncf:You are setting `forward` on an NNCF-processed model object.
|
| 182 |
+
NNCF relies on custom-wrapping the `forward` call in order to function properly.
|
| 183 |
+
Arbitrary adjustments to the forward function on an NNCFNetwork object have undefined behavior.
|
| 184 |
+
If you need to replace the underlying forward function of the original model so that NNCF should be using that instead of the original forward function that NNCF saved during the compressed model creation, you can do this by calling:
|
| 185 |
+
model.nncf.set_original_unbound_forward(fn)
|
| 186 |
+
if `fn` has an unbound 0-th `self` argument, or
|
| 187 |
+
with model.nncf.temporary_bound_original_forward(fn): ...
|
| 188 |
+
if `fn` already had 0-th `self` argument bound or never had it in the first place.
|
openvino_config.json
ADDED
|
@@ -0,0 +1,60 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{
|
| 2 |
+
"compression": {
|
| 3 |
+
"algorithm": "quantization",
|
| 4 |
+
"export_to_onnx_standard_ops": false,
|
| 5 |
+
"ignored_scopes": [
|
| 6 |
+
"{re}.*Embedding.*",
|
| 7 |
+
"{re}.*add___.*",
|
| 8 |
+
"{re}.*layer_norm_.*",
|
| 9 |
+
"{re}.*matmul_1",
|
| 10 |
+
"{re}.*__truediv__.*"
|
| 11 |
+
],
|
| 12 |
+
"initializer": {
|
| 13 |
+
"batchnorm_adaptation": {
|
| 14 |
+
"num_bn_adaptation_samples": 0
|
| 15 |
+
},
|
| 16 |
+
"range": {
|
| 17 |
+
"num_init_samples": 300,
|
| 18 |
+
"type": "mean_min_max"
|
| 19 |
+
}
|
| 20 |
+
},
|
| 21 |
+
"overflow_fix": "disable",
|
| 22 |
+
"preset": "mixed",
|
| 23 |
+
"scope_overrides": {
|
| 24 |
+
"activations": {
|
| 25 |
+
"{re}.*matmul_0": {
|
| 26 |
+
"mode": "symmetric"
|
| 27 |
+
}
|
| 28 |
+
}
|
| 29 |
+
}
|
| 30 |
+
},
|
| 31 |
+
"input_info": [
|
| 32 |
+
{
|
| 33 |
+
"keyword": "input_ids",
|
| 34 |
+
"sample_size": [
|
| 35 |
+
128,
|
| 36 |
+
66
|
| 37 |
+
],
|
| 38 |
+
"type": "long"
|
| 39 |
+
},
|
| 40 |
+
{
|
| 41 |
+
"keyword": "token_type_ids",
|
| 42 |
+
"sample_size": [
|
| 43 |
+
128,
|
| 44 |
+
66
|
| 45 |
+
],
|
| 46 |
+
"type": "long"
|
| 47 |
+
},
|
| 48 |
+
{
|
| 49 |
+
"keyword": "attention_mask",
|
| 50 |
+
"sample_size": [
|
| 51 |
+
128,
|
| 52 |
+
66
|
| 53 |
+
],
|
| 54 |
+
"type": "long"
|
| 55 |
+
}
|
| 56 |
+
],
|
| 57 |
+
"optimum_version": "1.14.1",
|
| 58 |
+
"save_onnx_model": false,
|
| 59 |
+
"transformers_version": "4.35.2"
|
| 60 |
+
}
|
openvino_model.bin
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:14e7a63cb6ed4e71f0fb41442750168d21fe4cc1f4a932ab908a126c9ffc3058
|
| 3 |
+
size 138739212
|
openvino_model.xml
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
original_graph.dot
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
pytorch_model.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
-
size
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:b6286ab040d463864d3a8e9d171df02ea044a878ecd16646b9aef131ecbea578
|
| 3 |
+
size 268172814
|
training_args.bin
CHANGED
|
@@ -1,3 +1,3 @@
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
-
oid sha256:
|
| 3 |
size 4600
|
|
|
|
| 1 |
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:77fea7f47d6f91af7370043dfc2a8acdf21fdd576a23651c0002ed739a6042b2
|
| 3 |
size 4600
|