violetch24 commited on
Commit
793cccf
1 Parent(s): aba9433

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -29,8 +29,7 @@ model-index:
29
 
30
  ### Post-training static quantization
31
 
32
- This is an INT8 PyTorch model quantized with [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
33
-
34
  The original fp32 model comes from the fine-tuned model [thyagosme/bert-base-uncased-finetuned-swag](https://huggingface.co/thyagosme/bert-base-uncased-finetuned-swag).
35
 
36
  The calibration dataloader is the train dataloader. The default calibration sampling size 100 isn't divisible exactly by batch size 8, so the real sampling size is 104.
@@ -44,11 +43,11 @@ The linear modules **bert.encoder.layer.2.output.dense, bert.encoder.layer.5.int
44
  | **Accuracy (eval-accuracy)** |0.7838|0.7915|
45
  | **Model size (MB)** |133|418|
46
 
47
- ### Load with Intel® Neural Compressor:
48
 
49
  ```python
50
- from neural_compressor.utils.load_huggingface import OptimizedModel
51
- int8_model = OptimizedModel.from_pretrained(
52
  'Intel/bert-base-uncased-finetuned-swag-int8-static',
53
  )
54
  ```
 
29
 
30
  ### Post-training static quantization
31
 
32
+ This is an INT8 PyTorch model quantized with [huggingface/optimum-intel](https://github.com/huggingface/optimum-intel) through the usage of [Intel® Neural Compressor](https://github.com/intel/neural-compressor).
 
33
  The original fp32 model comes from the fine-tuned model [thyagosme/bert-base-uncased-finetuned-swag](https://huggingface.co/thyagosme/bert-base-uncased-finetuned-swag).
34
 
35
  The calibration dataloader is the train dataloader. The default calibration sampling size 100 isn't divisible exactly by batch size 8, so the real sampling size is 104.
 
43
  | **Accuracy (eval-accuracy)** |0.7838|0.7915|
44
  | **Model size (MB)** |133|418|
45
 
46
+ ### Load with optimum:
47
 
48
  ```python
49
+ from optimum.intel.neural_compressor.quantization import IncQuantizedModelForMultipleChoice
50
+ int8_model = IncQuantizedModelForMultipleChoice.from_pretrained(
51
  'Intel/bert-base-uncased-finetuned-swag-int8-static',
52
  )
53
  ```