xinhe commited on
Commit
fbf0690
·
1 Parent(s): d70d94a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +0 -7
README.md CHANGED
@@ -25,12 +25,8 @@ The linear module **bert.encoder.layer.9.output.dense, bert.encoder.layer.10.out
25
 
26
  ### Test result
27
 
28
- - Batch size = 8
29
- - [Amazon Web Services](https://aws.amazon.com/) c6i.xlarge (Intel ICE Lake: 4 vCPUs, 8g Memory) instance.
30
-
31
  | |INT8|FP32|
32
  |---|:---:|:---:|
33
- | **Throughput (samples/sec)** |23.986|11.202|
34
  | **Accuracy (eval-f1)** |0.8997|0.9042|
35
  | **Model size (MB)** |120|418|
36
 
@@ -42,6 +38,3 @@ int8_model = OptimizedModel.from_pretrained(
42
  'Intel/bert-base-uncased-mrpc-int8-static',
43
  )
44
  ```
45
-
46
- Notes:
47
- - The INT8 model has better performance than the FP32 model when the CPU is fully occupied. Otherwise, there will be the illusion that INT8 is inferior to FP32.
 
25
 
26
  ### Test result
27
 
 
 
 
28
  | |INT8|FP32|
29
  |---|:---:|:---:|
 
30
  | **Accuracy (eval-f1)** |0.8997|0.9042|
31
  | **Model size (MB)** |120|418|
32
 
 
38
  'Intel/bert-base-uncased-mrpc-int8-static',
39
  )
40
  ```