qaihm-bot commited on
Commit
ef4751d
1 Parent(s): 49933fd

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -39,7 +39,7 @@ More details on model performance across various devices, can be found
39
 
40
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
41
  | ---|---|---|---|---|---|---|---|
42
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 3.44 ms | 0 - 19 MB | INT8 | NPU | [ConvNext-Tiny-w8a16-Quantized.so](https://huggingface.co/qualcomm/ConvNext-Tiny-w8a16-Quantized/blob/main/ConvNext-Tiny-w8a16-Quantized.so)
43
 
44
 
45
 
@@ -102,8 +102,8 @@ python -m qai_hub_models.models.convnext_tiny_w8a16_quantized.export
102
  Profile Job summary of ConvNext-Tiny-w8a16-Quantized
103
  --------------------------------------------------
104
  Device: Snapdragon X Elite CRD (11)
105
- Estimated Inference Time: 3.49 ms
106
- Estimated Peak Memory Range: 1.43-1.43 MB
107
  Compute Units: NPU (215) | Total (215)
108
 
109
 
 
39
 
40
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
41
  | ---|---|---|---|---|---|---|---|
42
+ | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 3.447 ms | 0 - 12 MB | INT8 | NPU | [ConvNext-Tiny-w8a16-Quantized.so](https://huggingface.co/qualcomm/ConvNext-Tiny-w8a16-Quantized/blob/main/ConvNext-Tiny-w8a16-Quantized.so)
43
 
44
 
45
 
 
102
  Profile Job summary of ConvNext-Tiny-w8a16-Quantized
103
  --------------------------------------------------
104
  Device: Snapdragon X Elite CRD (11)
105
+ Estimated Inference Time: 3.35 ms
106
+ Estimated Peak Memory Range: 0.29-0.29 MB
107
  Compute Units: NPU (215) | Total (215)
108
 
109