qaihm-bot commited on
Commit
af81ecb
1 Parent(s): 5d9c3b5

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -39,7 +39,7 @@ More details on model performance across various devices, can be found
39
 
40
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
41
  | ---|---|---|---|---|---|---|---|
42
- | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 1.717 ms | 0 - 261 MB | INT8 | NPU | [ConvNext-Tiny-w8a8-Quantized.so](https://huggingface.co/qualcomm/ConvNext-Tiny-w8a8-Quantized/blob/main/ConvNext-Tiny-w8a8-Quantized.so)
43
 
44
 
45
 
@@ -102,8 +102,8 @@ python -m qai_hub_models.models.convnext_tiny_w8a8_quantized.export
102
  Profile Job summary of ConvNext-Tiny-w8a8-Quantized
103
  --------------------------------------------------
104
  Device: Snapdragon X Elite CRD (11)
105
- Estimated Inference Time: 1.86 ms
106
- Estimated Peak Memory Range: 0.43-0.43 MB
107
  Compute Units: NPU (215) | Total (215)
108
 
109
 
 
39
 
40
  | Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
41
  | ---|---|---|---|---|---|---|---|
42
+ | Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 1.721 ms | 0 - 120 MB | INT8 | NPU | [ConvNext-Tiny-w8a8-Quantized.so](https://huggingface.co/qualcomm/ConvNext-Tiny-w8a8-Quantized/blob/main/ConvNext-Tiny-w8a8-Quantized.so)
43
 
44
 
45
 
 
102
  Profile Job summary of ConvNext-Tiny-w8a8-Quantized
103
  --------------------------------------------------
104
  Device: Snapdragon X Elite CRD (11)
105
+ Estimated Inference Time: 1.82 ms
106
+ Estimated Peak Memory Range: 0.45-0.45 MB
107
  Compute Units: NPU (215) | Total (215)
108
 
109