Upload README.md with huggingface_hub
Browse files
README.md
CHANGED
@@ -39,7 +39,7 @@ More details on model performance across various devices, can be found
|
|
39 |
|
40 |
| Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
|
41 |
| ---|---|---|---|---|---|---|---|
|
42 |
-
| Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 3.
|
43 |
|
44 |
|
45 |
|
@@ -102,8 +102,8 @@ python -m qai_hub_models.models.convnext_tiny_w8a16_quantized.export
|
|
102 |
Profile Job summary of ConvNext-Tiny-w8a16-Quantized
|
103 |
--------------------------------------------------
|
104 |
Device: Snapdragon X Elite CRD (11)
|
105 |
-
Estimated Inference Time: 3.
|
106 |
-
Estimated Peak Memory Range:
|
107 |
Compute Units: NPU (215) | Total (215)
|
108 |
|
109 |
|
|
|
39 |
|
40 |
| Device | Chipset | Target Runtime | Inference Time (ms) | Peak Memory Range (MB) | Precision | Primary Compute Unit | Target Model
|
41 |
| ---|---|---|---|---|---|---|---|
|
42 |
+
| Samsung Galaxy S23 Ultra (Android 13) | Snapdragon® 8 Gen 2 | QNN Model Library | 3.447 ms | 0 - 12 MB | INT8 | NPU | [ConvNext-Tiny-w8a16-Quantized.so](https://huggingface.co/qualcomm/ConvNext-Tiny-w8a16-Quantized/blob/main/ConvNext-Tiny-w8a16-Quantized.so)
|
43 |
|
44 |
|
45 |
|
|
|
102 |
Profile Job summary of ConvNext-Tiny-w8a16-Quantized
|
103 |
--------------------------------------------------
|
104 |
Device: Snapdragon X Elite CRD (11)
|
105 |
+
Estimated Inference Time: 3.35 ms
|
106 |
+
Estimated Peak Memory Range: 0.29-0.29 MB
|
107 |
Compute Units: NPU (215) | Total (215)
|
108 |
|
109 |
|