YOLO-NAS-Pose-JetPack5 / yolo_nas_pose_l_int8.onnx.int8.engine.err
Luigi's picture
Upload TensorRT & ONNX Model files
127e10d
raw
history blame
965 Bytes
[12/28/2023-20:16:35] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
[12/28/2023-20:16:35] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
[12/28/2023-20:16:40] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
[12/28/2023-20:31:04] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
[12/28/2023-20:31:04] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
[12/28/2023-20:31:04] [W] * GPU compute time is unstable, with coefficient of variance = 6.47493%.
[12/28/2023-20:31:04] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.