[12/28/2023-20:16:35] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32. | |
[12/28/2023-20:16:35] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped | |
[12/28/2023-20:16:40] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes. | |
[12/28/2023-20:31:04] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized. | |
[12/28/2023-20:31:04] [W] If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput. | |
[12/28/2023-20:31:04] [W] * GPU compute time is unstable, with coefficient of variance = 6.47493%. | |
[12/28/2023-20:31:04] [W] If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability. | |