File size: 965 Bytes
127e10d
 
 
 
 
 
 
1
2
3
4
5
6
7
8
[12/28/2023-19:27:37] [W] [TRT] onnx2trt_utils.cpp:375: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
[12/28/2023-19:27:37] [W] [TRT] onnx2trt_utils.cpp:403: One or more weights outside the range of INT32 was clamped
[12/28/2023-19:27:42] [W] [TRT] Calibrator won't be used in explicit precision mode. Use quantization aware training to generate network with Quantize/Dequantize nodes.
[12/28/2023-20:16:25] [W] * Throughput may be bound by Enqueue Time rather than GPU Compute and the GPU may be under-utilized.
[12/28/2023-20:16:25] [W]   If not already in use, --useCudaGraph (utilize CUDA graphs where possible) may increase the throughput.
[12/28/2023-20:16:25] [W] * GPU compute time is unstable, with coefficient of variance = 6.69708%.
[12/28/2023-20:16:25] [W]   If not already in use, locking GPU clock frequency or adding --useSpinWait may improve the stability.