[Step 1/11] Parsing and validating input arguments [ WARNING ] -nstreams default value is determined automatically for a device. Although the automatic selection usually provides a reasonable performance, but it still may be non-optimal for some cases, for more information look at README. [Step 2/11] Loading OpenVINO [ WARNING ] PerformanceMode was not explicitly specified in command line. Device CPU performance hint will be set to THROUGHPUT. [ INFO ] OpenVINO: API version............. 2022.1.0-6935-7cd3c8e86e9 [ INFO ] Device info CPU openvino_intel_cpu_plugin version 2022.1 Build................... 2022.1.0-6935-7cd3c8e86e9 [Step 3/11] Setting device configuration [ WARNING ] -nstreams default value is determined automatically for CPU device. Although the automatic selection usually provides a reasonable performance, but it still may be non-optimal for some cases, for more information look at README. [Step 4/11] Reading network files [ INFO ] Read model took 64.90 ms [Step 5/11] Resizing network to match image sizes and given batch [ WARNING ] Batch dimension is not specified for input 'input.0'. The first dimension will be interpreted as batch size. [ WARNING ] Batch dimension is not specified for input 'input.1'. The first dimension will be interpreted as batch size. [ WARNING ] Batch dimension is not specified for input 'input.2'. The first dimension will be interpreted as batch size. [ INFO ] Reshaping model: 'input.0': {1,384}, 'input.1': {1,384}, 'input.2': {1,384} [ INFO ] Reshape model took 0.05 ms [ INFO ] Network batch size: 1 [Step 6/11] Configuring input of the model [ INFO ] Model input 'input.0' precision i64, dimensions ([N,...]): 1 384 [ INFO ] Model input 'input.1' precision i64, dimensions ([N,...]): 1 384 [ INFO ] Model input 'input.2' precision i64, dimensions ([N,...]): 1 384 [ INFO ] Model output 'output.0' precision f32, dimensions ([...]): 1 384 [ INFO ] Model output 'output.1' precision f32, dimensions ([...]): 1 384 [Step 7/11] Loading the model to the device [ INFO ] Compile model took 788.57 ms [Step 8/11] Querying optimal runtime parameters [ INFO ] DEVICE: CPU [ INFO ] AVAILABLE_DEVICES , [''] [ INFO ] RANGE_FOR_ASYNC_INFER_REQUESTS , (1, 1, 1) [ INFO ] RANGE_FOR_STREAMS , (1, 152) [ INFO ] FULL_DEVICE_NAME , Intel(R) Xeon(R) Platinum 8368 CPU @ 2.40GHz [ INFO ] OPTIMIZATION_CAPABILITIES , ['WINOGRAD', 'FP32', 'FP16', 'INT8', 'BIN', 'EXPORT_IMPORT'] [ INFO ] CACHE_DIR , [ INFO ] NUM_STREAMS , 19 [ INFO ] INFERENCE_NUM_THREADS , 0 [ INFO ] PERF_COUNT , False [ INFO ] PERFORMANCE_HINT_NUM_REQUESTS , 0 [Step 9/11] Creating infer requests and preparing input data [ INFO ] Create 19 infer requests took 1.50 ms [ WARNING ] No input files were given for input 'input.0'!. This input will be filled with random values! [ WARNING ] No input files were given for input 'input.1'!. This input will be filled with random values! [ WARNING ] No input files were given for input 'input.2'!. This input will be filled with random values! [ INFO ] Fill input 'input.0' with random values [ INFO ] Fill input 'input.1' with random values [ INFO ] Fill input 'input.2' with random values [ WARNING ] Number of iterations was aligned by request number from 10000 to 10013 using number of requests 19 [Step 10/11] Measuring performance (Start inference asynchronously, 19 inference requests using 19 streams for CPU, inference only: True, limits: 10013 iterations) [ INFO ] Benchmarking in inference only mode (inputs filling are not included in measurement loop). [ INFO ] First inference took 37.43 ms [Step 11/11] Dumping statistics report Count: 10013 iterations Duration: 24302.14 ms Latency: Median: 43.95 ms AVG: 45.97 ms MIN: 39.02 ms MAX: 126.30 ms Throughput: 412.02 FPS