r0.010-squad-bert-b-mvmt-8bit / ir /log.bapp.tput.10k-iter
vchua
Add collaterals
4088e0c
[Step 1/11] Parsing and validating input arguments
[ WARNING ] -nstreams default value is determined automatically for a device. Although the automatic selection usually provides a reasonable performance, but it still may be non-optimal for some cases, for more information look at README.
[Step 2/11] Loading OpenVINO
[ WARNING ] PerformanceMode was not explicitly specified in command line. Device CPU performance hint will be set to THROUGHPUT.
[ INFO ] OpenVINO:
API version............. 2022.1.0-6935-7cd3c8e86e9
[ INFO ] Device info
CPU
openvino_intel_cpu_plugin version 2022.1
Build................... 2022.1.0-6935-7cd3c8e86e9
[Step 3/11] Setting device configuration
[ WARNING ] -nstreams default value is determined automatically for CPU device. Although the automatic selection usually provides a reasonable performance, but it still may be non-optimal for some cases, for more information look at README.
[Step 4/11] Reading network files
[ INFO ] Read model took 64.90 ms
[Step 5/11] Resizing network to match image sizes and given batch
[ WARNING ] Batch dimension is not specified for input 'input.0'. The first dimension will be interpreted as batch size.
[ WARNING ] Batch dimension is not specified for input 'input.1'. The first dimension will be interpreted as batch size.
[ WARNING ] Batch dimension is not specified for input 'input.2'. The first dimension will be interpreted as batch size.
[ INFO ] Reshaping model: 'input.0': {1,384}, 'input.1': {1,384}, 'input.2': {1,384}
[ INFO ] Reshape model took 0.05 ms
[ INFO ] Network batch size: 1
[Step 6/11] Configuring input of the model
[ INFO ] Model input 'input.0' precision i64, dimensions ([N,...]): 1 384
[ INFO ] Model input 'input.1' precision i64, dimensions ([N,...]): 1 384
[ INFO ] Model input 'input.2' precision i64, dimensions ([N,...]): 1 384
[ INFO ] Model output 'output.0' precision f32, dimensions ([...]): 1 384
[ INFO ] Model output 'output.1' precision f32, dimensions ([...]): 1 384
[Step 7/11] Loading the model to the device
[ INFO ] Compile model took 788.57 ms
[Step 8/11] Querying optimal runtime parameters
[ INFO ] DEVICE: CPU
[ INFO ] AVAILABLE_DEVICES , ['']
[ INFO ] RANGE_FOR_ASYNC_INFER_REQUESTS , (1, 1, 1)
[ INFO ] RANGE_FOR_STREAMS , (1, 152)
[ INFO ] FULL_DEVICE_NAME , Intel(R) Xeon(R) Platinum 8368 CPU @ 2.40GHz
[ INFO ] OPTIMIZATION_CAPABILITIES , ['WINOGRAD', 'FP32', 'FP16', 'INT8', 'BIN', 'EXPORT_IMPORT']
[ INFO ] CACHE_DIR ,
[ INFO ] NUM_STREAMS , 19
[ INFO ] INFERENCE_NUM_THREADS , 0
[ INFO ] PERF_COUNT , False
[ INFO ] PERFORMANCE_HINT_NUM_REQUESTS , 0
[Step 9/11] Creating infer requests and preparing input data
[ INFO ] Create 19 infer requests took 1.50 ms
[ WARNING ] No input files were given for input 'input.0'!. This input will be filled with random values!
[ WARNING ] No input files were given for input 'input.1'!. This input will be filled with random values!
[ WARNING ] No input files were given for input 'input.2'!. This input will be filled with random values!
[ INFO ] Fill input 'input.0' with random values
[ INFO ] Fill input 'input.1' with random values
[ INFO ] Fill input 'input.2' with random values
[ WARNING ] Number of iterations was aligned by request number from 10000 to 10013 using number of requests 19
[Step 10/11] Measuring performance (Start inference asynchronously, 19 inference requests using 19 streams for CPU, inference only: True, limits: 10013 iterations)
[ INFO ] Benchmarking in inference only mode (inputs filling are not included in measurement loop).
[ INFO ] First inference took 37.43 ms
[Step 11/11] Dumping statistics report
Count: 10013 iterations
Duration: 24302.14 ms
Latency:
Median: 43.95 ms
AVG: 45.97 ms
MIN: 39.02 ms
MAX: 126.30 ms
Throughput: 412.02 FPS