You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
From the benchmarking results above, We can deduce that the models initialize and load faster at FP16 precision. This faster models load time shortens the application's initialization time allowing the application to start up faster at FP16 model precision than when the models are at FP32 precision. The inference speed falls within a range of values but it peaks at the batch size of 1 for both precisions. Also, The average inference speed is lowest when the batch size is 50. The total inference time of the application spikes when batch size is 1 and drops to about 1.30s at bach size 25 and slowly declines as the batch size is increased.

454
461
462
+
463
+
455
464
#### Summary of Results
456
465
457
466
The VTune Amplifier was used to perform an analysis on the inference pipeline to fine the top hotspots in the application. The data collection and collation was performed in 7.661s, 2.180s out of which was used by the CPU to actively execute the application. A total of 21 threads were used by the CPU. The Hotspots insights section of the summary showed that the application utilized only 1 CPU core with a 13.3% parallelism and a microarchitecture usage of 23.2%.
@@ -482,7 +491,8 @@ The application also logs the Gaze vector outputs and inference statistics in th
482
491
483
492
- Batch count: Number of batches processed before inference was finished or interrupted.
484
493
485
-
#### Logged Last outputs
494
+
495
+
#### Final outputs that were logged
486
496
- Face coordinates: Coordinates of the first face in the frame detected by the face detection model.
487
497
488
498
- Left eye coordinates: Coordinates of the left eye of the face detected by the facial landmark model.
0 commit comments