Close

Presentation

Beyond Inference: Performance Analysis of DNN Server Overheads for Computer Vision
DescriptionDeep neural network (DNN) inference has become an important
part of many data-center workloads. This has prompted focused ef-
forts to design ever-faster deep learning accelerators such as GPUs
and TPUs. However, an end-to-end vision application contains
more than just DNN inference, including input decompression, re-
sizing, sampling, normalization, and data transfer. In this paper,
we perform a thorough evaluation of computer vision inference
requests performed on a throughput-optimized serving system. We
quantify the performance impact of server overheads such as data
movement, preprocessing, and message brokers between two DNNs
producing outputs at different rates. Our empirical analysis encom-
passes many computer vision tasks including image classification,
segmentation, detection, depth-estimation, and more complex pro-
cessing pipelines with multiple DNNs. Our results consistently
demonstrate that end-to-end application performance can easily
be dominated by data processing and data movement functions (up
to 56% of end-to-end latency in a medium-sized image, and ∼ 80%
impact on system throughput in a large image), even though these
functions have been conventionally overlooked in deep learning
system design. Our work identifies important performance bottle-
necks in different application scenarios, achieves
2.25× better throughput compared to prior work, and paves the
way for more holistic deep learning system design.
Event Type
Research Manuscript
TimeTuesday, June 254:00pm - 4:15pm PDT
Location3001, 3rd Floor
Topics
AI
Keywords
AI/ML Application and Infrastructure