Note: This video may require joining the NVIDIA Developer Program or login

GTC Silicon Valley-2019 ID:S9343:Accelerating Distributed Deep Learning Inference on multi-GPU with Hadoop-Spark

SeYoon Oh(Agency for Defense Development),Hunmin Yang(Agency for Defense Development)
Learn how to develop faster, scalable, and better GPU-Accelerated distributed inference on multi-node and multi-GPU cluster environments. In most of the benchmark cases, linear scalability for throughput performance is not guaranteed with increasing the number of GPUs and servers. We'll discuss present an efficient scale-out method for deploying deep learning-based object detection models on multi-node and multi-GPU clusters using Apache Hadoop and Spark. We'll explain how to deploy 120 deep learning models (YOLOv2) on our own video datasets with NVIDIA Tesla M60 GPU (30EA) Hadoop cluster. Our approach achieved about 20 percent faster inference throughput with super-linear scalability from one GPU server to 30 GPU cluster. This session will be a combination of lecture and videos about our GPU-Accelerated distributed inference platform for large-scale streaming data analytics.

View the slides (pdf)