Accelerating IO in the Modern Data Center: Network IO

This is the second post in the Explaining Magnum IO series, which described the architecture, components, and benefits of Magnum IO, the IO subsystem of the… 19 MIN READ
Artificial Intelligence

Massively Scale Your Deep Learning Training with NCCL 2.4

Imagine using tens of thousands of GPUs to train your neural network. Using multiple GPUs to train neural networks has become quite common with all deep… 8 MIN READ
Accelerated Computing

Scaling Deep Learning Training with NCCL

NVIDIA Collective Communications Library (NCCL) provides optimized implementation of inter-GPU communication operations, such as allreduce and variants. 6 MIN READ
AI / Deep Learning

NVIDIA Deep Learning SDK Update for Volta Now Available

At GTC 2017, NVIDIA announced Volta optimized updates to the NVIDIA Deep Learning SDK. Today, we’re making these updates available as free downloads to members… 2 MIN READ
Artificial Intelligence

Fast Multi-GPU collectives with NCCL

Today many servers contain 8 or more GPUs. In principle then, scaling an application from one to many GPUs should provide a tremendous performance boost. 10 MIN READ