Training Performance

NVIDIA’s complete solution stack, from GPUs to optimized containers on NVIDIA GPU Cloud (NGC), allows data scientists to quickly get up and running with deep learning. By leveraging NVIDIA’s platform, researchers can accelerate deep learning training across every framework and every type of neural network. See NVIDIA’s training performance data for more details:

Image Classification (CNNs)

Natural Language Processing (RNN) Translation

Inference Performance

NVIDIA® TensorRT™ running on NVIDIA GPUs enable the most efficient inference performance across multiple application areas and models. This provides wide latitude to data scientists to create the optimal low-latency solution. See NVIDIA’s inference performance data to learn how NVIDIA can accelerate your research needs:

Image Classification (CNNs) with TensorRT


Natural Language Processing (RNN) Translation with TensorRT