Technical Walkthrough 0

Minimizing Deep Learning Inference Latency with NVIDIA Multi-Instance GPU

Recently, NVIDIA unveiled the A100 GPU model, based on the NVIDIA Ampere architecture. Ampere introduced many features, including Multi-Instance GPU (MIG), that... 20 MIN READ
Technical Walkthrough 1

Optimizing NVIDIA AI Performance for MLPerf v0.7 Training

MLPerf is an industry-wide AI consortium that has developed a suite of performance benchmarks covering a range of leading AI workloads that are widely in use... 16 MIN READ
Technical Walkthrough 0

Int4 Precision for AI Inference

INT4 Precision Can Bring an Additional 59% Speedup Compared to INT8 If there’s one constant in AI and deep learning, it’s never-ending optimization to wring... 5 MIN READ
Technical Walkthrough 1

MLPerf Inference: NVIDIA Innovations Bring Leading Performance

New TensorRT 6 Features Combine with Open-Source Plugins to Further Accelerate Inference  Inference is where AI goes to work. Identifying diseases. Answering... 7 MIN READ
Technical Walkthrough 0

NVIDIA Boosts AI Performance in MLPerf v0.6

The relentless pace of innovation is most apparent in the AI domain. Researchers and developers discovering new network architectures, algorithms and... 10 MIN READ
Technical Walkthrough 0

Case Study: ResNet50 with DALI

Let’s imagine a situation. You buy a brand-new, cutting-edge, Volta-powered DGX-2 server. You’ve done your math right, expecting a 2x performance increase... 11 MIN READ