DEVELOPER BLOG
Tag: MIG
AI / Deep Learning
Jan 26, 2021
Adding More Support in NVIDIA GPU Operator
Reliably provisioning servers with GPUs can quickly become complex as multiple components must be installed and managed to use GPUs with Kubernetes.
6 MIN READ
AI / Deep Learning
Dec 18, 2020
Minimizing Deep Learning Inference Latency with NVIDIA Multi-Instance GPU
Recently, NVIDIA unveiled the A100 GPU model, based on the NVIDIA Ampere architecture. Ampere introduced many features, including Multi-Instance GPU (MIG)…
20 MIN READ
HPC
Dec 10, 2020
Supercharging the World’s Fastest AI Supercomputing Platform on NVIDIA HGX A100 80GB GPUs
Exploding model sizes in deep learning and AI, complex simulations in high-performance computing (HPC), and massive datasets in data analytics all continue to…
5 MIN READ
AI / Deep Learning
Nov 30, 2020
Getting the Most Out of the NVIDIA A100 GPU with Multi-Instance GPU
With the third-generation Tensor Core technology, NVIDIA recently unveiled A100 Tensor Core GPU that delivers unprecedented acceleration at every scale for AI…
18 MIN READ
AI / Deep Learning
Nov 30, 2020
Getting Kubernetes Ready for the NVIDIA A100 GPU with Multi-Instance GPU
Multi-Instance GPU (MIG) is a new feature of the latest generation of NVIDIA GPUs, such as A100. It enables users to maximize the utilization of a single GPU by…
13 MIN READ
AI / Deep Learning
Nov 13, 2020
Getting Immediate Speedups with NVIDIA A100 TF32
The NVIDIA A100 brought the biggest single-generation performance gains ever in our company’s history. These speedups are a product of architectural innovations…
5 MIN READ
