DEVELOPER BLOG

Tag: Automatic Mixed Precision (AMP)

AI / Deep Learning

Accelerating TensorFlow on NVIDIA A100 GPUs

The NVIDIA A100, based on the NVIDIA Ampere GPU architecture, offers a suite of exciting new features: third-generation Tensor Cores, Multi-Instance GPU (MIG)… 12 MIN READ
AI / Deep Learning

Develop Smaller Speech Recognition Models with NVIDIA’s NeMo Framework

As computers and other personal devices have become increasingly prevalent, interest in conversational AI has grown due to its multitude of potential… 7 MIN READ
AI / Deep Learning

Neural Modules for Fast Development of Speech and Language Models

As a researcher building state-of-the-art conversational AI models, you need to be able to quickly experiment with novel network architectures. 6 MIN READ
AI / Deep Learning

Creating an Object Detection Pipeline for GPUs

Earlier this year in March, we showed retinanet-examples, an open source example of how to accelerate the training and deployment of an object detection… 16 MIN READ
AI / Deep Learning

Automatic Mixed Precision for NVIDIA Tensor Core Architecture in TensorFlow

Whether to employ mixed precision to train your TensorFlow models is no longer a tough decision. NVIDIA’s Automatic Mixed Precision (AMP) feature for TensorFlow… 5 MIN READ
AI / Deep Learning

NVIDIA Apex: Tools for Easy Mixed-Precision Training in PyTorch

Most deep learning frameworks, including PyTorch, train using 32-bit floating point (FP32) arithmetic by default. However, using FP32 for all operations is not… 8 MIN READ