Posts by Dave Salvator
Technical Walkthrough
Apr 06, 2022
Getting the Best Performance on MLPerf Inference 2.0
NVIDIA delivered leading results for MLPerf Inference 2.0, including 5x more performance for NVIDIA Jetson AGX Orin, an SoC platform built for edge devices and robotics.
11 MIN READ
News
Nov 29, 2021
AWS Launches First NVIDIA GPU-Accelerated Graviton-Based Instance with Amazon EC2 G5g
The new Amazon EC2 G5g instances feature the AWS Graviton2 processors and NVIDIA T4G Tensor Core GPUs, to power rich android game streaming for mobile devices.
3 MIN READ
Technical Walkthrough
Sep 22, 2021
Furthering NVIDIA Performance Leadership with MLPerf Inference 1.1 Results
A look at NVIDIA inference performance as measured by the MLPerf Inference 1.1 benchmark.
6 MIN READ
News
May 14, 2021
Getting the Most Out of NVIDIA T4 on AWS G4 Instances
Learn how to get the best natural language inference performance from AWS G4dn instance powered by NVIDIA T4 GPUs, and how to deploy BERT networks easily using NVIDIA Triton Inference Server.
14 MIN READ
News
May 03, 2021
MLOps Made Simple & Cost Effective with Google Kubernetes Engine and NVIDIA A100 Multi-Instance GPUs
Google Cloud and NVIDIA collaborated to make MLOps simple, powerful, and cost-effective by bringing together the solution elements to build, serve and dynamically scale your end-to-end ML pipelines with the right-sized GPU acceleration in one place.
5 MIN READ
Technical Walkthrough
Apr 22, 2021
Extending NVIDIA Performance Leadership with MLPerf Inference 1.0 Results
In this post, we step through some of these optimizations, including the use of Triton Inference Server and the A100 Multi-Instance GPU (MIG) feature.
7 MIN READ