Ashwin Nanjappa

Ashwin Nanjappa is an engineering manager in the TensorRT team at NVIDIA. He leads the MLPerf Inference initiative to demonstrate the performance and energy efficiency of NVIDIA accelerators. He is also involved in improving the DL inference performance of the TensorRT library. Before joining NVIDIA, he worked on training DL models for CV, GPU-accelerated ML algorithms for depth cameras, and developing multimedia libraries for cellphones and DVD players. He has a Ph.D. in computer science from the National University of Singapore (NUS), with a focus on GPU algorithms for 3D computational geometry.
Avatar photo

Posts by Ashwin Nanjappa

Data Center / Cloud

NVIDIA Blackwell Platform Sets New LLM Inference Records in MLPerf Inference v4.1

Large language model (LLM) inference is a full-stack challenge. Powerful GPUs, high-bandwidth GPU-to-GPU interconnects, efficient acceleration libraries, and a... 13 MIN READ
An image of an NVIDIA H200 Tensor Core GPU.
Generative AI / LLMs

NVIDIA H200 Tensor Core GPUs and NVIDIA TensorRT-LLM Set MLPerf LLM Inference Records

Generative AI is unlocking new computing applications that greatly augment human capability, enabled by continued model innovation. Generative AI... 11 MIN READ
NVIDIA Jetson Orin modules.
Data Center / Cloud

Leading MLPerf Inference v3.1 Results with NVIDIA GH200 Grace Hopper Superchip Debut

AI is transforming computing, and inference is how the capabilities of AI are deployed in the world’s applications. Intelligent chatbots, image and video... 13 MIN READ
Image of Infiniband with decorative images in front.
Networking

New MLPerf Inference Network Division Showcases NVIDIA InfiniBand and GPUDirect RDMA Capabilities

In MLPerf Inference v3.0, NVIDIA made its first submissions to the newly introduced Network division, which is now part of the MLPerf Inference Datacenter... 9 MIN READ
Data Center / Cloud

Setting New Records in MLPerf Inference v3.0 with Full-Stack Optimizations for AI

The most exciting computing applications currently rely on training and running inference on complex AI models, often in demanding, real-time deployment... 15 MIN READ
Simulation / Modeling / Design

Full-Stack Innovation Fuels Highest MLPerf Inference 2.1 Results for NVIDIA

Today’s AI-powered applications are enabling richer experiences, fueled by both larger and more complex AI models as well as the application of many models in... 14 MIN READ