Posts by Jay Rodge
Technical Walkthrough
Jul 20, 2022
Optimizing and Serving Models with NVIDIA TensorRT and NVIDIA Triton
Imagine that you have trained your model with PyTorch, TensorFlow, or the framework of your choice, are satisfied with its accuracy, and are considering...
11 MIN READ
News
Dec 02, 2021
NVIDIA Announces TensorRT 8.2 and Integrations with PyTorch and TensorFlow
Today NVIDIA released TensorRT 8.2, with optimizations for billion parameter NLU models. These include T5 and GPT-2, used for translation and text generation,...
2 MIN READ
Technical Walkthrough
Dec 02, 2021
Optimizing T5 and GPT-2 for Real-Time Inference with NVIDIA TensorRT
The transformer architecture has wholly transformed (pun intended) the domain of natural language processing (NLP). Over the recent years, many novel network...
9 MIN READ
News
Nov 09, 2021
ICYMI: New AI Tools and Technologies Announced at NVIDIA GTC Keynote
At NVIDIA GTC this November, new software tools were announced that help developers build real-time speech applications, optimize inference for a variety of...
5 MIN READ
News
Oct 05, 2021
NVIDIA GTC: Can’t-Miss Sessions in AI and Deep Learning this November
Join NVIDIA November 8-11, showcasing over 500 GTC sessions covering the latest breakthroughs in AI and deep learning, as well as many other GPU technology...
4 MIN READ
News
Jul 20, 2021
NVIDIA Announces TensorRT 8 Slashing BERT-Large Inference Down to 1 Millisecond
Today, NVIDIA announced TensorRT 8.0 which brings BERT-Large inference latency down to 1.2 ms with new optimizations. This version also delivers 2x the accuracy...
3 MIN READ