Developer Blog: Speeding Up Deep Learning Inference Using TensorRT
By Nadeem Mohammad, posted Apr 22 2020 at 01:55PM
NVIDIA TensorRT is an SDK for deep learning inference. TensorRT provides APIs and parsers to import trained models from all major deep learning frameworks. It then generates optimized runtime engines deployable in the datacenter as well as in automotive and embedded environments.