The NVIDIA TensorRT inference server GA version is now available for download in a container from the NVIDIA GPU Cloud container registry. Announced at GTC Japan and part of the NVIDIA TensorRT Hyperscale Inference Platform, the TensorRT inference server is a containerized microservice for data center production deployments. As more and more applications leverage AI,

The post NVIDIA TensorRT Inference Server Available Now appeared first on NVIDIA Developer News Center.