Host ML models on Amazon SageMaker using Triton: TensorRT models Melanie Li AWS Machine Learning Blog
Sometimes it can be very beneficial to use tools such as compilers that can modify and compile your models for optimal inference performance. In this post, we explore TensorRT and how to use it with Amazon SageMaker inference using NVIDIA Triton Inference Server. We… Read More »Host ML models on Amazon SageMaker using Triton: TensorRT models Melanie Li AWS Machine Learning Blog