Enter your
email for instant access to
Mod Week AI/ML

Opt-in to Future Emails
Yes, I'd like Amazon Web Services (AWS) to share the latest news about AWS services and related offerings with me by email, post or telephone.

You may unsubscribe from receiving AWS news and offers at any time by following the instructions in the communications received. AWS handles your information as described in the AWS Privacy Notice.
Thank you!
Error - something went wrong!

Easily deploy models for the best performance and cost using Amazon SageMaker

Optimizing cloud resources to achieve the best cost and performance for your ML model is critical.

In this session, learn how Amazon SageMaker Inference Recommender automatically selects the compute instance type, instance count, container parameters, and model optimizations for inference to maximize performance and minimize cost.

You can then deploy your model to one of the recommended instances or run a fully managed load test on a set of instance types you choose without worrying about testing infrastructure.

You can review results of the load test in SageMaker Studio and evaluate the tradeoffs between latency, throughput, and cost to select the most optimal deployment configuration for your use case.

Click here and download the presentation deck to learn more!

Previous Video
Achieve high performance and cost-effective model deployment
Achieve high performance and cost-effective model deployment

Explore the different inference options available in Amazon SageMaker such as multi-container endpoints, in...

No More Videos

Questions about AI/ML?

Get in touch »