A company has set up and deployed its machine learning (ML) model into production with an endpoint using Amazon SageMaker hosting services. The ML team has configured automatic scaling for its SageMaker instances to support workload changes. During testing, the team notices that additional instances are being launched before the new instances are ready. This behavior needs to change as soon as possible. How can the ML team solve this issue? A. Decrease the cooldown period for the scale-in activity. Increase the configured maximum capacity of instances. B. Replace the current endpoint with a multi-model endpoint using SageMaker. C. Set up Amazon API Gateway and AWS Lambda to trigger the SageMaker inference endpoint. D. Increase the cooldown period for the scale-out activity.  Suggested Answer: A Community Answer: D Reference: https://aws.amazon.com/blogs/machine-learning/configuring-autoscaling-inference-endpoints-in-amazon-sagemaker/ This question is in MLS-C01 AWS Certified Machine Learning – Specialty Exam For getting AWS Certified Machine Learning – Specialty Certificate Disclaimers: The website is not related to, affiliated with, endorsed or authorized by Amazon. Trademarks, certification & product names are used for reference only and belong to Amazon. The website does not contain actual questions and answers from Amazon's Certification Exam.
Please login or Register to submit your answer