Topic 1 Question 220
A manufacturing company wants to monitor its devices for anomalous behavior. A data scientist has trained an Amazon SageMaker scikit-learn model that classifies a device as normal or anomalous based on its 4-day telemetry. The 4-day telemetry of each device is collected in a separate file and is placed in an Amazon S3 bucket once every hour. The total time to run the model across the telemetry for all devices is 5 minutes.
What is the MOST cost-effective solution for the company to use to run the model across the telemetry for all the devices?
SageMaker Batch Transform
SageMaker Asynchronous Inference
SageMaker Processing
A SageMaker multi-container endpoint
ユーザの投票
コメント(5)
Real-Time Inference is suitable for workloads where payload sizes are up to 6MB and need to be processed with low latency requirements in the order of milliseconds or seconds.
Serverless Inference: Serverless inference is ideal when you have intermittent or unpredictable traffic patterns.
Batch transform is ideal for offline predictions on large batches of data that is available upfront.
We are introducing Amazon SageMaker Asynchronous Inference, a new inference option in Amazon SageMaker that queues incoming requests and processes them asynchronously. This option is ideal for inferences with large payload sizes (up to 1GB) and/or long processing times (up to 15 minutes) that need to be processed as requests arrive. Asynchronous inference enables you to save on costs by autoscaling the instance count to zero when there are no requests to process, so you only pay when your endpoint is processing requests.
a
👍 2drcok872023/02/10- 正解だと思う選択肢: A
Based on the requirements and constraints given in the scenario, the MOST cost-effective solution for the company to use to run the model across the telemetry for all the devices is SageMaker Batch Transform.
SageMaker Batch Transform is a cost-effective solution for performing offline inference, as it allows for large amounts of data to be processed at a lower cost compared to real-time inference. In this case, the telemetry data for each device is collected hourly and can be processed in batches using SageMaker Batch Transform. This can help to reduce the cost of inference, as the data is not being processed in real-time and can be processed offline.
👍 2oso03482023/03/04 - 正解だと思う選択肢: A
I will go with A. The Async inference seems promising but the size of telemetry file is not known.
As per https://docs.aws.amazon.com/sagemaker/latest/dg/inference-cost-optimization.html
"Use batch inference for workloads for which you need inference for a large set of data for processes that happen offline (that is, you don’t need a persistent endpoint). You pay for the instance for the duration of the batch inference job". As you pay for the batch job duration, cost should not be an issue with Batch transform.
"Use asynchronous inference for asynchronous workloads that process up to 1 GB of data (such as text corpus, image, video, and audio) that are latency insensitive and cost sensitive. With asynchronous inference, you can control costs by specifying a fixed number of instances for the optimal processing rate instead of provisioning for the peak. You can also scale down to zero to save additional costs."
👍 2pan_b2023/03/13
シャッフルモード