Cyber Monday 2024! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Amazon Exam MLS-C01 Topic 2 Question 100 Discussion

Actual exam question for Amazon's MLS-C01 exam
Question #: 100
Topic #: 2
[All MLS-C01 Questions]

A data scientist uses Amazon SageMaker Data Wrangler to define and perform transformations and feature engineering on historical dat

a. The data scientist saves the transformations to SageMaker Feature Store.

The historical data is periodically uploaded to an Amazon S3 bucket. The data scientist needs to transform the new historic data and add it to the online feature store The data scientist needs to prepare the .....historic data for training and inference by using native integrations.

Which solution will meet these requirements with the LEAST development effort?

Show Suggested Answer Hide Answer
Suggested Answer: D

The best solution is to configure Amazon EventBridge to run a predefined SageMaker pipeline to perform the transformations when a new data is detected in the S3 bucket. This solution requires the least development effort because it leverages the native integration between EventBridge and SageMaker Pipelines, which allows you to trigger a pipeline execution based on an event rule. EventBridge can monitor the S3 bucket for new data uploads and invoke the pipeline that contains the same transformations and feature engineering steps that were defined in SageMaker Data Wrangler. The pipeline can then ingest the transformed data into the online feature store for training and inference.

The other solutions are less optimal because they require more development effort and additional services. Using AWS Lambda or AWS Step Functions would require writing custom code to invoke the SageMaker pipeline and handle any errors or retries. Using Apache Airflow would require setting up and maintaining an Airflow server and DAGs, as well as integrating with the SageMaker API.

References:

Amazon EventBridge and Amazon SageMaker Pipelines integration

Create a pipeline using a JSON specification

Ingest data into a feature group


Contribute your Thoughts:

Martha
3 months ago
I agree with Carey, Amazon EventBridge can trigger the transformations automatically with less effort.
upvoted 0 times
...
Lynette
3 months ago
Option D all the way! Easy peasy, just like my grandma's apple pie. Now, where's the feature store to put all those delicious data transformations?
upvoted 0 times
Jeannetta
2 months ago
Then you can prepare the historic data for training and inference using native integrations.
upvoted 0 times
...
Johnetta
2 months ago
Yeah, you can save the transformations to SageMaker Feature Store.
upvoted 0 times
...
Melissa
2 months ago
I think the feature store is in SageMaker, where you can save the transformations.
upvoted 0 times
...
Jill
2 months ago
Option D all the way! Easy peasy, just like my grandma's apple pie.
upvoted 0 times
...
...
Remona
3 months ago
But option A involves using AWS Lambda, which can automate the process.
upvoted 0 times
...
Carey
3 months ago
I disagree, I believe option D is more efficient.
upvoted 0 times
...
Remona
3 months ago
I think option A is the best choice.
upvoted 0 times
...
Audra
3 months ago
Personally, I'd go with Option D. Why complicate things with Lambda, Step Functions, or Airflow when EventBridge can do the job elegantly?
upvoted 0 times
Emeline
2 months ago
I agree, Option D with Amazon EventBridge seems like the most efficient choice here.
upvoted 0 times
...
Romana
2 months ago
Option D is definitely the way to go. It's the simplest solution for this scenario.
upvoted 0 times
...
...
Gennie
3 months ago
Option C with Apache Airflow seems a bit overkill for this use case. EventBridge is probably the simplest and most efficient solution.
upvoted 0 times
...
King
3 months ago
I'm partial to Option B with AWS Step Functions. It gives you a bit more control and visibility over the workflow compared to just using EventBridge.
upvoted 0 times
Bette
3 months ago
I agree, it provides more control and visibility over the workflow.
upvoted 0 times
...
Clorinda
3 months ago
Option B with AWS Step Functions sounds like a good choice.
upvoted 0 times
...
...
Coral
3 months ago
Option D looks like the most straightforward way to handle this. EventBridge can trigger the SageMaker pipeline automatically when new data arrives in the S3 bucket.
upvoted 0 times
Sabine
3 months ago
Yes, it definitely simplifies the workflow for the data scientist.
upvoted 0 times
...
Rikki
3 months ago
I agree, using Amazon EventBridge to automate the process seems efficient.
upvoted 0 times
...
Dorathy
3 months ago
Option D looks like the most straightforward way to handle this. EventBridge can trigger the SageMaker pipeline automatically when new data arrives in the S3 bucket.
upvoted 0 times
...
...

Save Cancel
az-700  pass4success  az-104  200-301  200-201  cissp  350-401  350-201  350-501  350-601  350-801  350-901  az-720  az-305  pl-300  

Warning: Cannot modify header information - headers already sent by (output started at /pass.php:70) in /pass.php on line 77