Amazon AWS Certified Machine Learning - Specialty
Prev

There are 204 results

Next
#101 (Accuracy: 100% / 3 votes)
A company operates large cranes at a busy port The company plans to use machine learning (ML) for predictive maintenance of the cranes to avoid unexpected breakdowns and to improve productivity.

The company already uses sensor data from each crane to monitor the health of the cranes in real time. The sensor data includes rotation speed, tension, energy consumption, vibration, pressure, and temperature for each crane. The company contracts AWS ML experts to implement an ML solution.

Which potential findings would indicate that an ML-based solution is suitable for this scenario? (Choose two.)
  • A. The historical sensor data does not include a significant number of data points and attributes for certain time periods.
  • B. The historical sensor data shows that simple rule-based thresholds can predict crane failures.
  • C. The historical sensor data contains failure data for only one type of crane model that is in operation and lacks failure data of most other types of crane that are in operation.
  • D. The historical sensor data from the cranes are available with high granularity for the last 3 years.
  • E. The historical sensor data contains most common types of crane failures that the company wants to predict.
#102 (Accuracy: 100% / 6 votes)
A company is building custom deep learning models in Amazon SageMaker by using training and inference containers that run on Amazon EC2 instances. The company wants to reduce training costs but does not want to change the current architecture. The SageMaker training job can finish after interruptions. The company can wait days for the results.

Which combination of resources should the company use to meet these requirements MOST cost-effectively? (Choose two.)
  • A. On-Demand Instances
  • B. Checkpoints
  • C. Reserved Instances
  • D. Incremental training
  • E. Spot instances
#103 (Accuracy: 100% / 5 votes)
A machine learning (ML) engineer at a bank is building a data ingestion solution to provide transaction features to financial ML models. Raw transactional data is available in an Amazon Kinesis data stream.

The solution must compute rolling averages of the ingested data from the data stream and must store the results in Amazon SageMaker Feature Store.
The solution also must serve the results to the models in near real time.

Which solution will meet these requirements?
  • A. Load the data into an Amazon S3 bucket by using Amazon Kinesis Data Firehose. Use a SageMaker Processing job to aggregate the data and to load the results into SageMaker Feature Store as an online feature group.
  • B. Write the data directly from the data stream into SageMaker Feature Store as an online feature group. Calculate the rolling averages in place within SageMaker Feature Store by using the SageMaker GetRecord API operation.
  • C. Consume the data stream by using an Amazon Kinesis Data Analytics SQL application that calculates the rolling averages. Generate a result stream. Consume the result stream by using a custom AWS Lambda function that publishes the results to SageMaker Feature Store as an online feature group.
  • D. Load the data into an Amazon S3 bucket by using Amazon Kinesis Data Firehose. Use a SageMaker Processing job to load the data into SageMaker Feature Store as an offline feature group. Compute the rolling averages at query time.
#104 (Accuracy: 93% / 12 votes)
A company is building a new version of a recommendation engine. Machine learning (ML) specialists need to keep adding new data from users to improve personalized recommendations. The ML specialists gather data from the users' interactions on the platform and from sources such as external websites and social media.
The pipeline cleans, transforms, enriches, and compresses terabytes of data daily, and this data is stored in Amazon S3.
A set of Python scripts was coded to do the job and is stored in a large Amazon EC2 instance. The whole process takes more than 20 hours to finish, with each script taking at least an hour. The company wants to move the scripts out of Amazon EC2 into a more managed solution that will eliminate the need to maintain servers.
Which approach will address all of these requirements with the LEAST development effort?
  • A. Load the data into an Amazon Redshift cluster. Execute the pipeline by using SQL. Store the results in Amazon S3.
  • B. Load the data into Amazon DynamoDB. Convert the scripts to an AWS Lambda function. Execute the pipeline by triggering Lambda executions. Store the results in Amazon S3.
  • C. Create an AWS Glue job. Convert the scripts to PySpark. Execute the pipeline. Store the results in Amazon S3.
  • D. Create a set of individual AWS Lambda functions to execute each of the scripts. Build a step function by using the AWS Step Functions Data Science SDK. Store the results in Amazon S3.
#105 (Accuracy: 92% / 5 votes)
A machine learning (ML) specialist wants to bring a custom training algorithm to Amazon SageMaker. The ML specialist implements the algorithm in a Docker container that is supported by SageMaker.

How should the ML specialist package the Docker container so that SageMaker can launch the training correctly?
  • A. Specify the server argument in the ENTRYPOINT instruction in the Dockerfile.
  • B. Specify the training program in the ENTRYPOINT instruction in the Dockerfile.
  • C. Include the path to the training data in the docker build command when packaging the container.
  • D. Use a COPY instruction in the Dockerfile to copy the training program to the /opt/ml/train directory.
#106 (Accuracy: 100% / 5 votes)
An ecommerce company sends a weekly email newsletter to all of its customers. Management has hired a team of writers to create additional targeted content. A data scientist needs to identify five customer segments based on age, income, and location. The customers' current segmentation is unknown. The data scientist previously built an XGBoost model to predict the likelihood of a customer responding to an email based on age, income, and location.
Why does the XGBoost model NOT meet the current requirements, and how can this be fixed?
  • A. The XGBoost model provides a true/false binary output. Apply principal component analysis (PCA) with five feature dimensions to predict a segment.
  • B. The XGBoost model provides a true/false binary output. Increase the number of classes the XGBoost model predicts to five classes to predict a segment.
  • C. The XGBoost model is a supervised machine learning algorithm. Train a k-Nearest-Neighbors (kNN) model with K = 5 on the same dataset to predict a segment.
  • D. The XGBoost model is a supervised machine learning algorithm. Train a k-means model with K = 5 on the same dataset to predict a segment.
#107 (Accuracy: 100% / 4 votes)
A Data Scientist needs to analyze employment data. The dataset contains approximately 10 million observations on people across 10 different features. During the preliminary analysis, the Data Scientist notices that income and age distributions are not normal. While income levels shows a right skew as expected, with fewer individuals having a higher income, the age distribution also shows a right skew, with fewer older individuals participating in the workforce.
Which feature transformations can the Data Scientist apply to fix the incorrectly skewed data? (Choose two.)
  • A. Cross-validation
  • B. Numerical value binning
  • C. High-degree polynomial transformation
  • D. Logarithmic transformation
  • E. One hot encoding
#108 (Accuracy: 100% / 5 votes)
A Machine Learning Specialist needs to move and transform data in preparation for training. Some of the data needs to be processed in near-real time, and other data can be moved hourly. There are existing Amazon EMR MapReduce jobs to clean and feature engineering to perform on the data.
Which of the following services can feed data to the MapReduce jobs? (Choose two.)
  • A. AWS DMS
  • B. Amazon Kinesis
  • C. AWS Data Pipeline
  • D. Amazon Athena
  • E. Amazon ES
#109 (Accuracy: 100% / 5 votes)
A data scientist has developed a machine learning translation model for English to Japanese by using Amazon SageMaker's built-in seq2seq algorithm with
500,000 aligned sentence pairs.
While testing with sample sentences, the data scientist finds that the translation quality is reasonable for an example as short as five words. However, the quality becomes unacceptable if the sentence is 100 words long.
Which action will resolve the problem?
  • A. Change preprocessing to use n-grams.
  • B. Add more nodes to the recurrent neural network (RNN) than the largest sentence's word count.
  • C. Adjust hyperparameters related to the attention mechanism.
  • D. Choose a different weight initialization type.
#110 (Accuracy: 100% / 4 votes)
A company hosts a machine learning (ML) dataset repository on Amazon S3. A data scientist is preparing the repository to train a model. The data scientist needs to redact personally identifiable information (PH) from the dataset.

Which solution will meet these requirements with the LEAST development effort?
  • A. Use Amazon SageMaker Data Wrangler with a custom transformation to identify and redact the PII.
  • B. Create a custom AWS Lambda function to read the files, identify the PII. and redact the PII
  • C. Use AWS Glue DataBrew to identity and redact the PII
  • D. Use an AWS Glue development endpoint to implement the PII redaction from within a notebook