A manufacturing company uses machine learning (ML) models to detect quality issues. The models use images that are
taken of the company's product at the end of each production step. The company has thousands of machines at the
production site that generate one image per second on average.
The company ran a successful pilot with a single manufacturing machine. For the pilot, ML specialists used an industrial PC
that ran AWS IoT Greengrass with a long-running AWS Lambda function that uploaded the images to Amazon S3. The
uploaded images invoked a Lambda function that was written in Python to perform inference by using an Amazon
SageMaker endpoint that ran a custom model. The inference results were forwarded back to a web service that was hosted
at the production site to prevent faulty products from being shipped.
The company scaled the solution out to all manufacturing machines by installing similarly configured industrial PCs on each
production machine. However, latency for predictions increased beyond acceptable limits. Analysis shows that the internet
connection is at its capacity limit.
How can the company resolve this issue MOST cost-effectively?
D
A retail company is selling products through a global online marketplace. The company wants to use machine learning (ML)
to analyze customer feedback and identify specific areas for improvement. A developer has built a tool that collects customer
reviews from the online marketplace and stores them in an Amazon S3 bucket. This process yields a dataset of 40 reviews.
A data scientist building the ML models must identify additional sources of data to increase the size of the dataset.
Which data sources should the data scientist use to augment the dataset of reviews? (Choose three.)
B D F
A company wants to classify user behavior as either fraudulent or normal. Based on internal research, a machine learning
specialist will build a binary classifier based on two features: age of account, denoted by x, and transaction month, denoted
by y. The class distributions are illustrated in the provided figure. The positive class is portrayed in red, while the negative
class is portrayed in black.
Which model would have the HIGHEST accuracy?
C
A Data Scientist needs to create a serverless ingestion and analytics solution for high-velocity, real-time streaming data.
The ingestion process must buffer and convert incoming records from JSON to a query-optimized, columnar format without
data loss. The output datastore must be highly available, and Analysts must be able to run SQL queries against the data and
connect to existing business intelligence dashboards.
Which solution should the Data Scientist build to satisfy the requirements?
A
A company wants to predict the sale prices of houses based on available historical sales data. The target variable in the
companys dataset is the sale price. The features include parameters such as the lot size, living area measurements, non-
living area measurements, number of bedrooms, number of bathrooms, year built, and postal code. The company wants to
use multi-variable linear regression to predict house sale prices.
Which step should a machine learning specialist take to remove features that are irrelevant for the analysis and reduce the
models complexity?
D
A machine learning specialist stores IoT soil sensor data in Amazon DynamoDB table and stores weather event data as
JSON files in Amazon S3. The dataset in DynamoDB is 10 GB in size and the dataset in Amazon S3 is 5 GB in size. The
specialist wants to train a model on this data to help predict soil moisture levels as a function of weather events using
Amazon SageMaker.
Which solution will accomplish the necessary transformation to train the Amazon SageMaker model with the LEAST amount
of administrative overhead?
C
A Machine Learning Specialist must build out a process to query a dataset on Amazon S3 using Amazon Athena. The
dataset contains more than 800,000 records stored as plaintext CSV files. Each record contains 200 columns and is
approximately 1.5 MB in size. Most queries will span 5 to 10 columns only.
How should the Machine Learning Specialist transform the dataset to minimize query runtime?
A
Explanation:
Using compressions will reduce the amount of data scanned by Amazon Athena, and also reduce your S3 bucket storage.
Its a Win-Win for your AWS bill. Supported formats: GZIP, LZO, SNAPPY (Parquet) and ZLIB.
Reference: https://www.cloudforecast.io/blog/using-parquet-on-athena-to-save-money-on-aws/
When submitting Amazon SageMaker training jobs using one of the built-in algorithms, which common parameters MUST be
specified? (Choose three.)
A E F
A Machine Learning Specialist is deciding between building a naive Bayesian model or a full Bayesian network for a
classification problem. The Specialist computes the Pearson correlation coefficients between each feature and finds that
their absolute values range between 0.1 to 0.95.
Which model describes the underlying data in this situation?
C
A company wants to classify user behavior as either fraudulent or normal. Based on internal research, a Machine Learning
Specialist would like to build a binary classifier based on two features: age of account and transaction month. The class
distribution for these features is illustrated in the figure provided.
Based on this information, which model would have the HIGHEST recall with respect to the fraudulent class?
C
A real-estate company is launching a new product that predicts the prices of new houses. The historical data for the
properties and prices is stored in .csv format in an Amazon S3 bucket. The data has a header, some categorical fields, and
some missing values. The companys data scientists have used Python with a common open-source library to fill the missing
values with zeros. The data scientists have dropped all of the categorical fields and have trained a model by using the open-
source linear regression algorithm with the default parameters.
The accuracy of the predictions with the current model is below 50%. The company wants to improve the model performance
and launch the new product as soon as possible.
Which solution will meet these requirements with the LEAST operational overhead?
A
Explanation:
Reference: https://docs.aws.amazon.com/deep-learning-containers/latest/devguide/deep-learning-containers-ecs-setup.html
A Data Scientist is developing a binary classifier to predict whether a patient has a particular disease on a series of test
results. The Data Scientist has data on 400 patients randomly selected from the population. The disease is seen in 3% of the
population.
Which cross-validation strategy should the Data Scientist adopt?
B
A data scientist is training a text classification model by using the Amazon SageMaker built-in BlazingText algorithm. There
are 5 classes in the dataset, with 300 samples for category A, 292 samples for category B, 240 samples for category C, 258
samples for category D, and 310 samples for category E.
The data scientist shuffles the data and splits off 10% for testing. After training the model, the data scientist generates
confusion matrices for the training and test sets.
What could the data scientist conclude form these results?
B
A Machine Learning Specialist kicks off a hyperparameter tuning job for a tree-based ensemble model using Amazon
SageMaker with Area Under the ROC Curve (AUC) as the objective metric. This workflow will eventually be deployed in a
pipeline that retrains and tunes hyperparameters each night to model click-through on data that goes stale every 24 hours.
With the goal of decreasing the amount of time it takes to train these models, and ultimately to decrease costs, the Specialist
wants to reconfigure the input hyperparameter range(s).
Which visualization will accomplish this?
B
A gaming company has launched an online game where people can start playing for free, but they need to pay if they
choose to use certain features. The company needs to build an automated system to predict whether or not a new user will
become a paid user within 1 year. The company has gathered a labeled dataset from 1 million users.
The training dataset consists of 1,000 positive samples (from users who ended up paying within 1 year) and 999,000
negative samples (from users who did not use any paid features). Each data sample consists of 200 features including user
age, device, location, and play patterns.
Using this dataset for training, the Data Science team trained a random forest model that converged with over 99% accuracy
on the training set. However, the prediction results on a test dataset were not satisfactory
Which of the following approaches should the Data Science team take to mitigate this issue? (Choose two.)
C D