MLA-C01 Flashcards

Cert Exam Study (111 cards)

1
Q

Before you can use auto scaling, you must have already created an Amazon SageMaker ______________.

A

model endpoint.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

You can have multiple model _____________for the same endpoint.

A

versions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Amazon SageMaker ____________ provides tools to help explain how machine learning (ML) models make predictions.

A

Clarify

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

An ____________can be thought of as the answer to a Why question that helps humans understand the cause of a prediction.

A

explanation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

On AWS, AI/ML practitioners can use Amazon Sagemaker ____________, which uses Shapley values to help answer how different variables influence model behavior.

A

Clarify

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Debug model output tensors from machine learning training jobs in real time and detect non-converging issues using Amazon SageMaker ____________.

A

Debugger

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

___________is the extent to which you can explain the internal mechanics of an ML or deep learning system in human terms.

A

Explainability

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Amazon SageMaker _________produces metrics that measure the predictive quality of machine learning model candidates.

A

Autopilot

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

The ratio of the number of correctly classified items to the total number of (correctly and incorrectly) classified items.

A

Accuracy

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

measures how well an algorithm predicts the true positives (TP) out of all of the positives that it identifies.

A

Precision

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

uses natural language processing (NLP) to extract insights about the content of documents. It develops insights by recognizing the entities, key phrases, language, sentiments, and other common elements in a document.

A

Amazon Comprehend

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

a text translation service that uses advanced machine learning technologies to provide high-quality translation on demand. use to translate unstructured text documents or to build applications that work in multiple languages.

A

Amazon Translate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

a fully managed, automatic speech recognition (ASR) service that makes it easy for developers to add speech to text capabilities to their applications.

A

Amazon Transcribe

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

a cloud service that converts text into lifelike speech. You can use to develop applications that increase engagement and accessibility.

A

Amazon Polly

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

a cloud-based image and video analysis service that makes it easy to add advanced computer vision capabilities to your applications.

A

Amazon Rekognition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

a fully managed service that uses statistical and machine learning algorithms to deliver highly accurate time-series forecasts.

A

Amazon Forecast

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

an AWS service for building conversational interfaces for applications using voice and text.

A

Amazon Lex

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

a fully managed machine learning service that uses your data to generate item recommendations for your users. It can also generate user segments based on the users’ affinity for certain items or item metadata.

A

Amazon Personalize

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

a machine learning (ML) service that automatically extracts text, handwriting, and data from scanned documents.

A

Amazon Textract

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

an intelligent search service that uses natural language processing and advanced machine learning algorithms to return specific answers to search questions from your data.

A

Amazon Kendra

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

allows you to conduct a human review of machine learning (ML) systems to guarantee precision.

A

Amazon Augmented AI (A2I)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

uses machine learning (ML) to make it easier for customers to accurately detect anomalies in their metrics.

A

Amazon Lookout for Metrics

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

a fully managed service enabling customers to identify potentially fraudulent activities and catch more online fraud faster.

A

Amazon Fraud Detector

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

a fully managed, generative-AI powered assistant that you can configure to answer questions, provide summaries, generate content, and complete tasks based on your enterprise data.

A

Amazon Q Business

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Amazon Polly is the Opposite of Amazon ____________.
Transcribe
26
______________measures how many actual positives were predicted as positive.
Recall
27
_____________is the harmonic mean of precision and recall.
F1-measure
28
It measures the ability of the model to predict a higher score for positive examples as compared to negative examples.
AUC (Area Under Curve)
29
_________is a method used in machine learning to reduce errors in predictive data analysis.
Boosting
30
____________improves machine models' predictive accuracy and performance by converting multiple weak learners into a single strong learning model.
Boosting
31
____________ are data structures in machine learning that work by dividing the dataset into smaller and smaller subsets based on their features
Decision trees
32
Boosting creates an ____________model by combining several weak decision trees sequentially.
ensemble
33
In ________, data scientists improve the accuracy of weak learners by training several of them at once on multiple datasets. In contrast, boosting trains weak learners one after another.
bagging
34
__________is a popular and efficient open-source implementation of the gradient boosted trees algorithm.
XGBoost
35
___________boosting is a supervised learning algorithm that tries to accurately predict a target variable by combining multiple estimates from a set of simpler models.
Gradient
36
Amazon SageMaker _____________ reduces data prep time for tabular, image, and text data from weeks to minutes.
Data Wrangler
37
With SageMaker ________________ you can simplify data preparation and feature engineering through a visual and natural language interface.
Data Wrangler
38
Sagemaker ____________ a no-code ML tool that helps business analysts generate accurate ML predictions without having to write code or without requiring any ML experience.
Canvas
39
Amazon SageMaker ____________ is a fully managed, purpose-built repository to store, share, and manage features for machine learning (ML) models.
Feature Store
40
___________are inputs to ML models used during training and inference.
Features
41
SageMaker ____________ tags and indexes feature groups so they are easily discoverable through the visual interface of Amazon SageMaker Studio.
Feature Store
42
Amazon SageMaker ____________ offers the most comprehensive set of human-in-the-loop capabilities, allowing you to harness the power of human feedback across the ML lifecycle to improve the accuracy and relevancy of models.
Ground Truth
43
You can complete a variety of human-in-the-loop tasks with SageMaker ___________, from data generation and annotation to model review, customization, and evaluation, either through a self-service or an AWS-managed offering.
Ground Truth
44
SageMaker _________helps identify potential bias during data preparation without writing code.
Clarify
45
SQL function used for anomaly detection on numeric columns in a stream
RANDOM_CUT_FOREST
46
is derived from “Linux” and “cluster
Lustre
47
a type of parallel distributed file system, for large-scale computing
Lustre
48
a fully managed Windows file system share drive
FSx for Windows File Server
49
a network drive you can attach to your instances while they run
EBS
50
Managed NFS (network file system) that can be mounted on many EC2
EFS
51
Data Warehouse vs Data Lake. Warehouse is ________________. Lake is ___________
Structured, Unstructured
52
Binary format that stores both the data and its schema
AVRO
53
Columnar storage format optimized for analytics.
Parquet
54
Find K “nearest” (most similar) rows and average their values
K Nearest Neighbor (KNN)
55
Find linear or non-linear relationships between the missing feature and other features
Regression
56
Duplicate samples from the minority class
Oversampling
57
Instead of creating more positive samples, remove negative ones
Undersampling
58
measures how “spread-out” the data is
Variance
59
________________ 𝜎 is just the square root of the variance.
Standard Deviation
60
Data points that lie more than one ___________________ from the mean can be considered unusual.
Standard Deviation
61
Bucket observations together based on ranges of values.
Binning
62
Create “buckets” for every category * The bucket for your category has a 1, all others have a 0
One Hot Encoding
63
______________ for deploying to edge devices
SageMaker NEO
64
___________values are the algorithm used to determine the contribution of each feature toward a model’s predictions
Shapley
65
Used on the final output layer of a multi-class classification problem
Softmax
66
Choosing an activation function: For multiple classification, use _________on the output layer
softmax
67
Choosing an activation function: ________do well with Tanh
RNN’s
68
Choosing an activation function: For everything else
Start with ReLU
69
Choosing an activation function: _________for really deep networks
Swish
70
When you have data that doesn’t neatly align into columns * Images that you want to find features within * Machine translation * Sentence classification * Sentiment analysis
Convlution Neural Network (CNN)
71
RNN’s: what are they for?
Time-series data
72
When you want to predict future behavior based on past behavior
Recurrent Neural Network (RNN)
73
Sequence to sequence, Sequence to vector, Vector to sequence, Encoder -> Decoder
RNN topologies
74
_________batch sizes tend to not get stuck in local minima
Small
75
____________batch sizes can converge on the wrong solution at random
Large
76
_________learning rates can overshoot the correct solution
Large
77
____________learning rates increase training time
Small
78
* ________________techniques are intended to prevent overfitting.
Regularization
79
Preventing overfitting in ML in general * A regularization term is added as weights are learned
L1 and L2 Regularization
80
* L1: sum of _______________
weights
81
L2: sum of ______________
square of weights
82
We need to understand true positives and true negative, as well as false positives and false negatives.
confusion matrix
83
Percent of positives rightly predicted
Recall
84
AKA Correct Positives
Precision
85
Plot of true positive rate (recall) vs. false positive rate at various threshold settings.
ROC Curve
86
The area under the ROC curve is… wait for it..
AUC
87
Generate N new training sets by random sampling with replacement
Bagging
88
Training is sequential; each classifier takes into account the previous one’s success.
Boosting
89
Define the hyperparameters you care about and the ranges you want to try, and the metrics you are optimizing for
Automatic Model Tuning
90
Don’t optimize too many hyperparameters at once * Limit your ranges to as small a range as possible * Use logarithmic scales when appropriate * Don’t run too many training jobs concurrently * This limits how well the process can learn as it goes * Make sure training jobs running on multiple instances report the correct objective metric in the end
Automatic Model Tuning: Best Practices
91
Stop training in a tuning job early if it is not improving the objective significantly
Early Stopping
92
Uses one or more previous tuning jobs as a starting point
Warm Start
93
Automates: * Algorithm selection * Data preprocessing * Model tuning * All infrastructure * It does all the trial & error for you
SageMaker Autopilot
94
Visual IDE for machine learning
SageMaker Studio
95
Create and share Jupyter notebooks with SageMaker Studio * Switch between hardware configurations (no infrastructure to manage)
SageMaker Notebooks
96
Organize, capture, compare, and search your ML jobs
SageMaker Experiments
97
Saves internal model state at periodical intervals * Gradients / tensors over time as a model is trained * Define rules for detecting unwanted conditions while training * A debug job is run for each rule you configure * Logs & fires a CloudWatch event when the rule is hit
SageMaker Debugger
98
Catalog your models, manage model versions * Associate metadata with models * Manage approval status of a model * Deploy models to production * Automate deployment with CI/CD
SageMaker Model Registry
99
___________________ is a visualization toolkit for Tensorflow or PyTorch * Visualize loss and accuracy * Visualize model graph * View histograms of weight, biases over time * Project embeddings to lower dimensions * Profiling
Tensorboard
100
Compile & optimize training jobs on GPU instances * Can accelerate training up to 50% * Converts models into hardware-optimized instructions * Tested with Hugging Face transformers library, or bring your own model * Incompatible with SageMaker distributed training libraries
SageMaker Training Compiler
101
Retain and re-use provisioned infrastructure * Useful if repeatedly training a model to speed things up * Use by setting KeepAlivePeriodInSeconds in your training job’s resource config
Warm Pools
102
Creates snapshots during your training * You can re-start from these points if necessary * Or use them for troubleshooting, to analyze the model at different points * Automatic synchronization with S3 (from /opt/ml/checkpoint)
Checkpointing
103
Run automatically when using ml.g or ml.p instance types * Replaces any faulty instances * Runs GPU health checks * Ensures NVidia Collective Communication Library is working
Cluster Health Checks and Automatic Restarts
104
You can of course run multiple training jobs in parallel * “job parallelism” * Individual training can also be parallelized * Distributed data parallelism * Distributed model parallelism
Distributed Training
105
Network device attached to your SageMaker instances * Makes better use of your bandwidth * Promises performance of an onpremises High Performance Computing (HPC) cluster in the cloud
Elastic Fabric Adapter (EFA)
106
______________ produces a weighted average of all token embeddings. The magic is in computing the attention weights.
Self-attention
107
A mask can be applied to prevent tokens from “peeking” into future tokens (words)
Masked Self-Attention
108
Chat! * Question answering * Text classification * i.e., sentiment analysis * Named entity recognition * Summarization * Translation * Code generation * Text generation * i.e., automated customer service
Applications of Transformers
109
Tokenization, token encoding * Token embedding * Captures semantic relationships between tokens, token similarities * Positional encoding * Captures the position of the token in the input relative to other nearby tokens * Uses an interleaved sinusoidal function so it works on any length
LLM Input processing
110
The stack of decoders outputs a vector at the end * Multiply this with the token embeddings * This gives you probabilities (logits) of each token being the right next token (word) in the sequence
LLM Output processing
111