Chapter 5 Flashcards

(70 cards)

1
Q

Before Machine Learning, we program ______ to recognize an apple given images of apple.

A

logic/algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

With Machine Learning, given new images of apple, ______ make the prediction.

A

Machine Learning Models

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

In Supervised Learning, we build a machine that can extract ______ between variables.

A

relationships

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Lazy learning algorithms generalize only when ______, rather than during the training phase.

A

queried

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

During the training phase of lazy learning, the algorithm accepts the data as input but refrains from ______ on it.

A

actively training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

In lazy learning, the actual model training occurs during the ______.

A

prediction phase

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

An example of a lazy learning algorithm is the ______ algorithm.

A

K-nearest neighbors (KNN)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Eager learning algorithms process data during the ______ phase.

A

training

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Examples of eager learning algorithms include Linear Regression, Logistic Regression, Support Vector Machines, Decision Trees, and ______.

A

Artificial Neural Networks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Instance-based learning involves using the ______ to make predictions.

A

entire dataset

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

In instance-based learning, the machine compares the new data to the instances it has seen before and uses the ______ to make a prediction.

A

closest match

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

In instance-based learning, ______ is created; instead, the machine stores all of the training data.

A

no model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Instance-based learning is often used in pattern recognition, ______, and anomaly detection.

A

clustering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Model-based learning involves creating a ______ that can predict outcomes based on input data.

A

mathematical model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

In model-based learning, the model can be thought of as a set of ______ that the machine uses to make predictions.

A

rules

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Parametric models make specific ______ about the relationship between input and output data.

A

hypotheses

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

In parametric models, assumptions concern a ______ of parameters and variables that impact the model’s result.

A

fixed number

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Parametric models are easier to understand as they build a certain ______ about the data input that has to be followed.

A

hypothesis

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

Because of the assumptions they make, parametric models often need ______ to reach a certain level of accuracy.

A

less data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

If the hypothesis is met, parametric models can be more efficient and perform better than ______.

A

non-parametric

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

A limitation of parametric models is that the assumption often ______ the problem.

A

simplifies

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

Parametric models have proven to be sensitive to outliers, show limited performance to ______ problems, and struggle to adapt to new unseen data.

A

nonlinearity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Non-parametric models don’t need to make ______ about the relations between the input and output to generate an outcome.

A

assumptions

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

Non-parametric models also don’t require a certain number of ______ to be set and learned.

A

parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Studies have shown that non-parametric models perform better on ______ and are more flexible.
large datasets
26
Common non-parametric algorithms are random forests, decision trees, Support Vector Machines with non-linear kernels, k-Nearest Neighbors (k-NN) algorithm and Neural networks with ______ activation functions.
non-parametric
27
The basic benefit of non-parametric models is that they are able to catch ______ and relationships without having to follow a hypothesis.
complex patterns
28
Non-parametric models can handle ______ and noisy data effectively.
outliers
29
A limitation of non-parametric models is that they require ______ in order to generate better predictions.
more data input
30
Non-parametric models are usually harder to understand and analyze as there are no ______ about the behavior of input and output.
functional assumptions
31
A machine learning model maps from ______ to prediction, represented as f(x) -> y.
features
32
Learning has three stages: Training, ______, and Test.
Validation
33
In the training stage of learning, one aims to optimize ______.
model parameters
34
The validation stage of learning involves intermediate evaluations to ______ model.
design/select
35
The test stage of learning is for final ______.
performance evaluation
36
Model design and 'hyper parameter' tuning is performed using a ______.
validation set
37
Sometimes, you need to split 'train' into a train set for learning parameters and a ______ for checking model performance.
val set
38
The key principle of machine learning states: if Xi is similar to Xj, then yi is probably ______ to yj.
similar
39
In K-Nearest Neighbors, an arbitrary instance x is described by the feature vector , where ar(x) denotes the value of the ______ of instance x.
rth attribute
40
The distance between two instances xi and xj in KNN is defined to be d(xi, xj), calculated using the ______ distance formula.
Euclidean
41
The nearest neighbor algorithm assigns the label/target value of the ______ training features for given test features.
most similar
42
For KNN classification, prediction is usually the ______ or most common class of the returned labels.
mode
43
For KNN regression, prediction is usually the ______ of the returned values.
arithmetic mean
44
Concept learning involves acquiring ______ from specific training examples.
general concepts
45
Each concept can be thought of as a ______ function defined over a larger set.
boolean-valued
46
Implementing KNN involves three steps: Calculate Distance, Identify Nearest Neighbors, and ______.
Aggregate Nearest Neighbors
47
In KNN Classification, the algorithm aggregates the class labels of the 'k' nearest neighbors to predict the class of the current data point, choosing the ______.
most common class label
48
In KNN Regression, the algorithm calculates the ______ of the k nearest training examples.
mean value
49
In nearest-neighbor learning, the target function may be either ______ or real-valued.
discrete-valued
50
The KNN classification algorithm returns f_hat(xq) <- argmax_v_in_V Σ_i=1_k ______.
δ(v, f(xi))
51
In the Distance-Weighted Nearest Neighbor algorithm, greater weight is given to ______.
closer neighbors
52
In Distance-Weighted Nearest Neighbor Classification, wi is defined as ______.
1 / d(xq, xi)^2
53
In Distance-Weighted Nearest Neighbor Regression, f_hat(xq) is calculated as (Σ_i=1_k w_i * f(x_i)) / ______.
Σ_i=1_k w_i
54
Considering all examples in global KNN with distance weighting means the classifier will run ______.
more slowly
55
Nearest neighbour classifier is also known as lazy learning or ______ or instance-based learning.
memory-based learning
56
Nearest neighbour classifiers are useful for large ______ datasets, when trained models become obsolete in a short time.
fast-changing
57
In a nearest neighbour classifier, the training phase involves ______.
do nothing
58
A ______ shows the decision boundaries in a nearest neighbour classifier.
Voronoi diagram
59
The distance-weighted k-NEAREST NEIGHBOR algorithm is robust to ______ and is effective for large numbers of training examples.
noisy training data
60
A limitation of KNN is the assumption that classification is most similar to instances nearby in ______.
Euclidean distance
61
Another limitation of KNN is that the distance between instances is calculated based on ______ of the instance.
all attributes
62
If only a few attributes are relevant, the presence of many irrelevant attributes can ______ the distance metric in KNN.
mislead
63
One solution to KNN's attribute limitation is to ______ each attribute differently when calculating distance.
weight
64
Weighting attributes differently in KNN corresponds to ______ the axes in the Euclidean space.
stretching
65
A more drastic alternative to attribute weighting in KNN is to completely eliminate the ______ attributes from the instance space.
least relevant
66
An additional practical issue in applying k-NEAREST NEIGHBOR is efficient ______.
memory indexing
67
One indexing method for KNN is the ______.
kd-tree
68
In statistical pattern recognition terminology, regression means approximating a ______ target function.
real-valued
69
In statistical pattern recognition, residual is the error ______ in approximating the target function.
f_hat(x) - f(x)
70
In statistical pattern recognition, the kernel function is the function of distance used to determine the ______ of each training example.
weight