MLA FA Flashcards
(32 cards)
The two phases of supervised ML process: Training, ________.
PREDICTING
Logistic Regression is an example of a regression algorithm.
FALSE
The _____ refers to the error from having wrong / too simple assumptions in the learning algorithm.
BIAS
Its primary objective is to map the input variable with the output variable.
Supervised Learning
These concepts helps to understand how well a model performs: Overfitting, Underfitting, _________.
GENERALIZATION
If your model performs well on the training set but poorly on the validation set.
Overfitting
When the model fits too closely to the training dataset.
Generalization
In k-NN, High Model Complexity is underfitting.
FALSE
K-nearest neighbors make a prediction for a new data point by finding the data that match from the training dataset.
FALSE
In k-NN, Low Model Complexity is:
Underfitting
In k-NN, when you choose a small value of k (e.g., k=1), the model becomes more complex.
TRUE
There is a regression variant of the k-nearest neighbors algorithm.
TRUE
In k-NN, High Model Complexity is:
Overfitting
There is a regression variant of the k-nearest neighbors algorithm.
FALSE
When comparing training set and test set scores, we find that we predict very accurately on the training set, but the R2 on the test set is much worse. This is a sign of underfitting.
FALSE
When comparing training set and test set scores, we find that we predict very accurately on the training set, but the R2 on the test set is much worse. This is a sign of:
Overfitting
Ridge regression is a linear regression model that controls complexity to avoid overfitting.
TRUE
Lasso uses L1 Regularization.
TRUE
The ‘slope’ parameter is also called _______ or coefficients.
Weight
Linear Regression is also known as Ordinal Least Squares.
FALSE
In supervised learning, market trend analysis is an example of:
REGRESSION
A model that performs poorly on both training and new data because it hasn’t learned enough from the training data.
Underfitting
Classification algorithms address classification problems where the output variable is categorical.
TRUE
This refers to the error resulting from sensitivity to the noise in the training data.
Not in the options