L5 - Regression - Bias-Variance Trade Off Flashcards

1
Q

Define bias…

A

Bias refers to the extent that a model produces errors due to under-fitting. High bias means that high model error is encountered due to underfitting.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Define variance…

A

Variance refers to the extent that a model has been overfit. High variance means a model will not perform accurately on unseen data due to overfitting the training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is the ideal bias-variance trade off for a regression model?

A

Low bias and low variance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the 3 causes of model error? Define each…

A

Bias - The extent to which a model predicts wrong.

Variance - The extent to which a model has learned data, and thus is overfit.

Irreducible Error - Refers to the unavoidable randomness of real world data that can cause errors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How can we tune bias and variance?

A

To increase bias we must reduce variance and vice versa.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is another way of wording the bias-variance trade off of a model?

A

Bias-variance trade off can also be called a Complexity trade off.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

If we achieve the optimal bias-variance ratio, what model complexity does this give us?

A

Optimal model complexity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What technique can we use to prevent our model from overfitting, and to encourage generalisation?

A

Linear Model Regularisation (Shrinkage)

This reduces the magnitude of the coefficients in the polynomial regression model.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the objective of Linear Model Regularisation?

A

Establish a trade off between bias and variance, resulting in optimal model complexity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does Linear Model Regularisation work?

A
  • Introduces a penalty to the models loss function
  • Penalty can be increased or decreased to increase or decrease complexity.
  • We want to push all coefficients towards 0.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What happens if the Tuning parameter of the Regression Models cost function is increased?

A
  • Results in less regularisation, which leans to overfitting ( increases variance )
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What happens if the Tuning parameter of the Regression Models cost function is decreased?

A
  • Results in more regularisation
  • Decreased model complexity which leans to underfitting ( increases bias )
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is L1 regularisation?

A
  • L1 ( LASSO ) applies a penalty value that is proportional to the sum of the absolute coefficient values.
  • Prevents overfitting and performs feature selection.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is L2 regularisation?

A
  • Ridge Regression
  • Prevents overfitting and improves model stability.
  • A penalty is applied that is proportional to the squared coefficient values.
  • Penalty imposes a bias, and thus can be used to control the bias-variance trade off.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

How do we establish the penalty value?

A
  • Through cross-validation, to find the lambda with the lowest variance.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What do L1 and L2 have in common?

A

Both push coefficients towards 0.

17
Q

Why do L1 and L2 push coefficients to 0?

A
  • Helps prevent overfitting
  • Improves generalisation
  • Reduces variance.
18
Q

How does L1 Regularisation perform feature selection?

A

By pushing coefficients to 0, the features with non-zero coefficients are known to be important.