Concepts week 4 - Generalisation Flashcards

1
Q

What is Ockham’s razor law?

A

“More things should not be used than are necessary.”
If there are two explanations of a hypothesis - the simper one might be the right one.

The less complex a ML model, the more likely that a good empirical result is not just due to the peculiarities of the sample.

Razor is a metaphor. Shaving off all unnecessary things.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is a test set methodology?

A

In contrast to generalisation theory - we use empirical approach. We take one set of data from a distribution and train the model on it - the training set. Then we use another set of data (test set) from the distribution to see how good the model is. If our model does a good job on test data - we have a good feeling that our model generalises well. ==> Generalisation

A machine learning model aims to make good predictions on new, previously unseen data. But if you are building a model from your data set, how would you get the previously unseen data? Well, one way is to divide your data set into two subsets:

training set—a subset to train a model.
test set—a subset to test the model.
Good performance on the test set is a useful indicator of good performance on the new data in general, assuming that:

The test set is large enough.
You don’t cheat by using the same test set over and over.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Describe Generalisation.

A

A ML model’s ability to make correct predictions on new, previously unseen data as opposed to the data used to train the model.

We train our model on a training set and verify the model with our test set. If our model does a good job on test set - we have a good feeling that our model generalises well on new and unseen data ==> Generalisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Describe overfit model

A

An overfit model gets a low loss during training but does a poor job predicting new data.

Overfitting is caused by making a model more complex than necessary (Ockham’s razor).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly