19. Learning by Example (ML) Flashcards

(22 cards)

1
Q

What is learning by example?

A

Learning by example involves inferring a function or decision boundary from input-output pairs (training data).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is the goal of supervised learning?

A

To learn a function that maps inputs to outputs, minimizing error on unseen examples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is a hypothesis in machine learning?

A

A candidate function from a hypothesis space that approximates the target function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is the hypothesis space?

A

The set of all functions a learning algorithm can choose from to approximate the target function.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does it mean to generalize?

A

To perform well on unseen data, not just the training set.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is overfitting?

A

When a model fits training data too closely, capturing noise and failing to generalize.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is underfitting?

A

When a model is too simple to capture the underlying pattern of the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the difference between training and test error?

A

Training error measures fit to seen data; test error measures performance on unseen data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the version space?

A

The set of hypotheses consistent with all training examples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is inductive bias?

A

Assumptions a learner uses to generalize beyond the training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the inductive learning hypothesis?

A

Any hypothesis that approximates the target function well over the training set will also do well on unseen data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is a consistent learner?

A

A learner that only outputs hypotheses consistent with all training examples.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the Find-S algorithm?

A

It finds the most specific hypothesis consistent with the training data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are limitations of Find-S?

A

It only works for conjunctive hypotheses and ignores inconsistent data or noise.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the Candidate Elimination algorithm?

A

It maintains the version space by updating specific (S) and general (G) boundaries.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What happens to the version space with more data?

A

It shrinks, ideally converging toward the target concept.

17
Q

What are the S and G sets in Candidate Elimination?

A

S contains the most specific hypotheses; G contains the most general ones consistent with data.

18
Q

How are S and G updated in Candidate Elimination?

A

S is generalized on positive examples; G is specialized on negative examples.

19
Q

What causes noise to be problematic in version space learning?

A

It can eliminate all hypotheses, as no consistent function may exist.

20
Q

What is inductive learning vulnerable to?

A

Noise, limited data, and incorrect inductive bias.

21
Q

How can hypothesis space design affect learning?

A

Too large → overfitting; too small → underfitting.

22
Q

Why is inductive learning considered impossible without bias?

A

Because multiple hypotheses may explain training data—bias is needed to prefer one.