Chapter 7 Flashcards

(18 cards)

1
Q

Rational behind Random Forests

A

“Wisdom of the crowd”
* Aggregate group of predictors (ensembles) to get better prediction
* Ensembles can be applied to different classifiers and regressors
* Random Forest uses many Decision Trees
– One of the most powerful Machine Learning algorithms

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Voting Classifiers

A

Train different classifiers (algorithms) on same data
Then predict by majority vote

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Weak learners of even 51% accuracy can still preform 75% accuracy with ensemble

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How do Voting classifiers do against individual classifiers

A

Voting classifier outperforms any individual classifier

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

When would you use Bagging and Pasting

A

When the same algorithm is used for all the voting classifiers

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How do Bagging and Pasting work

A

Use same algorithm, but different training data
– Choose random subset from training set

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

bagging

A

With replacement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

pasting

A

Without replacement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

How do Bagging and Pasting models predict

A

Aggregation of each predictor’s output:
* Statistical mode (most frequent prediction) for classification
* Average for regression
– Ensemble has same bias but lower variance than one predictor trained on data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How does Bagging and Pasting work?

A

Training and
predictions can
be done in
parallel on
different CPU
cores

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is Out-of-Bag Evaluation (OOB)

A

Bagging classifier samples m instances from training set of size m
– Sampling with replacement: some training instances will not be picked
Oob instances can be used for evaluation
– No need for separate validation or cross-validation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Does Random Forests uses bagging or pasting?

A

bagging

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

How can Random Forests can measure relative importance of each feature

A

Measure of how much tree nodes that use feature reduce impurity
– Weighted average across all trees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Explain Boosting

A

“Boosting” or “hypothesis boosting” combines weak learners
– Training of learners is done sequentially
– Each learner is trying to correct its predecessors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Most popular boosting methods

A

– AdaBoost (“Adaptive Boosting)
– Gradient Boosting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How does ada boost work

A

New predictor corrects
predecessor by paying
attention to outliers
– More focus on training
instances where
underfitting occurred
– Relative weight of
misclassified instances
is increased in next
iteration

17
Q

How does Gradient Boosting work

A
  • Also uses sequence of predictors
    – Instead of tweaking instance weights as in AdaBoost, fits to residual errors
  • When using Decision Trees as base estimators
    – “Gradient Tree Boosting” or “Gradient Boosted Regression Tree” (GBRT)
18
Q

Explain Stacking

A

“Stacking” or “Stacked
Generalization” trains
aggregation function
– Final predictor that aggregates
predictors is called “blender” or
“meta learner”
Training of blender
based on “hold-out set”
– Reserve some training
instances