Chapter 6 Flashcards

(19 cards)

1
Q

What tasks can decision trees accomplish

A

– Classification
– Regression
– Multioutput tasks

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Do decision trees require scaling or centering

A

Decision Tree training does not require feature scaling or centering

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Define Samples of a decision tree node

A

how many training instances

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Define Value of decision tree nodes

A

classification of training instances

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Define Class of Decision tree nodes

A

classification result

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What type of trees does Scitkit-Learn CART produce

A

Binary Trees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What value of a Gini Impurity makes a node “pure”

A

gini=0

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

For the Gini impurity algorithm what does the Pi,k stand for

A

𝑝𝑖,𝑘 is ratio of class 𝑘 instances among training instances of ith node

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Decision Tree can use leaf node to estimate class probabilities

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

CART

A

Classification and Regression Tree

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

How does the cart algorithm work?

A

Repeated splitting of training set
– Find split that produces the purest subsets (weighted by their size)
– Stop when maximum depth is reached or cannot find split that reduces impurity
- Greedy algorithm

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Which is faster to compute entropy or gini?

A

Gini

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the advantage of Entropy over Gini?

A

Entropy provides slightly more balanced trees

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

Decision Trees are a nonparametric model, what does that mean?

A

– No limit on number of parameters before training
– Typically leads to overfitting

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

Hyperparameters to constrain Decision Trees to avoid overfitting

A

– Maximum tree depth: max_depth
– Minimum samples in node before splitting: min_samples_split
– Minimum samples in leaf: min_samples_leaf or min_weight_fraction_leaf
– Maximum leaf nodes: max_leaf_nodes
– Maximum features evaluated for splitting: max_features

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How can we optimize Max_ and Min_ hyperparameters to regularize a model

A

Increasing min_* and decreasing max_* will regularize model

17
Q

Benefits of decision trees

A
  • Easy to use
    – Easy to understand and interpret
    – Versatile and powerful
18
Q

Short commings of decision trees

A

– Orthogonal decision boundaries
– Sensitive to small variations in training data
* Very different model if just one training value removed
* Stochastic algorithm may lead to different models on same data
(unless random_state hyperparameter is set)

19
Q

How do Random Forests address instability

A

Averaging predictions over many trees