Topic 18 Flashcards

(7 cards)

1
Q

How to best choose which features to have at the root of the decision tree?

A

Examine the probability distribution of the different outcomes and choose the features that split the probabilities most differently

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Entropy

A

A measure of the uncertainty associated with a random variable X, H(X) = -Σ n, i=1 P(xi)log2(Pxi) (Base 2 log)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Expected Entropy (over a child node)

A

EH = Σ k, i=1 Hi *|ni| / |n| where |ni| / |n| is the proportion of parent training instances that went to the child note. Allows us to compare the entropy before and after splitting a parent node into child nodes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Information Gain / Reduction in Entropy (from a split using some feature F)

A

Gain(Feat) = Hn - EH(Feat) or Entropy at Parent Node – (Expected) Entropy over the Child Nodes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Gini Index

A

Gini(X) = Σ n, i=1 P(xi)(1-P(xi))

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Advantages of Decision Trees

A

*
Relatively computationally inexpensive.
*
Can deal both non-numeric features and numeric features (i.e. use a numeric cut-off boundary). See next set of topic notes.
*
Can learn complex, non-linear decision boundaries.
*
Give an easy-to-understand classification process.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Disadvantages of Decision Trees

A

Disadvantages
*
Susceptible to overfitting
*
Lack robustness (small changes to training data can mean large changes to fitted model).
In particular, the order we assign feature decisions to nodes may change based on small changes in the data. See next set of topic notes.
*
Tend to be outperformed by other methods, and refinements to the basic decision tree idea. See next set of topic notes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly