Decision Trees Flashcards

a non-parametric supervised learning algorithm for classification and regression tasks

1
Q

Equation for Entropy

A

SUM(
-P(outcomeN)*log2(outcomeN)
)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Equation for Information Gain

A

Entropy(S) -
Sum
(
P(Rows with Feature=Value in Set)
*
Entropy(Feature=Value)
)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the two type of Decision Trees?

A

Classification(Target is discrete)
Regression(Target is continuous)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the stopping criteria for decision trees?

A

1) All or nearly all data has same class label
2) When there are no features to further split upon
3) When tree reaches a predefined depth

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What are feature selection measures?

A

deciding which features are to be considered as the root node and at each level.
Eg: Gini Index, Information Gain

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q
A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly