Chapter 15 Linear Discriminant Analysis Flashcards

1
Q

WHEN DO WE USE LDA? P71

A

Logistic regression is a classification algorithm traditionally limited to only two-class classification problems. If you have more than two classes then the Linear Discriminant Analysis is the preferred linear classification technique. But even with binary classification problems, it is a good idea to try both logistic regression and linear discriminant analysis.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

WHAT ARE THE LIMITATIONS OF LOGISTIC REGRESSION? P71

A

ˆ Two-Class Problems. Logistic regression is intended for two-class or binary classification problems. It can be extended for multi-class classification, but is rarely used for this purpose.
ˆ Unstable With Well Separated Classes. Logistic regression can become unstable when the classes are well separated.
ˆ Unstable With Few Examples. Logistic regression can become unstable when there are few examples from which to estimate the parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

WHAT ARE THE ASSUMPTIONS OF LDA? P72

A

That your data is Gaussian, that each variable is shaped like a bell curve when plotted.
That each attribute has the same variance; that values of each variable vary around the mean by the same amount on average.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

HOW CAN WE PREPARE DATA FOR LDA? P73

A

ˆ Classification Problems. This might go without saying, but LDA is intended for classification problems where the output variable is categorical. LDA supports both binary and multiclass classification.
ˆ Gaussian Distribution. The standard implementation of the model assumes a Gaussian distribution of the input variables. Consider reviewing the univariate distributions of each attribute and using transforms to make them more Gaussian-looking (e.g. log and root for exponential distributions and Box-Cox for skewed distributions).
ˆ Remove Outliers. Consider removing outliers from your data. These can skew the basic statistics used to separate classes in LDA such as the mean and the standard deviation.
ˆ Same Variance. LDA assumes that each input variable has the same variance. It’s almost always a good idea to standardize your data before using LDA so that it has a mean of 0 and a standard deviation of 1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

WHAT ARE THE NAMES OF THE EXTENSIONS TO LDA MODEL?
P 73

A

ˆ Quadratic Discriminant Analysis: Each class uses its own estimate of variance (or covariance when there are multiple input variables).
ˆ Flexible Discriminant Analysis: Where nonlinear combination of inputs is used such as splines.
ˆ Regularized Discriminant Analysis: Introduces regularization into the estimate of the variance (or covariance), moderating the influence of different variables on LDA.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What’s the similarity and difference between LDA and PCA?

External

A
  • Both LDA and PCA rely on linear transformations and aim to maximize the variance in a lower dimension. (reduce dimension)
  • Both method’s rank the new axis in order of importance (PC1 accounts for the most variation in the data, LD1 accounts for the most variation between the categories)
  • Unlike PCA, LDA finds the linear discriminants in order to maximize the variance between the different categories while minimizing the variance within the class. (LDA focuses on maximizing the separabality among known categories) but PCA reduces dimensionality by linearly transforming the data into a new coordinate system where (most of) the variation in the data can be described with fewer dimensions than the initial data.

Ref

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

When should we use LDA for dimensionality reduction?

External

A

we should be using LDA for dimensionality reduction if the objective is to maintain class separability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

If we have 3 classes and 18 features, LDA will reduce from 18 features to only ____ features

External

A

2

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Can we use LDA for a dataset with categorical features? Explain

External

A

LDA works on continuous variables. If the classification task includes categorical variables, the equivalent technique is called the discriminant correspondence analysis.

Ref

How well did you know this?
1
Not at all
2
3
4
5
Perfectly