Principal Components Analysis Flashcards

1
Q

How do we deal with high dimensionality (3)?

A
  • Use domain knowledge
    • Feature engineering (e.g. color historgrams for object detection)
  • Make assumptions
    • Independence
    • Smoothness
    • Symmetry
  • Reduce dimensionality
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the two methods for reducing dimensionality?

A
  • Feature selection
  • Feature extraction
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is feature selection?

A

Choosing a subset of the original features (e.g. highest infomation gain)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is feature extraction?

A

Contruct a new set of dimensions from a linear combination of the original

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does PCA try to preserve?

A

The structure (variance) in the data

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are principal components?

A

Eigen vectors with the largest eigen values

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What happens when you multiply a random vector with the covariance matrix?

A

It moves in the direction of greatest variance

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is an eigen vector?

A

A vector when multiplied by a matrix does not change direction, only magnitude

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is an eigen value?

A

The scaler for which an eigen vector grows

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

How do you find eigenvalues?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is the determinant of a 2x2 matrix?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

How do you find eigenvectors (given the eigen values)?

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

Which eigenvectors do we pick for principle components?

A

Unit length eigen vectors

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How do you project a coordinate x’ given ei, …, em eigen vectors?

A

(x’ - mu)Tej for j = 1…m

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What property does the eigen vector for a principle component have?

A

Its where the data is spread out the most

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How do we pick the amount of components to use (PCA)?

A
  • Pick the first m which explain some threshold of the total variance
  • Use a scree plot
17
Q

What are typical variance threshold values (PCA)?

A

0.9/0.95

18
Q

How do you compute what porportion of the variance m principle components explain? (given their are d dimensions)

A

lambda is the eigenvalues for each principle component

19
Q

What do we do before finding principle components (PCA)?

A

Center points (subtract mean)

20
Q

What is the advantages using eigen faces for simularity?

A

Insensitive to lighting, expression, orientation

21
Q

What are the pratical issues with PCA?

A
  • Sensitive to large values (large attribute -> large variance -> always picked as 1st component)
  • Always linear projection (line/hyperplane