Class Four Flashcards

1
Q

What is Principal Component Analysis (PCA)?

A

Principal Component Analysis (PCA) is a dimensionality reduction technique used to transform a high-dimensional dataset into a lower-dimensional representation by finding the principal components that capture the most significant variation in the data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the advantages of Principal Component Analysis (PCA)?

A

Advantages of PCA include reducing the dimensionality of the data, removing correlated features, and identifying the most informative features or patterns.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the limitations of Principal Component Analysis (PCA)?

A

Limitations of PCA include difficulty in interpreting the transformed components, sensitivity to outliers, and assumptions of linearity and normality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the different types of PCA?

A
  1. Randomized PCA quickly finds an approximation of the first principal components.
  2. Incremental PCA (IPCA) splits the training set into mini-batches and feed an IPCA algorithm one mini-batch at a time.
  3. Kernel PCA helps to perform complex nonlinear projections for
    dimensionality reduction.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Singular Value Decomposition (SVD)?

A

Singular Value Decomposition (SVD) is a matrix factorization technique used in linear algebra to decompose a matrix into three separate matrices to extract the underlying structure and reduce dimensionality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the advantages of Singular Value Decomposition (SVD)?

A

Advantages of SVD include its ability to handle missing values in data, extract latent features, and provide a low-rank approximation of the original matrix.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the limitations of Singular Value Decomposition (SVD)?

A

Limitations of SVD include its computational complexity for large matrices, difficulty in interpreting the singular values and vectors directly, and sensitivity to noise.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is dimensionality reduction?

A

Dimensionality reduction is the process of reducing the number of features or variables in a dataset while preserving important information, aiming to eliminate irrelevant or redundant features.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What are the advantages of dimensionality reduction?

A

Advantages of dimensionality reduction include simplifying the analysis and visualization of data, reducing computational complexity, and mitigating the curse of dimensionality.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What are the limitations of dimensionality reduction?

A

Limitations of dimensionality reduction techniques include potential loss of information, difficulty in selecting the appropriate number of dimensions, and potential distortion of the data’s structure.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are decision trees?

A

Decision trees are supervised machine learning models that recursively split the data based on features to create a tree-like structure for making decisions or predictions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are the advantages of decision trees?

A

Advantages of decision trees include interpretability, handling both numerical and categorical features, and automatic feature selection.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What are the limitations of decision trees?

A

Limitations of decision trees include overfitting, sensitivity to small changes in the data, and difficulty in capturing complex relationships or interactions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

When should we use Decision Trees?

A

We should consider decision trees when:

1.When our data is described using attributes (characteristics) and values. (attribute-value pairs)
2.When we want to predict outcomes that have specific categories or answers (like “yes” or “no”).
3.When we need to express choices that can be made by combining different conditions (like “if this or that”).
4.When our training data might have mistakes or some missing information, and we want a method that can handle those issues.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the Gini impurity measure?

A

The Gini impurity is a measure of impurity or homogeneity used in decision tree algorithms. It quantifies the probability of misclassifying a randomly chosen element if it were randomly labeled based on the distribution of classes in a given node. A lower Gini impurity indicates a more homogeneous node with a higher degree of purity.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What is the entropy impurity measure?

A

Entropy is a measure of impurity or disorder used in decision tree algorithms. It calculates the uncertainty or randomness of the distribution of classes in a given node. A higher entropy indicates a more heterogeneous node with a lower degree of purity. The goal in decision trees is to minimize entropy by splitting nodes in a way that increases the purity and separation of classes.

17
Q

What is reduction of entropy called?

A

It is called information gains, and it derives from Shannon information theory.

18
Q

What are ensembles, specifically Random Forests?

A

Ensembles combine multiple individual models to make predictions or decisions. Random Forests are ensembles of decision trees, where each tree is trained on a random subset of the data and features.

19
Q

What are the advantages of Random Forests?

A

Advantages of Random Forests include improved prediction accuracy, handling of high-dimensional data, resistance to overfitting, and feature importance estimation.

20
Q

What are the limitations of Random Forests?

A

Limitations of Random Forests include potential overfitting with noisy or irrelevant features, increased complexity and computation time compared to single decision trees, and difficulties in interpreting the final model due to the ensemble nature of Random Forests.

Additionally, Random Forests may struggle to handle imbalanced datasets, and their performance can be affected if there is high cardinality or multicollinearity among the features. Moreover, Random Forests are not suitable for tasks that require real-time predictions due to their inherent computational requirements.

21
Q

What is bagging in machine learning?

A

Bagging, short for bootstrap aggregating, is an ensemble learning technique where multiple models are trained on different bootstrap samples of the training data. These models are trained independently, and their predictions are combined through voting or averaging to make the final prediction. Bagging helps reduce variance and improve the overall stability and accuracy of the model.

Class:
Bagging builds many independent predictors and combine them using some
model averaging techniques. (e.g. weighted average, majority vote)
* Random Forest
* Handle overfitting

22
Q

What is boosting in machine learning?

A

Boosting is an ensemble learning technique that iteratively trains weak models to create a strong model. Each weak model is trained on a modified version of the training data, where the emphasis is placed on the instances that were misclassified by previous models. By combining the predictions of these weak models, boosting creates a powerful model that improves over time. Boosting focuses on reducing bias and improving the model’s predictive performance.

Class:
In Boosting, predictors are not made independently, but sequentially.
* Gradient Boosting
* Can overfit