Class Seven Flashcards

1
Q

What is Bayesian statistics?

A

Bayesian statistics is an approach to statistical inference that uses Bayes’ theorem to update prior beliefs or knowledge based on observed data, resulting in posterior probability distributions.

P(A|B)= (P(A)*P(B|A))/P(B)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What are the advantages of Bayesian statistics?

A

Advantages of Bayesian statistics include the ability to incorporate prior knowledge, flexibility in handling complex models, and the interpretation of results as probabilities.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What are the limitations of Bayesian statistics?

A

Limitations of Bayesian statistics include the need for informative prior distributions, computational complexity for complex models, and potential subjectivity in choosing priors.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are probabilistic classifiers?

A

Probabilistic classifiers are machine learning models that assign class labels to instances and provide a probability or likelihood estimate of the assigned label based on observed features or attributes.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is Naive Bayes classifier?

A

Naive Bayes classifier is a probabilistic classifier based on Bayes’ theorem and the assumption of feature independence given the class. It calculates the posterior probability of each class and assigns the instance to the class with the highest probability.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What are the advantages of Naive Bayes classifier?

A

Advantages of Naive Bayes classifier include simplicity, fast training and prediction times, and the ability to handle high-dimensional data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are the limitations of Naive Bayes classifier?

A

Limitations of Naive Bayes classifier include the assumption of feature independence (which may not hold in all cases), the sensitivity to irrelevant features, and the potential for poor performance with imbalanced datasets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the difference between Naïve Bayes and Decision Tree?

A

Naïve Bayes:
* Simultaneously combine all features.
* Training: 1 pass over data to count.
* Conditional independence assumption.
* Testing: look at all features.
* New data: just update counts.
* Accuracy: good if features almost
independent given label (text).

Decision trees:
* Sequence of rules based on 1 feature.
* Training: 1 pass over data per depth.
* Greedy splitting as approximation.
* Testing: just look at features in rules.
* New data: might need to change tree.
* Accuracy: good if simple rules based on
individual features work (“symptoms”).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is class imbalance in machine learning?

A

Class imbalance refers to a situation where the distribution of class labels in a dataset is uneven, with one class having significantly fewer instances than the others.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is SMOTE (Synthetic Minority Over-sampling Technique)?

A

SMOTE is a technique used to address class imbalance by generating synthetic minority class instances by interpolating between existing minority class instances.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are the advantages of SMOTE?

A

Advantages of SMOTE include the ability to balance class distribution, the generation of diverse synthetic samples, and the potential improvement of minority class prediction performance.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What are the limitations of SMOTE?

A

Limitations of SMOTE include the potential creation of noisy or unrealistic synthetic instances, sensitivity to the choice of neighbors, and difficulties in handling overlapping or borderline cases.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is ADASYN (Adaptive Synthetic Sampling)?

A

ADASYN is an extension of SMOTE that adaptively adjusts the generation of synthetic samples based on the difficulty of learning from the minority class instances.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What are the advantages of ADASYN?

A

Advantages of ADASYN include its ability to focus on more challenging minority class instances, better handling of overlapping classes, and potential performance improvement over SMOTE in imbalanced datasets.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are the limitations of ADASYN?

A

Limitations of ADASYN include the potential generation of noisy synthetic instances, sensitivity to the choice of neighbors, and the need for careful parameter tuning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

How can class imbalance be addressed in machine learning?

A

Class imbalance can be addressed using techniques such as oversampling the minority class (e.g., SMOTE, ADASYN), undersampling the majority class, using ensemble methods, or applying cost-sensitive learning techniques.