Topic 17 Flashcards

(11 cards)

1
Q

Bayes Theorem

A

P(Y|X) = P(X|Y)P(Y) /P(X) Where P(Y|X) is the posterior probability (description given all the available data), P(X|Y) is the likelihood (predicts the data), P(Y) is the prior probability (based on prior information) and P(X) is the normalising constant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Product Rule for Independent Events

A

P(X|Y) = P(x1, x2, … , xn|Y) = P(x1|Y)P(x2|Y)…P(xn|Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Conditional Probability

A

The probability that A and B are true is the same as the probability that A is true and that B is true given that A is true: P(A and B) = P(A) x P(B|A)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

Naive Bayes General Formula

A

P(Y|x1, … , xn) ∝ 𝑃(𝑌)ෑ
Σ 𝑖=1,𝑛 𝑃(𝑥𝑖 |𝑌)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

Y With Maximum Probability

A

𝑌𝑝𝑟𝑒𝑑 = 𝑎𝑟𝑔𝑚𝑎𝑥𝑦𝑃(𝑌) Σ n, i=1 𝑃(𝑥𝑖 |𝑌)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

Laplace/ Plus One Smoothing

A

Adding 1 to all occurences / data points to avoid having a probability of 0 that will ruin Naive Bayes calculations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Multinomial Naive Bayes

A

Document classification (eg. assigning an article to sports, politics etc or spam vs not spam)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Bernoulli Naive Bayes

A

Similar to Multinomial. but the features / predictors are boolean variables (eg. tennis weather example)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

Gaussian Naive Bayes

A

When features are not discrete and take up a continuous value (assumed that the values are sampled from a Gaussian distribution), estimate a normal distribution with some mean and SD for each feature

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

Naive Bayes Advantages

A

*
It is easy and fast to predict a class of test dataset
*
Naïve Bayes classifier performs better compare to other models assuming independence
*
It performs well in case of categorical input variables compared to numerical variables

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

Naive Bayes Disadvantage

A

Independent Predictor Assumption (variables might not be independent)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly