Chapter 2- Probabilistic Models Flashcards

1
Q

what is a random variable?

A

a numeric quantity whose values map to the possible outcomes of an experiment

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

what is a sample space, or alphabet?

A

consists of all possible events

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

what is an event?

A

any subset of values that X can take

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

what is the first rule of probability theory?

A

probabilities add up to 1

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

what is the difference between estimated and true probabilities?

A

an estimate comes from a sample- it is a sample estimate

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

what is another name for the true probability?

A

a population parameters

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

what is joint probability?

A

AND

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

from conditional probabilities, p(x,y) = ?

A

p(x|y)p(y) and vice versa, p(y|x)p(x)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

what is the rule for independent events?

A

p(x,y) = p(x)p(y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

how is bayes theorem derived?

A

p(x,y) = p(x|y)p(y)
and vice versa p(x,y) = p(y|x)p(x)
equate these

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

what is another word for joint probability?

A

marginalisation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

what is the formula for joint probability, p(X=x)?

A

sum for each y: p(x|y)p(y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

what is the conditional independence assumption? (in words)

A

the features are conditionally independent of each other, given the class value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

what is the conditional independence assumption p(x1,x2,x3|y) = ?

A

p(x1|y)p(x2|y)p(x3|y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

what is the conditional independence assumption p(X|y) = ?

A

multiply for each x: p(x|y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

what is the naive bayes model?

A

we solve problems by making the conditional independence assumption

17
Q

a bayesian network is an example of?

A

a directed acyclic graph

18
Q

when might over/underfitting occur in a bayesian network?

A

if we have more links, the more complicated the probability distribution and hence more data is needed

19
Q

what is one of the great advantages of bayesian networks?

A

you can very naturally encode human knowledge about a given problem

20
Q

what is the chain rule for probability: P(A n B) = ?, and P(A1 n A2 n A3 n A4) = ?

A

P(A n B) = p(B|A)P(A)
or for more
P(A1 n A2 n A3 n A4) = p(A4 | A1 n A2 n A3)
= P(A4 | A3 n A2 n A1) p(A3 | A2 n A1)p(A2 | A1) p(A1)

21
Q

what does the chain rule tell us how to compute? Events A and B

A

The chain rules tells us how to compute the probability of event A happening AND then event B occurring afterwards.

22
Q

what is a random experiment?

A

random experiment is any experiment in which the outcome is uncertain (not known or determined in advance).

23
Q

what is P(A or B) if

a) they are disjoint
b) they are joint

A

a) P(A) + P(B)

b) P(A) + P(B) - P(A and B)

24
Q

what is a probability mass function?

A

discrete random variables take on a finite number of values.
Each value is associated with a probability of it occurring. t
The collection of these probabilities is the probability mass function

25
give the bernoulli distribution
``` P(X = 0) = 1 - p p(X = 1) = p ```
26
give the binomial distribution, P(X=k) =
P(X = k) = (nCk)(p^k)(1-p)^(n-k)
27
when would we use the binominal distribution
if experiment is repeated n times, the probability that we will see k successes is given by this probability mass function
28
give the geometric distribution
P(X=x) = (1-p)^x-1 (p)
29
when would we use the geometric distribution
it is useful for modelling the first occurrence of an outcome after repeated identical trials
30
give the poisson distribution
P(X=x) = { lambda^x e(-lambda) } / x!
31
when would we use the poisson distribution
if we had a rate e.g. 5 times a year
32
if a discrete r.v. X has a pmf f(X) what is the expected value E[g(x)]
sum i: g(Xi)f(Xi)
33
if a discrete r.v. X has a pmf f(X) what is the variance V[g(x)]
E[(g(X) - E(g(X)))^2] = E[g(X)^2] - E[g(X)]^2
34
properties of Expectations | E[aX + b] =
aE[X] + b
35
properties of variance: | V[aX+b] =
a^2V[X]
36
what are the two schools of probability?
frequency based belief based
37
Describe the gamblers fallacy and how it demonstrates iid
A sequence of outcomes of spins of a fair or unfair roulette wheel is i.i.d The outcome of the previous turn doesnt impact the next turn The distribution of probabilities is the same each time i.e. 50/50 if not biased. So, even if we've had 20 reds, there is still the same chance of having a black vs red next go
38
Let the two events be the probabilities of persons A and B getting home in time for dinner, and the third event is the fact that a snow storm hit the city. If they live in different areas, are A and B conditionally independent given C?
Yes. That is, the knowledge that A is late does not tell you whether B will be late. If they lived in the same neighbourhood they would not be.
39
if two events are independent, are they also conditionally dependent?
not necessarily, it depends on the third event. rolling a dice twice are two independent events. if the third event was the sum of them being 7, they would then be independent but not conditionally independent given c