lecture 4 Flashcards

(30 cards)

1
Q

What is the focus of probabilistic models in machine learning?

A

Using probability theory to select models for given data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

What is conditional probability?

A

p(X | Y) is the probability of X given Y.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

What is Bayes’ theorem?

A

p(X | Y) = p(Y | X) * p(X) / p(Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What is a useful probability rule?

A

p(X, Y) = p(X | Y) * p(Y)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What is (Conditional) Independence?

A

When knowing Y provides no information about X.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What is expectation in probability?

A

The weighted average of all possible values of a random variable.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What are common probability distributions in ML?

A

Bernoulli, Categorical, and Normal distributions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What does p(Data | θ) represent?

A

The likelihood of observing data given parameters θ.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What is the goal in probabilistic learning?

A

To infer the parameters θ given only the observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is the frequentist view of probability?

A

The true model exists but is unknown, and we estimate its parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What is maximum likelihood estimation (MLE)?

A

Finding parameters that maximize the likelihood of observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the likelihood function L(θ)?

A

p(X | θ), the probability of observed data given parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is the MLE (maximum likelihood estimation) formula?

A

θ̂ = argmax_θ p(X | θ)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is a simple analogy for probabilistic learning?

A

A ‘machine’ generates data, and we infer its parameters.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What does the model space represent?

A

All possible models that could explain the observed data.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

What does maximum likelihood estimation (MLE) assume?

A

That the best model is the one that makes the observed data most likely.

17
Q

What is a practical example of MLE?

A

Estimating if a coin is fair or biased based on observed flips.

18
Q

What is the difference between frequentist and Bayesian inference?

A

Frequentists estimate a fixed parameter, while Bayesians use probability distributions over parameters.

19
Q

What is the Bayesian approach to learning?

A

Updating beliefs about parameters using observed data and prior knowledge.

20
Q

What is a prior in Bayesian inference?

A

The initial belief about a parameter before observing data.

21
Q

What is a posterior distribution?

A

The updated belief about a parameter after observing data.

22
Q

What is the likelihood in Bayesian inference?

A

The probability of observed data given a parameter.

23
Q

What is the Bayes rule formula for updating beliefs?

A

Posterior ∝ Likelihood × Prior

24
Q

What is a conjugate prior?

A

A prior that, when updated, results in a posterior with the same functional form.

25
What is the benefit of using conjugate priors?
They simplify Bayesian inference calculations.
26
What is MAP estimation?
Maximum a posteriori estimation, which finds the most probable parameter given both data and prior.
27
What is the difference between MAP and MLE?
MAP considers prior knowledge, while MLE relies only on data.
28
What is the purpose of probabilistic modeling in ML?
To incorporate uncertainty and make robust predictions.
29
What is the relationship between likelihood and probability?
Likelihood measures how well parameters explain data, while probability represents chance.
30
What is the key takeaway from probabilistic models?
They provide a framework for handling uncertainty in machine learning.