CHAPTER THREE Bayesian Decision Theory Flashcards

1
Q

What is deductive reasoning?

A

A logical process where if the premises are accepted, the conclusion must also be accepted.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Provide an example of a valid but factually incorrect syllogism.

A

Plants are good for you; tobacco is a plant; ergo tobacco is good for you.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

Who is credited with the idea of deductive reasoning?

A

Aristotle.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

What are the two strong syllogisms described by E. T. Jaynes?

A
  • If A is true, then B is true.
  • A is true.
  • Therefore B is true.
  • If A is true, then B is true.
  • B is false.
  • Therefore A is false.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

What does A ∨ B represent in Boolean algebra?

A

Both A and B are true (conjunction).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

What does A ∧ B mean in Boolean algebra?

A

At least one of A and B is true (disjunction).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

What is the implication represented by A → B?

A

If A is true, then B is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

What is the output of an AND gate?

A

The output is true if both inputs are true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What does an OR gate do?

A

It outputs true if at least one of its inputs is true.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What is a NOT gate’s function?

A

It outputs true if it is not receiving some input.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are NAND gates capable of?

A

They can be used to create all other types of gates.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

What is the limitation of classical logic?

A

It deals only with absolute truths (ones and zeros).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What does Bayesian reasoning allow us to do?

A

It helps us deal with probabilities and uncertainties.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

What is Bayes’ theorem used for?

A

It provides a mathematical framework for updating beliefs based on new evidence.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What is the difference between frequentist and Bayesian approaches in decision theory?

A

Decision theorists cannot use frequentist math.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

In the lottery example, what is the prior probability of winning with one ticket?

A

1 in 131,115,985.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

What happens when the box beeps while testing a lottery number?

A

You must consider the likelihood ratio and the prior probability to update your belief.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

What is the likelihood ratio in the beeping box example?

A

4:1.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

What is the new posterior probability after the box beeps?

A

4:131,115,985.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

What does Bayesian reasoning rely on?

A

It combines prior information with new data to form a revised understanding.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

Fill in the blank: Bayesian logic allows us to deal with ____ in reasoning.

A

shades of gray.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

True or False: Bayesian reasoning is only applicable in scientific contexts.

A

False.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

What is the probability mass of now?

A

1/32,778,996

This represents a very small chance in a probabilistic context.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

How many wrong combinations does the box beep for on average?

A

8,194,749 wrong combinations

This indicates the challenge of identifying the correct combination amidst many false positives.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
How many times must you run a ticket through the box for it to be likely the right one?
Fourteen times ## Footnote This emphasizes the improbability of identifying the correct ticket easily.
26
What concept in thermodynamics is compared to Bayes' theorem?
Carnot engine ## Footnote The Carnot engine represents the most efficient theoretical model for heat engines.
27
What does Bayes' theorem help approximate in decision-making?
It helps approximate Bayesian decision theory ## Footnote Decisions under uncertainty are better when they approximate Bayes' theorem.
28
What does E. T. Jaynes argue about Bayes' theorem?
It allows for plausible reasoning beyond Aristotelian logic ## Footnote Jaynes emphasizes that using probabilities can extend logical reasoning.
29
What is an example of a logical conjunction represented in Bayesian terms?
p(A ∧ B) = p(C) ## Footnote This shows how the probability of two events occurring together can be expressed in Bayesian logic.
30
What does the likelihood ratio indicate in the context of wet pavements?
It tells how much to update beliefs about rain based on evidence ## Footnote The likelihood ratio quantifies how much more plausible a hypothesis becomes given new evidence.
31
What is the prior probability of rain at a certain time of year in the example?
33 percent ## Footnote This serves as the baseline probability before considering additional evidence.
32
What is the posterior probability of it having rained given wet pavements?
66 percent ## Footnote This result derives from applying Bayes' theorem with prior probabilities and likelihoods.
33
What is Cromwell's rule in Bayesian decision theory?
Never assign probabilities of one or zero to anything except logical truths ## Footnote This rule encourages keeping an open mind to possibilities.
34
What happens to posterior probability if the prior is set to zero?
The posterior probability remains zero ## Footnote This illustrates the issue of being too certain about initial beliefs.
35
How are odds calculated from probabilities?
By dividing the probability by 1 minus the probability ## Footnote This transformation highlights the relationship between probabilities and odds.
36
What is the odds representation of a probability of 0.9?
9:1 ## Footnote This indicates a strong likelihood in favor of the event occurring.
37
What is a critical distinction between probabilities of 1 and 0.999999?
1 equals infinity to 1 ## Footnote This distinction shows the mathematical implications of assigning absolute certainty.
38
Why should you avoid assigning a probability of one?
Because it implies absolute certainty, which is unrealistic ## Footnote Real-world events often have uncertainties that should be accounted for.
39
What is the implication of assigning a probability of zero or one?
You should never assign anything a probability of zero or one. ## Footnote This means acknowledging that while some events are extremely unlikely, they are not impossible.
40
What does a very small probability indicate?
Very, very small probabilities are very, very small and should not be confused with impossibility. ## Footnote For example, a one-in-a-quadrillion chance exists but is extremely unlikely.
41
What is the conservation of expected evidence?
You can't go looking for new evidence to support your theory; the absence of expected evidence counts as evidence against your hypothesis. ## Footnote This is a principle in Bayesian decision theory.
42
How does the absence of expected evidence affect belief?
If you expect to see evidence and do not find it, your belief should shift significantly in the opposite direction. ## Footnote For example, if you expect to see evidence of wrongdoing and do not, your belief in the wrongdoing decreases.
43
What is the relationship between expected evidence and posterior probability?
The more strongly you expect something, the less your posterior probability changes when you find it. ## Footnote Conversely, unexpected evidence causes a more significant shift in belief.
44
Fill in the blank: The absence of evidence is, in fact, _______.
evidence of absence. ## Footnote This principle suggests that not finding evidence for a belief can strengthen the belief that it does not exist.
45
What is utility in decision theory?
Utility describes how much you care about something in decision-making under uncertainty. ## Footnote It is often treated as equivalent to money for simplicity in calculations.
46
What does expected value combine?
Expected value combines probability and utility. ## Footnote This helps in making decisions based on the anticipated outcomes of those decisions.
47
How is expected value calculated using a lottery example?
Expected value is calculated by dividing the value of the jackpot by the chance of winning it. ## Footnote For example, if a lottery ticket costs £1 and has a jackpot of £150 million with a 1 in 131,115,985 chance, the expected value is positive.
48
What is a Dutch book in betting theory?
A Dutch book occurs when a person's beliefs about probabilities lead to guaranteed losses regardless of the outcome. ## Footnote It demonstrates irrationality in betting based on inconsistent probability assessments.
49
What did John von Neumann contribute to decision theory?
John von Neumann developed game theory and sought a normative way to make decisions under uncertainty to maximize expected well-being. ## Footnote His work laid the foundation for understanding complex decision-making in economics.
50
What challenge arises when trying to maximize group utility?
Conflicts of interest between individuals complicate the process of maximizing group utility. ## Footnote Different individuals may prioritize different desires, making it difficult to achieve a consensus on utility maximization.
51
Fill in the blank: Classical economics assumes that while you can rank people's preferences, you cannot _______.
compare them. ## Footnote This is particularly true when preferences conflict between individuals.
52
What is a key axiom proposed by von Neumann regarding people's desires?
People's desires need to be transitive.
53
Define transitive preferences in the context of decision-making.
If a person prefers A to B and B to C, then they must prefer A to C.
54
What happens if preferences are intransitive?
The individual can become a money pump.
55
What are the three necessary conditions for preferences according to von Neumann?
* Transitive * Continuous * Monotonic
56
What does it mean for preferences to be continuous?
There are no sudden jumps in people's preferences as outcomes change.
57
What does monotonicity imply about decision-making?
A decision with a 50% chance of £10 should be indifferent to a decision with a 100% chance of £5.
58
What is meant by substitutability in preferences?
If indifferent between cake and jelly, one shouldn't care about the probabilities of receiving each.
59
What is the utility theorem proposed by von Neumann?
People have preferences that can be assigned a numerical value, called 'utils'.
60
What is the concept of 'utils'?
A unit of measure for preferences in von Neumann's utility theorem.
61
How did von Neumann apply his theories to the scenario of Holmes and Moriarty?
He analyzed their decision-making strategies using expected utilities.
62
What is expected utility?
The average utility of possible outcomes weighted by their probabilities.
63
What should Moriarty do to maximize his expected utility?
Be unpredictable in his actions.
64
What is a key challenge in decision-making under uncertainty?
Knowing the expected utility of any decision due to lack of information.
65
What is Occam's razor?
A principle stating that simpler explanations are preferred over complex ones.
66
Who is Occam's razor named after?
William of Ockham.
67
What is minimum message length in decision theory?
The shortest computer program that describes a given output.
68
What does Kolmogorov complexity refer to?
The complexity of an object in terms of the length of the shortest description.
69
What example illustrates the concept of minimum message length?
Three eleven-digit strings of numbers, where one is predictable and two are not.
70
What is the significance of randomness in the context of algorithm complexity?
Truly random sequences require longer descriptions than predictable ones.
71
What trade-off is discussed regarding hypothesis selection?
Between the complexity of the algorithm and the confidence in predicting the output.
72
How is information defined in the context of decision-making?
A single 'bit' of information can halve the probability space.
73
What is the role of priors in Bayesianism?
They are initial probabilities that influence the outcome of Bayesian inference.
74
True or False: Complexity in hypotheses should always be minimized regardless of fit.
False.
75
Fill in the blank: The __________ is the simplest explanation in a set of hypotheses.
[Occam's razor]
76
What does assigning p = 0.02 to the remaining doors signify?
It indicates that the search space has been halved, increasing the probability mass on each remaining option.
77
What is the trade-off mentioned between complexity and good fit?
If an extra bit of information doesn’t allow you to halve the search space, it’s not worth it.
78
How should one choose between two hypotheses according to the text?
Assign higher prior probability to the simpler hypothesis to write as a computer program.
79
What is the relationship between modern AI systems and Bayesian principles?
Modern AI systems operate in a manner consistent with Bayesian decision-making under uncertainty.
80
What is a hyperprior?
It's uncertainty about what parameter to use, reflecting a higher-level prediction about the shape of the world.
81
How does the example of a Bayesian AI playing hide-and-seek illustrate hyperpriors?
It demonstrates adapting prior probabilities based on changing evidence, like the opponent's hiding behavior.
82
What is the significance of having multiple hypotheses in evaluating claims?
It complicates the evaluation because different priors can lead to different conclusions despite the same evidence.
83
What example is given to illustrate the skepticism of psychic claims?
The Mysterious Barry guessing numbers correctly, where even many correct guesses may not convince skeptics.
84
What were the results of Samuel Soal's card-guessing experiment?
Two subjects scored significantly better than chance, raising questions about psychic powers.
85
What is the effect of alternative hypotheses on belief in psychic powers?
Alternative explanations, like fraud or design flaws, can overshadow belief in psychic phenomena.
86
How does prior probability influence the interpretation of new evidence?
Individuals with different priors may interpret the same evidence in opposing ways, reinforcing their beliefs.
87
What is the core function of artificial intelligence as described?
To predict uncertain outcomes, fundamentally based on Bayesian reasoning.
88
What is supervised learning in AI?
It's a model that predicts labels for new data based on labeled training data.
89
How does an AI update its probability after seeing a picture?
It shifts from a prior probability to a posterior probability based on new information.
90
What statistical concept is used to find the line of best fit in AI?
Linear regression.
91
Fill in the blank: A simple AI might predict the likelihood of an image being a lion with a prior probability of _______.
p ≈ 0.33
92
True or False: Bayesian machine learning explicitly mimics Bayes' rule.
True
93
What happens when evidence suggests a hypothesis is false?
Individuals might conclude the source of the evidence is untrustworthy rather than changing their belief.
94
What is the relationship between shoe size and height as observed in a random sample?
On average, taller people have larger feet.
95
What is the line of least squares used for?
To draw a line through data points that minimizes the sum of squared errors.
96
How is the error calculated in the line of least squares?
By measuring the vertical distance from the line to each dot and squaring that distance.
97
What does the sum of squared error represent?
The total error across all data points.
98
What is the purpose of the AI in the context of predicting height from shoe size?
To estimate a person's height based on their shoe size using the line of least squares.
99
What factors influence the AI's confidence in its predictions?
The amount of training data and the variance in that data.
100
What is the Bayesian process mentioned in the context of AI?
It involves updating prior beliefs with new data to form a posterior distribution.
101
What type of curves might be used instead of a straight line for data fitting?
Curved lines, such as exponential curves, S-shaped curves, or sine waves.
102
What does it mean for a model to be 'underfitting'?
When the model is too simple to accurately capture the underlying data.
103
What is 'overfitting' in the context of AI models?
When a model fits the training data too closely and fails to generalize to new data.
104
What are hyperparameters in AI modeling?
Parameters that control the model's capacity to fit the data, such as the degree of freedom in curve fitting.
105
What is meant by hyperpriors?
The AI's prior beliefs about hyperparameters.
106
What is the primary function of AIs like ChatGPT?
To predict what a human would say or draw in response to a prompt.
107
How does ChatGPT generate responses?
By predicting the next word in a sequence based on prior text.
108
What does the term 'stochastic parrots' refer to?
A term used to describe language models that predict sequences without genuine understanding.
109
What is the significance of building a model of the world in AI?
It helps the AI make better predictions about future data.
110
What was the focus of the study involving Othello-GPT?
To determine whether the AI built an internal representation of the game rather than just memorizing statistics.
111
What does 'probing' refer to in AI research?
A technique to examine the internal states of the AI and its decision-making process.
112
What conclusion was drawn about LLMs and their ability to predict?
They likely build internal representations that assist in making predictions.
113
True or False: Large language models like ChatGPT understand the world in the same way humans do.
False.
114
Fill in the blank: The process of predicting the next token in a sequence is an inherently ______ process.
Bayesian.
115
What does it imply if an AI can make original, legal moves in a game it hasn't seen before?
It suggests the AI has formed a model of the game instead of merely memorizing moves.