12.4 Decision Making Flashcards
Normative decision theories
define how people ought to make decisions that yield the optimal choice. Very often, such theories fail to predict what people actually choose.
Descriptive decision theories
attempt to describe what people actually do, not what they should do.
action-outcome decisions
where the decisions involves some from of evaluation of the expected outcomes
stimulus-response decisions
if the outcome is consistent, it becomes a stimulus-response decision.
Model-based
means that the agent has an internal representation of some
aspect of the world and uses this model to evaluate different actions.
Model-free
means that you have only an input–output mapping, similar to
stimulus–response decisions.
primary reinforcers
they have a direct benefit for survival fitness
secondary reinforcers
sch as money and status, are rewards that have no intrinsic value themselves, but become rewarding through their association with other forms of reinforcement.
Value represented in the brain
in the ACC, anterior cingulate cortex, the LPFC and the OFC. Overall, the neurophysiological and neuroimaging studies indicate that the OFC plays a key role in the representation of value.
temporal discounting
is the observation that the value of a reward is reduced when we have to wait to receive that reward.
dopamine (DA).
Dopaminergic cells are scattered throughout the midbrain, sending axonal projections to many cortical and subcortical areas. Two of the primary loci of dopaminergic neurons are two brainstem nuclei, the substantia nigra pars compacta (SNc) and the ventral tegmental area (VTA).
Dopaminergic neurons that originate in the VTA project through two pathways:
The mesolimbic pathway travels to structures important to emotional processing, and the mesocortical pathway travels to the neocortex, particularly to the medial portions of the frontal lobe.
reward prediction error (RPE)
Rather than thinking of the spike in DA neuron activity as representing the reward, he suggested that it should be viewed as a reward prediction error (RPE), s the difference between the expected reward and what is actually obtained. The RPE is used as a learning signal to update value information as expectancies and the valence of rewards change. The activity of some DA neurons provides a neuronal
code of prediction errors.