final Flashcards

1
Q

conditioned inhibition

A

the CS announces no US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

conditioned inhibition example

A

good response > click!; wrong response > no, thank you!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

conditioned inhibition dog trainers

A

call this a “non-reward marker” — something signals nothing will happen (good or bad)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

safety signals

A

can be used to calm down — when this is present, nothing happens

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

conditioned inhibition definition

A

procedure in which a CR elicited by a CS is inhibited when another concurrently trained CS signals the absence of a US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

CS can be…

A

xcitors or inhibitors — these constructs are crucial in our discussion of motivation (can either slow things down or accelerate things)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

conditioned excitation

A

CS being associated with a US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Method 1 (intuitive): differential inhibition or discriminative inhibition procedure

A

classical conditioning, simply pairing with a US or not (same with habituation/dishabituation). Based on a pairing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

differential/discriminative inhibition pairings

A

CSa <> US
CSx <> no US
CSa = clicker, CSx = note

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

differential/discriminative inhibition animal responds to

A

Animal responds only to CSa…in other words, it can discriminate between the two

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

differential/discriminative inhibition: two sounds

A

Trying to discriminate between two sounds. Condition one with food and the other without. Eg. If the animal is salivating, C, therefore not G:
What if CSa is C (musical note “do”) and CSx is G (musical note “sol”)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Method 2 (could be competing): conditioned inhibition procedure

A

getting two reinforced stimuli to compete

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

conditioned inhibition procedure trials

A

US is presented (reinforced trials): CSa > US
US is omitted (non reinforced trials) CSa + CSx > nothing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

conditioned inhibition outcome

A

Subjects respond to CSa
Subjects do not respond to CSa + CSx
Subjects also do not respond to CSx (= conditioned inhibitor!)
CSx by itself signals no US (food)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

conditioned inhibition CSx

A

CSx on its own was never on its own predicting the food.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

conditioned inhibition designed by

A

Pavlov

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

information value in conditioning

A

CS provides information on the US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

rescorla experiments

A

contingencies between CS (tones) and US (shock)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

rescorla experiments groups

A

3 groups:
positive, zero, negative contingency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

what is key in rescorla

A

probability of CS followed by US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

rescorla positive contingency

A

if CS is followed by US in a relatively predictable way — excitation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

rescorla zero contingency

A

completely unpredictable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

rescorla negative contingency

A

inhibition, reverse of number one. If you have a US that is not immediately preceded by the CS (almost like backward conditioning) — inhibition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

appetitive conditioning

A

cues can announce something pleasurable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

appetitive conditioning example

A

Coyotes associate the human scent or voice with food (litter on trails)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

aversive conditioning

A

Cues can announce something not pleasurable

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

aversive conditioning example

A

Humans use human scent with hazing methods to scare the coyotes away

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

powerful appetitive USs

A

food, sex, drugs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

aversive control, appetitive conditioning

A

get coyotes afraid of what they are supposed to avoid. Coyotes predicted the presence of food with humans, food and humans became the same thing — this classical conditioning was the problem. Became operant when the coyotes actually acted on the human to obtain to the food (an attack)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

aversive conditioning, hazing program

A

hazing program may not be successful with coyotes that already formed that association. Could be useful in young coyotes, to begin to conditioning coyotes that humans are dangerous. Have to be consistent in these pairings.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

main method of extinction

A

you stop pairing of CS-US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

is extinction unlearning or forgetting?

A

yes, but also learning a new response — learning what happens when CS is followed by US of one nature and of another nature

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

extinction is actually

A

Responses are suppressed or not expressed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

extinction is likely just..

A

a suppression of response. It is not permanent, if the right conditions are back together the response comes back

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
35
Q

renewal effect

A

CS-US pairing in context X
Extinction in context Y — now bringing the organism to a different environment on purpose to extinguish the response
CS in context X > renewed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
36
Q

renewal effect two associations

A
  1. CS — US
  2. CS — no US
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
37
Q

extinguishing is…

A

is context specific — extinction must be done in context of acquisition

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
38
Q

extinguishing is important in

A

therapy: eg. Might be able to handle it in the therapists office but not in the real environment
If you experience of fear of explosions in war, being exposed just to the sound in a clinic may not help you

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
39
Q

extinction is NOT

A

loss of information

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
40
Q

compound conditioning/CS

A

The presentation of 2 or more CSs, ether
- simultaneous
- serial

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
41
Q

simultaenous compound CS

A

CSs presented at same time

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
42
Q

serial compound CS

A

CSs presented in a sequence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
43
Q

compound conditioning example

A

Tone + light > US (food)
This will condition normally, as if one S
But, if later, the tone is tested alone…the CR won’t be as strong — compound stimuli will be stronger
If you really want a good response to the tone, condition the tone alone

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
44
Q

isolating one part of compound conditioning suggests that

A

cues CSs are competing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
45
Q

compound conditoning hints at

A

the importance of having salient cues

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
46
Q

salient cues

A

Salient cues = more intense. The more intense cue is more likely to get conditioned and be noticed

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
47
Q

importance of saliency related to what theory

A

mackintosh attentional theory of conditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
48
Q

CS competition — how to win the competition?

A
  1. Be more noticeable > overshadowing (because of saliency)
  2. Be there sooner > blocking (because of temporal priority) — when something happens first, in the sequence of conditioning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
49
Q

overshadowing

A
  • More salient cues get more learning in compound conditioning, still less than if trained alone
  • This brings us back to the idea of an information value of cues (cognitive interpretation): we use here the concept of information
  • If you really want to have a clear effect of one CS, try to isolate that completely and train it alone
    Neural networks/information theory involved here in classical conditioning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
50
Q

compound conditoning example

A

CSa + CSb > US is compound conditioning, ie. competing CSs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
51
Q

when two CSs compete, if one is more salient it will…

A

overshadow the less salient ones

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
52
Q

compound conditioning analogy

A

Picture book for kids. The book contains text and pictures, but the pictures will most likely overshadow the text
In other words, learning to read this way is not optimal

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
53
Q

compound conditioning spatial learning

A

Similarly, in spatial navigation, landmark cues may overshadow more general spatial cues — most of the time when you navigate proximal cues influence you most. The most important cues are the most salient

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
54
Q

compound potentiation

A

This is the opposite of overshadowing
In this case, there is more conditioning to a weak conditioned stimulus if it is combined with a more salient conditioned stimulus during conditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
55
Q

compound potentiation example

A

The conditioning of a weak odour may be strong if it is combined with a salient taste during conditioning
Eg. Gadbois aversion for even the smell of whisky in his taste aversion for whisky
Taste and odour go together — they are intimately linked stimuli. Smell is usually the early CS that something good/bad is coming.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
56
Q

kamins blocking effect example scenario

A

Scenario: visit in-laws every Sunday. Your FIL offers an anise based liqueur, but you drink it every time even though it makes you sick. One day, he offers you bagels with the anise.
Will you develop an aversion to bagels? No. Even if you get a bit sick, you will attribute your malaise to the anise not the bagels.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
57
Q

kamins blocking effect assigned variables

A

CSa: anise
Cab: bagels

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
58
Q

kamin blocking

A

you were exposed to the aversive first, then you combined the bagel (neutral). Only the aversive maintains the negative response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
59
Q

kamin original experiment

A

The experimental group gets the pairing as described above
When bagels are presented without anise (Test phase) almost no CR
Less response (CR) in the experimental group than the control group

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
60
Q

kamin what does prior conditoning do

A

prior conditioning with the anise has blocked conditioning to the bagels
Basically, the anise already predicted the sickness. Adding the bagels did nothing.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
61
Q

kamin what is the control group

A

the basic compound conditioning explained previously

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
62
Q

original experiment control group

A

compound conditioning group

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
63
Q

original experiment experimental group

A

pre-trained

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
64
Q

table in slides

A

basics of classical conditioning slide 56

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
65
Q

Is overshadowing different from blocking?

A

It occurs in only one conditioning phase

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
66
Q

overshadowing application

A

See application mentioned by Boynton for chemotherapy: scapegoating
Give patients a salient new flavour just before chemo (eg. Ginger candy)
This new flavour will overshadow any other flavour eaten in recent meals (just before the chemo)
The ginger candy becomes the scapegoat…problem is…the patient may never be able to stomach ginger candies in the future

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
67
Q

information provider

A
  • The CS-US association is an information provider unless the informational value is blocked
  • to be useful, a cue needs to be non-redundant (eg. Needs to be a non-redundant predictor)
  • Redundant cues are typically not very informant
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
68
Q

learning only occurs if…

A

the CS provides new information on the US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
69
Q

CSs may

A

compete with each other

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
70
Q

a more informative CS will win over

A

a less informative CS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
71
Q

Two areas of experimental psych have produced solid quantitative models, even “laws”

A
  • Psychophysics and the study of sensation and perception
  • Learning theory and the study of conditioning and learning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
72
Q

The Rescorla-Wagner model

A
  • most well known mathematical model in experimental psych and has been applied beyond learning theory
  • Tries to explain blocking
  • A lot about the unconditioned stimulus
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
73
Q

The Mackintosh model

A
  • A lot about conditioned stimulus
  • Attentional theory (saliency)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
74
Q

The Pearce-Hall model (including Kaye & Pearce)

A

US and CS modulation, attention, negative transfer

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
75
Q

Wagner’s original model (SOP) and AESOP

A

Priming (Attentional priming), affective dimension

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
76
Q

Rescorla-Wagner model

A

We know that cues compete with each other for associative strength
We know and will reaffirm that contiguity (in time and space) is not the full story. Information produced by the cues needs to be:
Reliable
Non-redundant

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
77
Q

Rescorla-Wagner model counterintuitive

A

difference between phase 1 and 2. In all experiments in learning theory we have an acquisition phase and after. In acquisition phase, you do want that CS/US pairing to be tight and predictable. Later, if the CS always announces the US, your animal loses attention (redundancy). Then work with the saliency of the CS/US or make it less predictable. Uncertainty is what facilitates learning after a while (NOT initially)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
78
Q

error detection and error correction

A
  1. Prediction error: The difference between what is predicted, and what actually happens
  2. Error detection: Ability to detect errors or discrepancies between the predicted and actual occurrences (detect prediction errors)
  3. Error correction learning: Similar to “trial and learning”. Learning to reduce the error or discrepancy between predicted and actual occurrences
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
79
Q

error correction learning

A
  • This type of learning applies to classical conditioning and many other types of learning, ie. skill and motor learning
  • This includes prediction error and error detection
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
80
Q

pavlovian classical experiment follows this scheme:

A

Bell — food, Salivate salivate
Or
CS — US, CR UR

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
81
Q

when conditioning to one CS generalizes to another we have…

A

generalization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
82
Q

generalization graphic

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
83
Q

Historically: two views on classical conditioning

A

SR learning
SS learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
84
Q

SR learning

A

CS-UR ~ note that this form of learning is known to occur in some case

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
85
Q

SS learning

A

US » CS = stimulus substitution theory or “cognitive” theories of classical conditioning; this view was confirmed by Rescorla in 1978.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
86
Q

SR vs. SS learning graphic

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
87
Q

Is it possible to respond to a CS (e.g., tone) that was never paired with the US (e.g., food)? (method)

A

One method can help answer this question: US devaluation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
88
Q

Is it possible to respond to a CS (e.g., tone) that was never paired with the US (e.g., food)? (forms of learning)

A
  • Second order or higher order conditioning
  • Sensory pre-conditioning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
89
Q

conclusion about US-CS?

A

association is essential!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
90
Q

figure from textbook?

A

???

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
91
Q

is sign tracking resistant to outcome devaluation?

A

Sign tracking but not goal tracking is resistant to outcome devaluation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
92
Q

We will see later that cues paired with a US/reward can…

A

gain “incentive salience”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
93
Q

what happens if you get a temporary seperation of the CS and US in classical condtioning

A

you will decrease sign tracking, but increase goal tracking.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
94
Q

second-order condiotning and sensory preconditoning graphics

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
95
Q

conclusion, learning can occur…

A

without an US > pro-SR learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
96
Q

second order conditioning example

A

You get bitten (US) by a dog (CS) » Fear (CR)
The later you associate the park (CS2) with dogs and get fearful (CR) entering the park

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
97
Q

second order conditoning graphic

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
98
Q

sensory preconditioning graphic

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
99
Q

sensory preconditioning appears to be…

A

pro SR learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
100
Q

sensory preconditioning analogy

A

Guilt by association” analogy
- Peter and John hang-out together. Peter is arrested for using cocaine. You assume John is using cocaine.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
101
Q

what happens with sensory preconditioning

A

wo neutral S get associated: The target CS (tone) never gets paired with the response it evokes.
In sensory preconditioning, the same two stimuli are paired before the light can even elicit a CR. So with that logic, it is a case of S-S learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
102
Q

sensory preconditioning is therefore

A

pure S-S learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
103
Q

generalization

A

concise def
???

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
104
Q

higher (second) order conditioning

A

concise def
???

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
105
Q

sensory preconditoning

A

concise def
???

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
106
Q

methods in classical conditioning: motor responses

A
  • Proboscis Extension Response (PER) conditioning to artificial odours in bees
  • Eyeblink conditioning in rabbits
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
107
Q

methods in classical conditioning: emotional responses

A
  • Fear conditioning in rats
  • Autoshaping in pigeons
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
108
Q

methods in classical conditioning: motivational responses

A
  • Appetitive conditioning in rats
  • Taste aversion learning
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
109
Q

eyeblink conditioning in rabbits associations

A

UR = eyeblink
CR = eyeblink
US = puff of air
CS = tone/light

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
110
Q

fear conditioning/threat conditioning/conditioned emotional response/conditioned suppression associations

A

UR = bar pressing (from training)
CR = fear/suppression of bar pressing
US = shock
CS = tone/light/noise

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
111
Q

conditioned fear

A

the fear response is often “freezing”.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
112
Q

the CS suppresses the bar pressing, therefore

A

The response (pressing the bar) is suppressed.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
113
Q

in conditioned fear, the suppresion is

A

the dependent variable, i.e., a measure of conditioned fear » suppression ratio

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
114
Q

suppression ratio

A

responding CS/(responding CS + responding pre-CS)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
115
Q

When the CS does not change the bar pressing rate, the S.R

A

= 0.5. A complete suppression is 0.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
116
Q

fear conditioning is an important concept in…

A

CBT for fear and anxiety disorders, even some elements of PTSD.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
117
Q

conditioned emotions are important in…

A

motivational aspects of instrumental conditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
118
Q

autoshaping in pigeons: sign tracking

A

focus on the CS’s, or cues predicting the US.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
119
Q

autoshaping in pigeons: goal tacking

A

focus on the USs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
120
Q

autoshaping therefore is actually

A

sign tracking

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
121
Q

autoshaping is a convenient way to train animals in

A

Skinner boxes. Alternative: successive approximations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
122
Q

signal-food conditioning protocol

A
  • You light the “pecking key” for a few seconds, then the pellet magazine opens.
  • After a number of pairings (45 or so, more or less), the pigeons will peck at the key (it announces the delivery of food, and it is signtracking… that cue becomes associated with the food).
    Note: the food delivery is not contingent upon the pecking. The pigeons does not have to peck the key.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
123
Q

appetitive conditoning in rats

A
  • This is the “magazine approach procedure”: CS (any) - US (food)
  • The approach is more likely to happen when the CS is produced. This is goal tracking as defined earlier.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
124
Q

taste aversion learning associations

A

UR = nausea
CR = nausea
US = drug injection (emetic drug or nausea/vomiting inducing)
CS = flavour S (saccharin)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
125
Q

taste aversion learning CS/US

A

CS-US can be separated in time by hours and still the conditioning will take place.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
126
Q

taste aversion clear and dramatic example

A

Clear and dramatic example in humans: Observed in humans with chemotherapy as well.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
127
Q

strength of conditioning factors

A

Time / timing
Novelty of CS and US
Intensity of CS and US
Pseudoconditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
128
Q

four main types of CS/US associations

A
  1. Delay conditioning
    • Short-delay
    • Long-delay
  2. Trace conditioning
  3. Simultaneous conditioning
  4. Backward conditioning
    • Other factor: Trial spacing
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
129
Q

types of conditioning graphic

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
130
Q

conditoning works better if…

A

the CS occurs before the US .

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
131
Q

the CS _____ the US

A

announces

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
132
Q

does backward condtioning work well?

A

typically no

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
133
Q

delay conditoning

A

The interval of 1me depends on the type of conditioning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
134
Q

delay conditoning example

A

Example: Eyeblink (very short) vs taste aversion (very long) conditioning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
135
Q

trace conditoning

A

Gap between CS and US = “trace interval”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
136
Q

is trace conditioning good

A

Fine procedure, but decreased efficiency with increased delay.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
137
Q

trace conditoning possible issues

A
  • The memory of the CS.
  • Inability for the animal to differentiate between the trace interval from the time between trials.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
138
Q

simultaneous condtioning

A

Full temporal overlap of the CS and US.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
139
Q

simultaneous condtioning validity

A

Usually not great results, but some exceptions.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
140
Q

backward conditoning

A

May be a signal of “no US” (i.e., a conditioned inhibitor).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
141
Q

backward conditioning may be associated with…

A

with relief in fear conditioning.
= “safety signal”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
142
Q

conditioning works best if…

A

trials are spaced-out over time (spaced trials); The ITI is crucial.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
143
Q

ISI?

A

interstimulus interval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
144
Q

ITI

A

intertrial interval

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
145
Q

are massed trials efective

A

no

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
146
Q

summary of conditioning

A

Short ISI’s
Long ITI’s

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
147
Q

ratio between ITI and duration of CS graphic

A
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
148
Q

novelty of the CS and US is…

A

very important: Pre-exposure to CS and US before conditioning can interfere with learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
149
Q

CS pre-exposure

A

“latent inhibition”. You can habituate to a potential CS with repeated exposures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
150
Q

US pre-exposure

A

Randich & LoLordo . Will delay subsequent conditioning. You can habituate to a potential US with repeated exposures.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
151
Q

CS and US pre-exposure both have

A

real-life implications, e.g., for fear/phobia acquisition.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
152
Q

intensity of the CS/US: US

A

The CR will be strong, if the US is strong. So the intensity of food or shocks, etc., will influence the magnitude of the response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
153
Q

intensity of the CS/US: CS

A

Strong salient CS’s will positively influence learning as well. So a strong tone, or flash of light, etc., will influence the magnitude of the response.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
154
Q

what is best for intensity of CS/US?

A

With CS’s salient, but not scary (overpowering) is good.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
155
Q

Counterfeit (fake) conditioning: Pseudoconditioning and sensitization
: What if the CS elicits a response like the one that you are trying to condition?
(Example)

A

You use CS = light.
Response (UR) = blinking.
Natural blinking or conditioned blinking?
So what is going on? » impossible to tell

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
156
Q

Counterfeit (fake) conditioning: Pseudoconditioning and sensitization : culprit 1

A

Sensitization

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
157
Q

Counterfeit (fake) conditioning: Pseudoconditioning and sensitization
: culprit 2

A

Pseudoconditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
158
Q

counterfeit/pseudo: sensitization example

A

Blinking for the camera. Initially a natural response, can get sensitized and anticipated.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
159
Q

counterfeit/pseudo: psuedoconditoning example

A

Mere exposure to the US!
“Increased responding that may occur to a stimulus whose presentations are intermixed with presentation of a US in the absence of the establishment of an association between the stimulus and the US”. (Domjan, 2015) Experiments need to control for both sensitization and pseudoconditioning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
160
Q

Rescorla and Wagner model: If no CS or a novel CS is followed by a US

A
  • The US will be unexpected
  • You have a positive prediction error
  • The larger the error, the greater the learning!
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
161
Q

Rescorla and Wagner model: If a well learned CS is followed by the expected US

A
  • The US is expected…
  • There is no error prediction (because no discrepancy)
  • There is no learning.
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
162
Q

Rescorla and Wagner model: If there is a CS-US association established, but then no US occurs

A
  • The absence of the US will be unexpected
  • You have a negative prediction error
  • There is a decrease in CS-US association
  • This is extinction
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
163
Q

What happens if you trained a dog every time the CS is produced the US follows (clicker training, phase 1)

A

he theory is that technically it would suggest the CS (clicker) itself is reinforcing. After a while the dog might not even need the treat, the click is reinforced enough. If the association between the US/CS is weakened, therefore the clicker would not become as good of a predictor overtime and would lose its reinforcing value.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
164
Q

Rescorla and Wagner ___ predict that the CS can be the ____ itself

A

do not, reward

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
165
Q

rescorla and wagner…

A

do not think clicker training would work if the clicker is used in isolation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
166
Q

Basic assumptions of RS

A
  • Each CS has an association weight, or strength of association, with the US.
  • Cues (CS’s) compete for associative strength.
  • This tells us that information, or the surprise value of the stimuli involved in important
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
167
Q

key element in RW

A

the element of suprise!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
168
Q

what is surprise RW?

A
  • What is surprise? It is what is unexpected.
  • Part of this may be about attention — how much attention are you giving to elements in that model
  • Either you think its coming or its not, or vice versa, or a completely novel experience
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
169
Q

premise of surprise RW

A

The US needs to be “surprising”. With time, the US, in any conditioning sessions, becomes less and less “surprising”.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
170
Q

why is surprise needed?

A

Why? Because the CS becomes a solid predictor of the US.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
171
Q

example of CS being solid predictor o US

A

clicker training in dogs

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
172
Q

learning wont occur anymore when…

A

the CS predicts the US perfectly.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
173
Q

suprise clicker training example

A

Clicker training, your click pouch should be full of whole bunch of different intensity treats (surprise element). Slight variation of the US is good — this is for appetitive learning
Could apply to a person as well. If a CS announces a shock, same principles apply especially if it is always the same small shock. If you start varying the intensity of the shock, then the person will start paying attention again.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
174
Q

If the CS/US link is highly predicable there is…

A

no learning, learning depends on not being exactly sure how that works.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
175
Q

what happens with repetition in learning

A

Repetitions get boring after a while, if you can always predict things then you start getting bored. That means though that you have already learned the task. But how do you keep a strong response/motivation — this is where unpredictability might help
- You don’t want the response to start to extinguish
- But if you maintain the link between CS/US it shouldn’t extinguish

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
176
Q

how do you avoid extinguishing of connection

A

maintain the link between CS/US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
177
Q

Does it matter if “learning does not occur anymore”?

A

you are maintaining the strength of the association… so maybe not.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
178
Q

Does that mean that you can drop the US?

A
  • Then you create a negative prediction error (see above)
  • There is a decrease in CS-US association…
  • So a US is important (at least some of the time…)
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
179
Q

what is important in surprise/learning

A

Uncertainty/intermittent reward is good for this

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
180
Q

If you drop the US you get…

A

a negative prediction error, you are affecting the CS/US relationship

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
181
Q

mathematical model, a.k.a

A

the neural network-like model:

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
182
Q

mathematicla model peak of learning is

A

the associative strength between the US and the CS.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
183
Q

how is associative strength developed

A

Based on strengthening of the relationship between the CS and US. The more you do click/treat, with time the dog learns that the click predicts the treat.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
184
Q

learning ceiling is denoted by ____, it is

A

This learning ceiling effect is called lambda, it is the asymptote of the learning curve.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
185
Q

what is the learning ceiling

A

When dog completely knows that click means treat
Asymptote is very much determined by the value of the US (Whether it is good or bad)
If it was about shocks, probably a more intense shocks get you to a higher lambda.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
186
Q

what is really important in lambda

A

All about the US — that is the important part of this. Relatively of the US is very important

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
187
Q

the associative strength of the US and CS is V (y axis) =

A

predictive value

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
188
Q

the asymptote is determined by

A

the value of the US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
189
Q

Change in associative strength (formula)

A

△V = ⍺β(ƛ-V)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
190
Q

A

CS = ⍺ (it comes first) = salience of CS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
191
Q

β

A

US = β = salience of US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
192
Q

V

A

V = associative strength = learning parameter = how well the CS predicts the US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
193
Q

ƛ

A

ƛ = ceiling effect (how much that animal is going to learn) — stands for the US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
194
Q

A

△ = delta = change

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
195
Q

(ƛ-V)

A

(ƛ-V) = surprisingness of the US = prediction error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
196
Q

V = ⍺β(ƛ-V) is pretty much

A

the Rescorla Wagner model

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
197
Q

(ƛ-V)

A

The difference between the US and what the CS predicts (ƛ-V) is called the prediction error.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
198
Q

as trials of CS-US associations accumulate

A

the CS becomes a better and better predictor of the US.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
199
Q

the more pairings of the CS/US…

A

the easier it is for the animal to learn

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
200
Q

with time, the prediction error becomes

A

smaller and smaller

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
201
Q

classical conditioning =

A

djusting the prediction error

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
202
Q

If you have a large (magnitude) US (𝛃), lambda will be…

A

larger too (higher asymptote)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
203
Q

More likely to hit the lambda with…

A

a strong/salient US

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
204
Q

So how can you get the blue line in clicker training? (larger magnitude)

A
  • Use a strong US
    e.g., in dog training, liver treats instead of regular kibble
  • Easier to make you get nervous about the possibility of getting a shock when the shock is very painful
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
205
Q

can lambda move?

A

The lambda can move (you can bypass the ceiling of (1.0), all depending on the strength of 𝛃

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
206
Q

if the CS is really salient (⍺)…

A

the learning will be faster (i.e., it affects the rate of learning).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
207
Q

how can you get faster learning

A

get a salient CS (⍺)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
208
Q

salient CS example

A

e.g., in dog training, a stronger, louder, better defined (crispier) clicker sound.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
209
Q

strong CS, great US =

A

lot of learning and fast learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
210
Q

can boredom occur over time

A

With time is it possible things get boring, boredom can take place. It is technically possible that after a while these parings get boring. Then you could replace the clicker with a whistle or something to add an element of surprise/novelty.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
211
Q

A strong US (liver treats), or 𝛃…

A

will impact the associative strength/asymptote or lambda (𝛌).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
212
Q

A salient CS (crisp click), or 𝛂…

A

will impact the rate of learning.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
213
Q

is it possible for other CSs to be present during conditioning?

A

yes

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
214
Q

If 2 or 3 CS’s are present…

A

they all contribute to the conditioning (and may compete… as you may remember with blocking)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
215
Q

how did they add the sum of associative strengths to the formula?

A

This is why they added sigma to the formula, the sum of the associative strengths — acknowledge the potential of sum variability in adding CS.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
216
Q

when might a number of associative strengths be at play

A

In the acquisition of a traumatic experience it might not always be clear what caused it, but rather an association of many things

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
217
Q

compound conditioning is also known as

A

potentiation effect

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
218
Q

formula with compound conditioning

A

𝚫V = 𝛂𝛃(𝛌-𝚺V).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
219
Q

what does sigma refer to

A

sum of

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
220
Q

the Kamin blocking experiment: phases

A
  1. Phase 2: The tone competes with the light
  2. Phase 3: The tone loses
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
221
Q

why does the kamin experiment play out that way

A

Why? The light already accurately predicts the US (shock).
- The light already acquired the strength of the CS/US pairing

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
222
Q

Medical diagnosis of dietary (GI) intolerance example

A

Phase 1: Patient drinks milk » stomach ache
Phase 2: Patient drinks milk + and has garlic » stomach ache
Phase 3: Could the garlic cause the stomach ache?
Diagnosis: “Unlikely, eat garlic, but avoid milk”

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
223
Q

Think of this example: Medical diagnosis of dietary (GI) intolerance – implication (garlic/milk)

A

Implication (cognitive): Clearly the physician is thinking of a milk allergy here. But… if she/he had considered that both milk and garlic are common triggers for IBS (therefore a milk — and garlic — intolerance, not a milk allergy), then garlic would have been seen as a potential player.
- Have to test things one by one. If you start combining things you will never be able to isolate the cause — not just because of the metabolic response, but also the chance of classical conditioning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
224
Q

Suppression ratio

A

There is less learning with a high suppression ratio.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
225
Q

If a new stimulus is presented with an other that was previously conditioned…

A

then little or no conditioning will occur (resulting in a high suppression ratio).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
226
Q

Remember that you initially pair a tone or noise (CS) to a shock (US).
Then, the same noise AND a light are paired to the shock (US).
the outcome is…

A

that the response to the light is not well acquired (high suppression ratio).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
227
Q

what did the RW model address

A

Rescorla-wagner explained blocking, big victory. However, it does not explain a lot of other things.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
228
Q

how did the RW model explain blocking

A

By showing that the second stimulus does not acquire an associative value in the second wave

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
229
Q

how to unblock

A

modulate the US (saliency of US is the focus of the model)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
230
Q

UNBLOCKING =

A

Increasing the US (shock) in phase 2.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
231
Q

neural network explanation

A

???

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
232
Q

what decreases during extinction

A

In other words, V, or the associate strength, will decrease during extinction

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
233
Q

what happens to lambda in extinction

A

So lambda becomes 0 (there is no US anymore)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
234
Q

what happens when you reintroduce a US

A

There is very strong spontaneous recovery as soon as you reintroduce the US. What may actually work is if you repair the clicker with the food (marking or keep going signal). Fear is that if you never pair the clicker with the US again, it may change the meaning of the click. The meaning of the click is how it was associated with the US. What if sometimes you click and its not followed by food, but sometimes you click and it is followed by food. That model here does not address that.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
235
Q

Conditioned inhibition à la Pavlov:

A

L (CS) + Shock (US)
L (CS) + N (CS) + no US > the noise becomes an inhibitor

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
236
Q

Protection from extinction

A

If you combine an inhibitor with an excitor during extinction, the inhibitor will protect the excitor from losing associative strength (V).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
237
Q

what does protection from extinction help to explain

A

This may explain relapse in exposure therapy for fear or phobias — you would try to extinguish the link

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
238
Q

in exposure therapy for fear/phobias you would…

A

extinguish excitors. While doing so, you may have cues that become inhibitory.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
239
Q

initially during fear conditioning…

A

it looks like you have no more fear. But if later you test the excitor by itself, then fear comes back.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
240
Q

why does that occur during fear conditioning?

A
  • Well remember the concept of surprise.
  • The noise (inhibitory) predicts no US (no fear). But if you present the light again… fear is back!
  • Surprise is the key for learning!
  • In some cases you can have two excitors
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
241
Q

what happens if you have two excitors?

A
  • Then you have a very efficient extinction!
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
242
Q

To get very efficient extinction trials…

A

combine the CS with other excitors (not inhibitors).

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
243
Q

negative contingency

A

US is less probable in the presence of the CS
- This scenario leads to inhibitory conditioning of the CS

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
244
Q

neutral contingency

A

US is equally probable in the absence or presence of the CS
- This leads to no learning

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
245
Q

Anytime you have a CS or US that is conditioned

A

context is always processed at the same time (comparator theory)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
246
Q

Mackintosh and Turner (1971)

A

How the attentional theory of theory (attentional learning of classical conditioning)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
247
Q

mackintosh and turner groups

A

Group 1: used noise followed by shock, no phase 2, light noise than much better shock, then test light and see what happens
Group 2: noise followed by shock, light/noise shock, light noise than much bigger shock, then test

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
248
Q

what happened in mackintosh and turner experiment

A
  • Based on Rescorla and Wagner, the groups should not have differed, but the control group (1), sees more conditioning (remember, with the SR, low scores indicate more conditioning)
  • Group 1 (control) learns because of unblocking, i.e., the use of a larger shock.
  • Group 2 (experimental): The LN - Shock trial interfered with the learning.
  • Group 2, phase 2: Subjects learn that the light (L) is redundant, so they pay less attention to it!
How well did you know this?
1
Not at all
2
3
4
5
Perfectly
249
Q

what were mackintosh and turner trying to show

A

that we learn to ignore redundant predictors of a US, good or bad.

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
250
Q

what do we learn to do with redundant predictors of a US

A

learn to ignore them!

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
251
Q

So it may not be about the ineffectiveness of the ___ (RescorlaWagner), but rather the ineffectiveness of the __ (Mackintosh).

A

US, CS

252
Q

so how well does the CS predict the US?

A

If the CS is a good predictor, then we pay attention to it. If not… we don’t

253
Q

Attention paid to the CS

A

how well the CS predicts its consequences

254
Q

___ is the predictor

A

CS

255
Q

Mackintosh model is all about the…

A

attention we pay to the CS

256
Q

the more ___ the ___, the faster the learning.

A

salient, CS

257
Q

moral of the story

A

US and CS are important

258
Q

Group dog clicking classes — how does it work if everything you hear in the room is click?

A

Think about it. How does the dog know which click is theirs? There must be a mechanism for that dog to understand which click is important to them. Attention is important!

259
Q

problems with RW model?

A
  1. Does not explain well extinction of inhibition
  2. Does not explain well latent inhibition (pre-exposure to a CS)
  3. May have some issues even with the blocking effect
  4. Miller, barnet & Grahame identity 23 predictive failures of the model
260
Q

who raised issues with blocking effect of RW?

A

Mackintosh raised some issues with how the US needs to be surprising, thinks attention is more important

261
Q

mackintosh and turner main idea

A

The more you repeat something, the better the outcome — talking about elemental isolated stimuli that get a certain type of learned irrelevance

262
Q

mackintosh and turner finding about CS?

A

CS that is not redundant (or surprising) may be good

263
Q

what is at the forefront of mackintosh and turner model?

A

CS

264
Q

conclusion of mackintosh and turner?

A

we learn to ignore redundant predictors of a US
- So it may not be about the ineffectiveness of the US but rather the ineffectiveness of the CS

265
Q

so how well does the CS predict the US?

A

If the CS is a good predictor, than we pay attention to it. If not…we don’t

266
Q

cognitive: attention paid to CS «<

A

how well the CS predicts its consequences

267
Q

mackintosh model?

A
  • Attention to the CS = alpha in the RW formula
  • CS salience (alpha) is key, the more salient — the faster the learning
  • Game is to pay attention to the CSs that are the best predictors of the US
268
Q

what is latent inhibition the same as

A

CS pre-exposure effect

269
Q

what is latent inhibition/CS pre-exposure effect?

A

if a stimulus is not suprising anymore, after a while you will stop paying attnetion

270
Q

does the mackintosh model always work?

A

no. thats the problem

271
Q

the pearce hall model…

A

changed the landscape of models for classical conditioning

272
Q

pearce hall two groups?

A

Group 1 = tone and shock, tone and SHOCK
Group 2 = light and shock, tone and SHOCK

273
Q

pearce hall results

A

We know here that group 1 should learn quickly, because the tone is present in phase 1 and 2, with a US of stronger value. What we have here is the opposite due to negative transfer

274
Q

premise of pearce hall model?

A

Why pay attention to a well established CS? (Elements of mackintosh here)
Instead, we pay attention to unknown/unconditioned CS — really coming back with idea of surprise

275
Q

main point of pearce hall

A

organisms attend to reposed to uncertain predictors

276
Q

pearce hall perspective aligns well with…

A

modern attentional thoery

277
Q

what is alpha based on in pearce hall model?

A

As in mackintosh, alpha is attention but the value of alpha is based on how surprising the US is — important because alpha brings attention in

278
Q

what matters in pearce hall model

A

both CS and US, could argue all elements of RW model are here

279
Q

kaye and pearce study

A

Step 1: Condition rats to light CS — initially orient to it (eg. Sign tracking). If they orient to it it means they are paying attention
Step 2. Three groups
Continuous
Partial
None

280
Q

continuous group?

A

CS/US pairing on every trial (R&W in support of this)

281
Q

partial group?

A

50% CS/US paring

282
Q

none group?

A

CS never paired with US

283
Q

kaye and pearce study results

A

Found that partial pairing actually worked the best. Continuous pairing is fine but not great, no pairing does not perform with. Partial pairing is really when the rats are paying attention, in other words, a little uncertainty is not bad.

284
Q

conclusion of pearce/hall and kaye/pearce

A
  • How surprising the US is on preceding trial is crucial
  • How surprising the US is, depends on the CS
  • Little bit of uncertainty may actually maintain responses
  • The vale of alpha remains high
  • US component (R&W), CS component (Mackintosh) — this pairing creates uncertainty and ends up being in your advantage
  • Explains negative transfer
  • Explains latent inhibition
285
Q

The Pearce and hall model, under the kay and Pearce experiment reinforces that…

A

uncertainty is a good thing and both CS/US are important

286
Q

pearce and hall model summary of negative transfer: two groups?

A

Group 1 = tone and shock, tone and SHOCK
Group 2 = light and shock, tone and SHOCK

287
Q

pearce and hall model summary of negative transfer: phase 1

A

Phase 1: CS becomes a perfect predictor of the US
- Problem, US becomes unsurprising with time — the repetition, fact that CS is losing some of its predictive value is problematic. With time, less and less attention is payed to the CS.

288
Q

pearce and hall model summary of negative transfer: alpha value at beginning of phase 2

A

is very low

289
Q

pearce and hall model summary of negative transfer:

A

initially little learning is possible
- Negative transfer occurs: you pay less and less attention to the CS

290
Q

pearce and hall model summary of negative transfer:

A

Synthesis is to in a sense, take all of the components together
Importance of CS comes together with Pearce and Hall

291
Q

what are we paying attention to? according to pearce and hall

A
  1. Cues with high predictive value (mackintosh)
  2. Cues high in uncertainty (pearce-hall): organisms attend to uncertain predictors
292
Q

what does RW suggest (main idea!)

A

Surprisingness of the US — high saliency of US gets you to pay a lot more attention to predictors

293
Q

attnetion is important but depends on…

A

previous learning

294
Q

two things we have learned…

A

can condition attention AND can condition motivation as well. Cognitive processes of attention and motivation can be conditioned, these are essential to learning

295
Q

modern version of short term memory?

A

working memory

296
Q

who proposed working memory?

A

werner honig

297
Q

what is working memory?

A

Memory that you use as you are doing a task. Could include rats in a skinner box, or people being given a phone number. Form of short term memory that stays as long as the info is relevant (usually not very long)

298
Q

how does sensory memory go to STM?

A

self-generated priming

299
Q

how does LTM go to STM?

A

retrieval-generated priming

300
Q

what does wagner do in his new model alone?

A

Brings back the surprise element here
Brings back primed component of STM

301
Q

wagner priming (problem?)

A

If you are primed, in most memory tasks that primed element will be easier to retrieve from memory later. Here, a primed element is less surprising — therefore we have a bit of a problem

302
Q

two types of priming

A

self generated
retrieval generated

303
Q

self generated priming?

A

via sensory memory (basic structure of the stimulus)

304
Q

retrieval generated priming?

A

via LTM (actually from a complex retrieval process of info — already most likely in LTM)

305
Q

surprise can be reduced in ___ ways

A

2

306
Q

retrieval generated priming?

A

From long term memory via a retrieval cue

307
Q

self-generated priming?

A

From sensory memory via a recent presentation

308
Q

blocking =

A

retrieval generated priming

309
Q

retrieval from LTM?

A

the US is primed by the CS

310
Q

what is priming of the US with the CS?

A

blocking

311
Q

priming can explain…

A

blocking
- because the surprisingness of the US is dampened

312
Q

continuous learning =

A

in the moment, self-generated

313
Q

long pause, later CS is presented =

A

long term memory, retrieval generated priming

314
Q

priming of the CS

A
  • RW: combined processing of the CS and US in STM is important
  • Wagner thought that maybe he could look into latent inhibition (CS pre exposure effect again)
  • With this model, exposure to the CS before conditioning should reduce the CS surprisingness.
315
Q

wagner can explain…

A

habituation

316
Q

wagner model what is habituation

A

habituation is simply a decrease in the “surprisingness” of the stimulus with both types

317
Q

SOP

A

Standard operating procedure or sometimes opponent process

318
Q

what kind of model is SOP?

A

This is a connectionist model, very much a neural network model — parallel processing

319
Q

who massively influenced SOP?

A

donald hebb

320
Q

what was SOP originally called?

A

PDP or parallel distributed processing

321
Q

what does PDP entail

A
  1. Processing in the brain is parallel, not serial (like the information processing models always suggested).
  2. Processing in the brain is distributed, not localized or modular (like emphasized in [clinical] neuropsychology and evolutionary psychology).
322
Q

Those models (PDP) have been very successful in some areas of cognitive psychology and neuroscience…

A
  1. Learning and memory (including priming)
  2. Perception / pattern recognition
323
Q

But (PDP) not so successful in other areas (mostly “higher” cognitive processes):

A
  1. Language
  2. Problem solving
324
Q

what does SOP do that is new?

A

address time

325
Q

SOP (3 features)

A
  1. Addresses the timing of the CS and US
  2. Addresses backward conditioning (US presented before the CS). Explains how you can get conditioned inhibition; timing is still important here for this effect.
  3. Compound conditioning: Leads to inhibition.
326
Q

AESOP who proposed?

A

wagner and brandon

327
Q

AESOP adds?

A

Wagner and Brandon added the emotional qualifies of the stimuli, particularly the US.

328
Q

what does AE mean in AESOP?

A

affective extension - emotional aspect

329
Q

AE suggests that new nodes are activated by a CS

A

sensory node
affective node

330
Q

result of AESOP

A

CS gets linked to both of these nodes

331
Q

Potentially every time you have CS/US paring…

A

you are tapping into the sensory system, likely near the limbic

332
Q

AESOP is the only theory that says…

A

pairing can be both sensory and emotional

333
Q

what is AESOP useful in discussing

A

It is particularly useful when discussing US’s that may have a strong emotional valence (positive or negative).

334
Q

what can AESOP help to explain

A

how the emotional response may potentiate the sensory response

335
Q

AESOP and trauma?

A

So this adds an other variable/factor to conditioning. Think about flashbacks in PTSD!
Physical stimuli and emotions being combined

336
Q

Suddenly AESOP is much about the ___ or ____ and is less stuck in a “____” view of the US.

A

Suddenly this theory is much about the “sign” or “stimulus” and is less stuck in a “poverty of the stimulus” view of the US.

337
Q

what is also responsible for poverty of stimulus?

A

Cognitive psych is also responsible for poverty of stimulus because it focuses on representation of the stimulus by itself. Here we are saying that none of that is terribly valid if you don’t take the emotional value into consideration.

338
Q

what is the parallel processing dimension of AESOP?

A

simultaneous sensory and affective processing.

339
Q

what does AESOP explain

A

Explains why conditioning works better (often worse), with strong reactions — eg. Trauma

340
Q

potentiation between sensory and affective suggests that..

A

limbic system is working with sensory areas of brain

341
Q

CS in AESOP

A

They (Wagner & Brandon) were influenced by Pearce (from Pearce and Hall)

342
Q

pearce came up with?

A

configural thoery of learning

343
Q

what is configural learning

A

The set, or configuration, is what is conditioned to the US

344
Q

what is elemental learning?

A

Each CS conditioned independently to the US

345
Q

pearce configural theory of learning?

A

configural theory of learning — means that if you have many CS, Pearce saw them as typically being a whole. In other words, they are more than just the parts (elemental view)

346
Q

the configurational perspective is ___ in nature

A

gestalt

347
Q

configurational perspective gestalt

A

It means that the whole is more than the sum of its parts.
Example: Play a chord on your piano or guitar, say A minor.
- Means that configurations are meaningful and you can actually be conditioned by configurations

348
Q

So compound stimuli, especially if in the same sensory modality, are…

A

more that the sum of the elements.

349
Q

So just adding different CS’s just…

A

weakens the value of the single, elemental CS (they are in fact, literally not the same).

350
Q

From a neural network theory point of view, and consistent with SOP, if you condition ABC as CS’s, but then just test AB, you would …

A

lose 33% of the conditioning (from R-W, remember that CS’s share a lambda).

351
Q

order of progression of theories

A

Wagner’s original theory ~ SOP ~ AESOP

352
Q

AESOP explains

A
  • Compound conditioning effects (Rescorla-Wagner)
  • Attentional and priming effects (Mackintosh, Pearce-Hall)
  • Time constraints in conditioning
  • Sensory and emotional dimensions of conditioning
  • Stimulus configuration, and even generalization
353
Q

operant conditioning, focus on…

A

Focus on response-outcome learning

354
Q

applications of operant conditioning

A
  • Parenting
  • Training (animals, humans, etc.)
  • Teaching (behaviourist methods)
  • Cognitive Behavioural Therapies (and others), behaviour modification
  • Programming of gaming software, gambling games and machines (e.g., slot machines)
  • Advertisement
  • Any social situations (making compliments, ignoring people, etc)
  • The tipping game!
    Etc. (so many more examples)
355
Q

differences between classical and operant conditoning

A
  • Contingency between outcome and response is extremely clear in operant — the is why an animal learns
  • In classical conditioning this is very fuzzy because it is often automatic/unconscious
356
Q

similarities in classical and operant conditioning

A
  • Both have a learning curve.
  • Both have extinction.
357
Q

thordnike theory

A

Thorndike: Reinforcement theory » law of effect.

358
Q

guthrie theory

A

Contiguity theory » Contiguity is enough.

359
Q

guthrie contiguity examples

A

(Why the dog doesn’t learn to hear the porcupine? Because of lack of contiguity in time and space)
Often dogs get hit by porcupine, may feel little bit of pain but there is an analgesic in the quills. Very quickly the animal doesn’t feel pain, the more they start moving the more the quills go in. Why is it that my dog is not learning the porcupine is a big thing from the initial hit? Because of continuity. They end up learning the most pain they experience that day is hours later at the vet clinic. That is what they learn to be afraid of.

360
Q

stimulus sampling theory

A

stimulus elements are at the centre of this theory:

361
Q

Learning curves show an improvement because…

A

the organism notices more stimulus elements.

362
Q

for guthrie, a reinforcer is…

A

just a very salient stimulus — we know it is more than that now.
How close time and space associations are happening is the key for Guthrie

363
Q

tolmans approach

A

One of the first cognitive theories (uses the term mental map or cognitive map).

364
Q

for tolman behaviour is…

A

inherently flexible.

365
Q

tolman rats main point

A

He showed that rats in mazes learn better about places than responses
Main point: They can learn on non-reinforced trials. It becomes obvious when the reward is provided. This is called latent learning

366
Q

learning is not…

A

performance!

367
Q

the role of the reinforcer is to…

A

provider motivation (idea developed by spence and hull)

368
Q

tolmans approach order of variables

A

SD > R > O
SD = discriminative stimulus
R = response
O = outcome

369
Q

discriminative stimulus gives…

A

a response for a specific outmode (Eg. Kibbles in spinner box) — operating conditioning. Contingencies must be vert explicit.

370
Q

thorndike box SD/R/O values

A

SD = box
R = sequence of events to open
O = outcome

371
Q

for thorndike and guthrie ____ associations

A

SR

372
Q

for tolman ___ associations

A

SS

373
Q

rewards thorndike idea?

A

they reinforce

374
Q

rewards guthrie?

A

they are a very salient stimulus

375
Q

rewards tolman/hull/spence?

A

they motivate

376
Q

what do rewards actually do (consensus)

A

In fact they are/do all three — reinforce, salient stimuli and actually motivate.

377
Q

two types of priming

A

positive and negative

378
Q

positive priming

A

pre-exposure to stimulus facilitates received of that stimulus

379
Q

negative priming

A

Wagner SOP
the exposure to the stimulus gets in the way of learning

380
Q

negative transfer

A

when any type of prior knowledge interferes with current knowledge (learning)

381
Q

pearce-hall negative transfer

A

Peace-Hall negative transfer: pairing a CS with a weak US slows-down conditioning when the CS is subsequently paired with a strong US

382
Q

classical view of dog bit situation

A

CS is dog, US is bite

383
Q

AE-SOP view of dog bite situation

A

CS is still dog, now there is two US — one purely sensory (pain), one affective (learned fear)

384
Q

AESOP explains ____ very well

A

AESOP explains conditioned emotional responses, often based on fear conditioning, very well

385
Q

contiguity

A

all about spatial and temporal (how close in space and time things are)
- CS co-occurs with the US: they are contiguous, or close together, in space and time

386
Q

contingency

A

the CS predicts the US: the occurrence of the US is contingent on the prior occurrence of the CS.

387
Q

two main methods of a skinner box

A

discrete trial method
free-operant method

388
Q

discrete trial method

A

Stand-alone trials, and many of them, in a session. This controlled by the experimenter.
- Eg. Gadbois lab — presenting dog with a problem. Accumulate trials

389
Q

free-operant method

A

The animal or the human control the apparatus/computer.
- Typically used for skinner box — animal controls flow of experiment by giving specific responses to specific stimuli

390
Q

free operant set up

A

Typical learning curve as response rate showing acquisition and extinction (when you stop producing the kibbles). — response for minute
The cumulative recorder taking the cumulative responses.
As you learn they increase, as it extinguishes they decrease

391
Q

SD → R → O

A

Able to identify SD as something that should provoke a response and outcome

392
Q

S△ → R → ØO

A

The other stimulus, the one you don’t want to respond to because it won’t provide an outcome

393
Q

Sdelta or?

A

S-

394
Q

Operant way of teaching discrimination (nose hold in position) example

A

Eg. Want a dog that can find specific snakes (eg. Gartner snake) in a park that is full of different kinds of snakes. Need dogs to tell difference between ribbon and garter snakes. How do you teach a dog to discriminate just ribbon snakes? We rely on their nose. It is very simple, show the ribbon snake smell and encourage the dog for sniffing (give treat or click). Now grab the other stimulus (garter snake) and do not encourage the dog for sniffing (no treat). Do you train no response (just walk away) or a distinct response for a no (train them to sniff, then sit down because it is wrong). S+ or yes response would be hold nose.
Contrasting approach: systematically reinforcing a response where they initially just show interest and ignore with the S-.

395
Q

concept of discriminative stimulus

A

Light on » Press lever » get food
Light off » Press lever » no food
So after a while, the rat presses the lever only or mostly when the light is on.

396
Q

what does it mean when an operant response is brought under stimulus control

A

a stimulus will set the occasion for the response.
- The light is a discriminative stimulus (also labelled S^D or S+): It is associated with reinforcement. The other stimulus is the S delta or S-.

397
Q

occasion-setting in discriminative stimulus is…

A

operant

398
Q

____ is necessary for all sorts of operant learning

A

stimulus control

399
Q

shaping, autoshaping and ____ behaviours according to skinner

A

superstitious

400
Q

superstituous behaviours example

A

If there is too much of a delay before giving reward, can occur because they have to hold their nose for 5 seconds, you can develop superstitious behaviours. Smarter breeds are more likely to develop superstitious behaviours (border collies) — because they are trying to determine what you want.
Idea is that you need to extinguish the response, may feel like you are going backwards. You have to get them to step back, do a normal entrance, then reinforce. How do you do this? Often going backwards. Get them to enter, put them right in front of the vial before the trial starts and reshape just the nose hold — not the whole entrance. This helps to reduce the possibility of reinforcing the wrong behaviour.

401
Q

habit slips

A

In maze running, food encountered in the maze won’t stop the running to the goal (the end that is baited).
- If you put the reinforcement before the end goal, often the rat will just run by it and complete the task. That just means they have really learned the task and developed the habit of going through the full actions (almost the opposite of superstitious behaviours in some ways)

402
Q

Habit slips, or strong discriminative stimuli »

A

response associations are not uncommon.

403
Q

habit slip example

A
  • Going to pick-up somebody at the airport but you take the Dartmouth Crossings exit because you go there often.
  • These behaviours can be really hard to extinguish
404
Q

the ______ of the response, does not matter if you focus on the goal

A

specificity

405
Q

lashley showed…

A

that rats that learn to run dry maze, will immediately learn to swim it if it is flooded.

406
Q

can all dogs trained in the lab transfer to the field

A

no only about 50% - dog knows the odours but the change in context is enough to confuse them

407
Q

transfer of learning is a form of

A

generalization

408
Q

specificity of response errors in change of context

A

might have to do with focus of control, or sign tracking/goal tracking

409
Q

EDT

A

errorless discriminative training

410
Q

EDT Terrace example

A

The dog learns to respond to the S+ (e.g., scent of lavender).
The S- (scent of oregano) is introduced early.
The S- is not presented in the first trial, and then is presented in weak form at first, and gradually strengthened: this is “fading” (fading-in).

411
Q

attentional dimension of EDT

A

Manipulation of the saliency of the S-. Difficult to do if you don’t have the right equipment

412
Q

EDT is all about

A

what you are getting the animal/human to pay attention to. Based on simple premise — during learning you commit the least amount of error possible and also the outcome is that (often) the performance of the animal/human will be with less errors than with typical discrimination training.

413
Q

EDT example (long)

A

S+ (one you want the dog to respond to) is aways shown at full strength. You present the two stimuli at the same time. S+ (lavender) always shown at full saliency, but not the oregano. Present a very weak, varying, amount of the S- (oregano). Sometimes they will look towards the oregano, but ignore that and they typically go towards the more salient stimulus. Idea is that through all the trials, you make the saliency of S- stronger until they are at the same level. This works because you are indicating right away to the dog which stimulus is important and which to ignore. In some cases, they barely learn to ever respond to the S- so by the end of the task they have responded only to the S+, even though the S- is at full strength at the end.

414
Q

why is EDT considered to be errorless

A

you diminish the amount of potential error that can be created

415
Q

EDT tried to be used with what population

A

This got the attention of people working with a autistic children. If you go with trial/error method they often get very frustrated when it is not reinforced. How to teach autistic children without them making errors, created techniques that resembled this in some ways.

416
Q

main advantages of EDT

A

Large reduction of errors compared to traditional discrimination learning
Less mistakes are possible during training.
No negative emotions during training (e.g., frustration, helplessness, etc.)
The training is potentially very fast

417
Q

EDT conclusions

A

Mistakes are NOT necessary for learning!
Evidence from neural networks that sometimes it is
It is likely an attentional mechanism: You learn (early!) to ignore the irrelevant stimulus and basically never respond to it (or at very low rate and only initially in the training)
Trial and error is NOT necessary, in principle you can learn without the error just by focusing on attentional processes right from the get go

418
Q

main disadvantages of EDT

A

Modifications or reversals in training are difficult: e.g., if lavender (S+) becomes the S- and you try to train the S- (oregano) as the S+. Acquisition of the S+ will be delayed and/or impaired.
Example: Training a dog to discriminate (for field search and detection) a specific species of snake within the Thamnophis genus.
If you ever want that dog to switch to Garter snakes as the target species, it may be challenging.

419
Q

___ of dogs wont train through EDT easily

A

2/3

420
Q

examples of how classical conditioning is involved in operant conditioning

A

Money — there is a delay, humans still understand this
The clicker in animal training or TAG teach (TAG = Teaching with Acoustical Guidance; e.g., used to train movement precision in dance or sports) — make the exact moment they did the exact right movement

421
Q

the clicker is a ______ stimulus

A

classically conditioned discriminative

422
Q

the clicker is a _____ for food (US)

A

pavlovian CS

423
Q

why is the clicker used in operant conditioning

A

because it has acquired a reinforcing value in itself

424
Q

where did the idea of using just the clicker come from

A

Situation arises first when trying to train orcas and dolphins in captivity — idea is that you dont want to interfere with animals during training. How do you tell them “yea you got that right, continue”. The conditioned reinforcer was created, not giving a primary but marking it with a secondary (the clicker or a whistle)
That in itself is literally telling you that the clicker does not need to be followed by the reinforcer, at least every time.

425
Q

history of conditioned reinforcer

A

When pigeons were in the Skinner box, when being shaped and giving the right response, the food magazine would drop but make a click while doing so. Quite accidentally, researchers noticed that when the magazine would empty, pigeons would continue to respond for a long time, and most likely we responding to the click as a conditioned reinforcer.

426
Q

what are conditioned reinforcers now used in

A

Used a lot with kids, eg. Stickers.
These are secondary reinforcers — they have been conditioned to come with a reinforcer later.

427
Q

CS/US link may ____ over time if you stop paring the click with food for animals

A

deteriorate

428
Q

how to use the clicker two steps

A
  1. Pair the clicker sound with the primary reinforcer , e.g. food (with a short of a delay between the two as possible).
  2. When that pairing is well established, then:
429
Q

when there is both a clicker and food…

A

the clicker announces the food

430
Q

what happens when you use the clicker in isolation

A

In this case, the clicker announces the possibility of food.
“… as long as the secondary reinforcer is occasionally followed by food, the behaviour is maintained”
Or the clicker is used as a marker (yes, this was the right answer, but continue what you are doing; e.g., in “chaining” — more on this below).
Could be called “keep going signal”
Becomes useful in a number of situations — solves issue of distance, need for spatial/temporal precision,

431
Q

why use the clicker

A
  1. Necessary because of delay in delivering the primary, usually because of distance between the trainer and the trainee.
  2. Spatial and/or temporal precision is necessary.
  3. You are training behavioural chains, and use the clicker as a “keep going” signal (similar arguments as per #1):
432
Q

chaining

A

You use chaining to train complex sequences

433
Q

chaining is common in

A

sports, music, dance, and other high motor skill learning — eg. Flying an airplane

434
Q

backward chaining

A

Backward chaining: 3, then 2, then 1 (reverse order of actions).
Learn the whole sequence by starting at the end and go backwards. Could be useful in correction of superstitious behaviour in a dog for example.

435
Q

chaining is often use to…

A

train new pilots and flight simulators

436
Q

“Stimuli associated with primary reinforcement can strengthen behaviour because these stimuli acquire….

A

their own reinforcing value”

437
Q

conditioned (secondary) reinforcers and motivation

A

If not that, then at least it is feedback that informs the organism it will soon get the primary (sometimes called a “ keep going signal” in training circles).

438
Q

still the idea that ____ is what gets ____ to happen

A

Idea that learning/dopamine is what gets attention/learning to happen

439
Q

negative contrast

A

Rats (and humans) are sensitive to the contrast between reinforcers of different values.

440
Q

negative contrast three rat groups

A

Group 1: plain water as reinforcement — fine, especially if animal had been water deprived. But water is not a terribly good reinforcer. Food when hungry works better than water when thirsty.
Group 2: Sweetened water as reinforcement — glucose gives it more value
Group 3: Sweetened water then plain water — they will get less responses to plain water.

441
Q

is reinforcement or punishment better

A

Science suggests that reinforcement is the way to go, but that punishment works (albeit with negative consequences) if just talking about performance, punishment does work.
Question is what supports reinforcement — it is ethical. Important to make an ethical argument, what ethics tell us is that you well develop a better relationship with the human/animal if you use reinforcement as opposed to punishment

442
Q

punishment works to suppress a behaviour, but at a cost…

A

Emotional distress, frustration, toxic relationship, etc.
And not even necessary… other things work.

443
Q

Concurrent reinforcement to punishment may still keep the behaviour:

A

Sam is misbehaving at school and gets punished by the teachers… but the attention he gets from his peers is more reinforcing.

444
Q

When the contingency are figured-out, it may lead to cheating:

A

till producing the “bad” behaviour when the punisher is not around.

445
Q

best technique of reinforcement/punishment

A

DRA or differential reinforcement of alternative (other) behaviours

446
Q

From a ethical perspective people say to always go with

A

positive reinforcement — that is better because it won’t damage relationship with the person. 98% of the time this is the ethical and scientific perceptive

447
Q

is timing important for…

A

conditioned reinforcers

448
Q

timing is particularly important for

A

early training — kibble has to come immediately after the behaviour you want is produced

449
Q

continuous reinforcement

A

every time the animal produces the behaviour it is reinforced.

450
Q

if you want a behaviour to be easily extinguishable reinforce it _____

A

continuously.
- Behaviour that is reinforced every time, when you stop reinforcing it will extinguish immediately. Typical example in parenting — you give a reward to your child for doing something, one day you dont and they won’t do the behaviour anymore.

451
Q

continuous reinforcement sets up…

A

an expectation

452
Q

When you initial train an animal (or human for that matter), you typically start with

A

a continuous reinforcement (CRF) schedule.

453
Q

types of intermittent reinforcement

A

Fixed ratio
Variable ratio
Fixed interval
Variable interval

454
Q

how are intermittent reinforcement/schedule of reinforcement recordered

A

Those schedules are recorded by a cumulative recorder and the output is a cumulative record. It shows the pattern of response based on the reinforcement delivery (the schedule).

455
Q

ratio schedules

A

Ratio is based on a count of the number of occurrences or trials — each trial is every chance you would have to reinforce the organism. In a ratio schedule, you do not reinforce every trials — instead, skip a few then reinforcer
- Ratio schedules are based on the rate of response, i.e., a x number of responses (e.g., you reinforce the rat every 5 good responses, this would be a FR 5).

456
Q

interval schedules

A

Interval is about the timing that elapses before the reinforcement, not the number of trials.
- Interval schedules are based on an interval of time between reinforcements, i.e., regardless of the number of correct bar presses (assuming some good responses occurred), the rat is reinforced at an interval of time (e.g., every 2 minutes, this would be a FI 2).

457
Q

what kind of reinforcement were dog trainers taught to use

A

Dog trainers were taught to do continuous reinforcement, because ‘otherwise the dog would bark or get frustrated’ — gadbois says if that is happening you are doing it wrong

458
Q

what is important for intermittent reinforcement

A

Intermittent reinforcement for every experiment people were taught to go slowly. Eg. If you see any stress or aggravation in your rat you are going too quickly.
Observing your organism is important — figure out what they like or dont. If you sense any frustration stop because something is wrong.

459
Q

95% of the time when you start applying intermittent reinforcement and you get frustration it is because…

A

you are going too fast

460
Q

study using intermittent reinforcment with dogs

A

Study said dogs in intermittent reinforcement got more stressed — they went from 100% continuous reinforcement immediately to 60%. Imagine your dog gets constantly reinforced and then suddenly only 60% of the time and they wonder why the dog is stressed. Thats because you were going way to fast. Actual way to do it is very slowly omit a few of the rewards, with time increasingly omit more. It is that simple. If you start to see any aggression, reinforce a little more.
You are setting up the animal to understand that when they give a right answer — give them feedback they did, but not always a reward. Eventually you will have pigeons that will give 3000 responses without reinforcement — that is how powerful intermittent reinforcement is when used right.

461
Q

what is needed for the transition from continuous to intermittent reinforcement

A

need an adaptation period

462
Q

fixed ratio

A
  • FR 5 = every five answers you give a reward
  • If you are continuously reinforcement start with FR2 for example, needs to be low early on.
  • FR 5 is extremely predicable — they know when the reinforcement is coming. They will tap the lever five times quickly to get the reward, they will accelerate.
  • They have an expectation and that is the issue with the fixed ratio — they learn they just have to produce five responses
463
Q

variable ratio

A
  • Variable ratio = VR 5 +/- 2
  • You set the computers to come up with these numbers, what it means is that they know they will get a response but not every time.
464
Q

fixed interval

A
  • FI 10
  • Unit of time
  • Regardless of number of responses they give, they will not be reinforced until the interval is over — eg. 5 minutes
  • Now they know if they wait a certain time they will get a reinforcement
  • Problem is that they figure out what they do doesn’t really matter until the end of the internal
  • Can correct this problem using variable interval
465
Q

variable interval

A
  • VI 10+/- 3 minutes
  • Dont know exactly the interval of time where they will start getting a reward — therefore will work really hard in those conditions
  • Will typically get you a more consistent reinforcement, lower chance of extinction
466
Q

what methods of reinforcement is preferred

A

Anticipation and learning connected, anticipation will be better when you dont know when you are getting the reward therefore fixed and variable intervals are better.

467
Q

psyches of dog trainers now are wrong

A

Most dogs don’t care about the treat because they are so into the task (working breeds). Whole point of doing intermittent reinforcement is that you train your brain that this is fun because you dont know when you are getting the reward — its a surprise! Very resistant to extinction — falls into category of habit learning
When you keep habit learning it becomes pretty much unextinguishable. Problematic when you do therapy on behaviours you are trying to get rid of that have become habits.

468
Q

problems with fixed ratio or interval

A

they are predictable!

469
Q

what is the benefit of variable ratio or variable interval?

A

By applying those, the rat or pigeon is unsure about when the reward is delivered and is more likely to respond eagerly and the behaviour acquired is much more resistant to extinction

470
Q

real life is…

A

intermittent

471
Q

what relationships are hard to snap out of

A

Relationships that are hard to snap out of are relationships that are unpredictable, they’re exciting — there is a lot of anticipation
That is because you are on a variable ratio — most social interactions are on intermittent reinforcement
Reinforcement matters in social behaviour, that means that all principles you know from learning theory applies to these situations.

472
Q

fixed ratio (example and consequence)

A

Reward every xth response. Consequence: Produces a good amount of behaviours, but lots of post reinforcements pauses. You know that after you are done that ratio, there is no need to rush.

473
Q

variable ratio (example and consequence)

A

The ratio is harder to predict. Consequence: High and persistent rates of responding.
- Slot machines!

474
Q

fixed interval (example and consequence)

A

Reward at a fixed (xth) interval of time. Consequence: Low rate response after each reward. Same problem as fixed ratio because it is relatively easy to predict.

475
Q

variable interval (example and consequence)

A

Steady rates of behaviour because the outcome is difficult to predict in time.
Social media likes or messages

476
Q

variable interval: social media

A

social media companies have people that are trained in learning theory to control this, they control when they let you know, how they let you know, when you stop responding and they look at your rate of responses. They use these intermittent schedules of reinforcement, mostly variable ratio/interval
Can be used for good and bad things!

477
Q

Variable ratio/interval gives…

A

more persistent behaviour less chance of extinction

478
Q

Ratio schedule

A

Strong correlation between the rate of reinforcement and the rate of behaviour.
Faster response rates (the “hurried-up” nature of the behaviours produced, e.g., lever-pressing).

479
Q

Interval schedule

A

Reinforcement rates can be steady, but the rate of responding can vary immensely.
More pauses and wait times between responses.

480
Q

compound schedules

A

Use of more than on schedule. In fact, outside of the laboratory, these are by far the most common schedules of reinforcement.

481
Q

different schedule examples

A

Mixed/multiple/tandem/chained

482
Q

matching law

A

When training an organism you want to see what they respond to best — test what works best. There is some evidence that some species will not deal well with variable intervals (as much as we said they are good if you go slow enough), some species will not work unless it is continuous reinforcement
- Skinner thought there were universal laws here but that is not the case

483
Q

matching law example

A

Give two different schedules to a pigeon
A: pecking at disk A is reinforced on a VI 2 min schedule
B: pecking at disk B is reinforced on a VI 1 min schedule
- In this experiment you get more for VI 1 min, this could change with time and will depend on how hungry the animal is. Eventually, they may switch to VI 2.

484
Q

relative rate of responding to a particular choice (response alternative) will match (equal) the…

A

relative rate of reinforcement for that choice (response alternative).

485
Q

what does matching law explain the relationship between

A

explains the relationship between payoff and choice
Idea is that animals will try to maximize reinforcement at a lower cost

486
Q

what explains the matching law

A
  1. Animals are likely to try to maximize the rate of reinforcement: How much you get per unit of time / session.
  2. They try to improve the local rate of reinforcement, so they shift between the options (choices). This is called melioration
  3. Rate is not the same as strength, so both are factors (the same way as speed and mass influence momentum in kinematics): This brings us to momentum
487
Q

melioration

A

this depends on emotional state — will often simply switch to a better schedule when it fits their current metabolic needs better

488
Q

what is reinforcement a test of

A

persistence rather than competence

489
Q

delayed reinforcement

A

ometimes reinforcement is delayed. How well animals take delayed reinforcement matters (if too long, organisms do not take it well). If humans are being explained to — going to work two weeks for a lot but then get paid, humans are okay with that.

490
Q

Impulsiveness vs self-control (impulse control)

A

impulsive people are more likely to want their goods right away

491
Q

procrastination and working priorities

A

a lot of this has to do with your ability for delayed gratification. Idea is that you have ease into it and develop a technique of finding small gratifications.

492
Q

implications of reinforcement

A
  1. delayed reinforcement
  2. impulsiveness vs. self control
  3. procrastination and working priorties
493
Q

Example of tasks testable with people and animals: Choice between a large delayed reward vs. a smaller immediate reward.

A

The general observation: Animals and humans are impulsive, i.e., prefer smaller immediate rewards.

494
Q

example of delayed reinforcers

A

biweekly paycheque

495
Q

delayed reinforcers are…

A

not as reinforcing as immediate ones.

496
Q

If you are trying to intervene with things that have regular schedules it can be challenging — really hard to get people to adhere to things unless you have …

A

a pre-commitment strategy

497
Q

what do pre commitment strategies have applications or

A

studying, weight loss, cessation of smoking, snoozing the alarm in the morning, etc.

498
Q

what are pre commitment strategies

A

get people to announce what they are going to do (eg. New years resolution). You are dangling the possibility of shame if you dont go through with what you said you would
Example: Dry January and other resolutions that are “publicly” announced.

499
Q

what are factors in delayed gratification

A

temperament / personality, development and psychopahtologies

500
Q

personality example

A

impulsivity

501
Q

development example

A

Impulse control (especially in boys), etc.

502
Q

psychopathologies example

A

DHD, etc.
ADHD is complicated: idea that attention may be a problem. If you cannot focus attention, learning is not going to happen.

503
Q

strategies for delayed gratification

A
  1. Make the immediate reward less “rewarding” (or even coming at a cost).
  2. Make abstention profitable: You get a reward if you don’t procrastinate.
  3. Make the associated delayed reward explicitly less appealing if the choice is to get rewarded immediately
504
Q

Make the immediate reward less “rewarding” (or even coming at a cost).

A

Put a cost on engaging in the behaviour now. They still get a choice in the matter.

505
Q

Make abstention profitable

A

Ironically, that profit can be delayed, it will still work, e.g., “If you finish your homework now, you can play video games for longer tomorrow”.
Appetitive reinforcement?

506
Q

Make the associated delayed reward explicitly less appealing if the choice is to get rewarded immediately

A

Personality difference here, impact may depend on whether you can be patient.

507
Q

what can help delayed grat

A

Much of this can work if you make “anticipation” exciting (as it should be). It tends to be more rewarding to some than others. “Waiting” does not have to be aversive.
Humans have working memory and prospective memory (planning) — ability to deal with things that haven’t happened yet.

508
Q

making anticipation exciting depends on

A
  • Working memory: Good working memory = reduced impulsiveness.
    Distraction from the appeal of the immediate reward is the strategy here. Remember that working memory is conceptually overlapping with attention.
  • Prospective memory and planning: To process delayed rewards, you need to be able to process the future (i.e., mental “time travel”)
509
Q

what are Impulsiveness / sensation-seeking / risk-taking implicated in

A

asal ganglia/dopaminergic system

510
Q

Personality traits that may modulate and influence behaviour in both humans and animals.
Many of these traits are associated with pathologies or risky behaviours:

A

Gambling
Addiction
Risk junkies (sky-diving, bungee jumping, risky sexual behaviours, etc)
ADHD
Borderline personality disorder — characterized by risky sexual behaviours

511
Q

dopaminergic system dogs 1992 study

A

Dogs in 1992: paper that showed a massive difference between breeds that have high vs. Low dopamine.
High = border collie, jack Russel, Australian Shepard (more energetic)
Low = great dane, saint bernard, Newfoundlander (more lazy)
Difference is activity level and impulsiveness
Can be applied to friends: you may have friends that prefer skydiving vs. Friends that prefer laying on couch

512
Q

We now know through neuroimaging that there is a correlation between…

A

raits and dopaminergic system — idea of addictive personalities. We know this is a thing, but we don’t know what to do about it.
We think the key to additive personalities is sensitivity to reinforcement — responding to variable intervals, excitement, etc.
Idea that it is dopamine quantity and sensitivity both — unknown

513
Q

day-to-day applications for managing delayed grat

A
  1. Give choices — with any organism, test your procedure, reinforcement and schedule — every individual may be different
  2. Help make better choices
  3. Be clear about the impact of the choices — explain consequences
514
Q

clinical approaches for managing delayed grat

A
  1. Contingency management — directly from learning theory
  2. Incentive-based treatment
515
Q

Behavioural economics and neuroeconomics (not huge on exam)

A
  • These two fields have massively influenced research in behaviour and neuroscience, and have been influenced by those sciences as well. Go a little beyond basic cognitive psych and neuro — weird interaction between rational and irrational behaviour
  • We know that humans love to think we are cognitive and rational, good at following rules. When actually, we are unbelievably limbic in the way we make our decisions. What we do is limbic not cortical, based on emotions/raw motivations/reinforcement — applies to psych, bio, etc.
516
Q

theories of reinforcement

A
  1. Drive reduction theory
  2. The Premack principle
  3. Behavioural regulation theory
  4. Selection by consequences
517
Q

Interestingly, Skinnerian (radical behaviourism) theories are ______ in explaining reinforcement.

A

not very helpful

518
Q

skinnerian reinforcment

A
  1. They avoid mentalistic concepts, hypothetical/theoretical constructs, e.g., “motivation”. — had to adopt one to explain how motivation works (drive theory)The Skinnerian arguments to explain reinforcement are circular arguments (tautologies).
  2. They do not have a theoretical framework to explain reinforcement, nor do they seem to mind.
519
Q

Clark Hull and drive reduction theory

A

theory of attrition/deficit, missing something so you get it — reducing drive

520
Q

Drive reduction theory:

A

Reinforcers reduce drives.

521
Q

drive reduction theory example

A

Example: Treats reduce hunger (while training a dog; logically, this means that a hungry dog would learn better, faster, etc.).

522
Q

drive reduction theory is a ____ theory

A

It is a deficit theory (you do or learn something because you need to reduce a drive, e.g., hunger, thirst, etc.).
Hull eventually realized the drive theory had limitations and started to talk about “incentives”.
In the section on motivation, we will cover the incentive theory that does a much better job at explaining motivation and its role in learning (and behaviour in general).
Neuro and psych have adopted incentive theory — works better than drive because you can classically condition motivation

523
Q

premack prinicple aka

A

the differential probability principle

524
Q

classical/traditional principle

A

Contingency between a behaviour and the reinforcement.

525
Q

premack principle

A

The contingency is between two behaviours — contrast between pressing a lever and getting fed, to him this is the essence of reinforcement. More probable response will reinforce less probable responses

526
Q

premack prinicple in clear terms

A

more probable responses will reinforce less probable responses
The question: What is the contingent behaviour? In the Skinner box, what are the two behaviours?

527
Q

skinner box: premack

A

Skinner box:
Bar pressing
Eating
- But the rat would rather eat… In other words, reinforcement happens when the (instrumental) behaviour gives access to a preferred behaviour.
- The bar pressing gets you to eat which is the motivation
- Theory about contrasting two different processes

528
Q

solution for reinforcement!

A

Do a preference test! See what is more reinforcing to the subject: Food, toy, play, etc.
Find the behaviour (instrumental act) that provides access to a more preferred behaviour.

529
Q

premack 1959

A

play a pinball machine or have candy
Some kids chose candy (they spent more time eating candy than play pinball).
With these kids, if candy is made contingent on playing pinball, pinball playing increases.
The reverse is true for kids that chose pinball over candy.

530
Q

preference is…

A

really important!

531
Q

preference lab example

A

Dog owners bringing dog preferred treats to lab

532
Q

Behavioural regulation theory (Timberlake et al): premise

A

al behaviours have a preferred level. If you restrict access to that behaviour, the organism will engage in an other behaviour or behaviours to gain back access to that behaviour.

533
Q

Behavioural regulation theory: staddon

A

Bliss point and the minimal distance model. How the organism distributes its behaviours to maximize reinforcement (the bliss point).
So all organisms will learn to maintain a preferred distribution of behaviours.

534
Q

behavioural regulation theory main assumption

A

The main assumption of this model is that humans and animals will seek and attempt to maintain an optimal combination of activities (the bliss point).
Idea is that as organisms we try to meet our metabolic needs

535
Q

Selection by consequences (Donahoe et al.): Premise

A

Learning is like natural selection.

536
Q

selection by consequences, premise is like…

A

“weeding-out” (Bouton’s term) or “pruning” (my term) by reinforcement (or punishment…).

537
Q

“ingredients” for natural selection

A
  1. You need a trait to select. Here it is a behaviour.
  2. You need a selective agent. Here it is reinforcement (via the teacher/trainer/computer).
538
Q

selection by consequences mechanisms are similar to…

A

imilar to natural selection — reinforces the theory that learning is not just about acquiring information, but also pruning information

539
Q

selection by consequences prerequisites

A
  1. variation
  2. fitness consequences
  3. mode of inheritance
  4. limited resources
540
Q

variation

A

This is now variability in forms of behaviours, across situations, etc.

541
Q

fitness consequences

A

This is now simply reinforcement contingencies

542
Q

mode of inheritance

A

The learning process itself (operant conditioning).

543
Q

limited resources

A

The constraints on the process, and deprivation (e.g., of food) is the analogy.
- eg. If you are reinforcing with food, maybe you need hunger. Some dogs are suggested to skip breakfast, so that they will work harder to increase the drive

544
Q

learning is…

A

ontogenetic (it happens during a lifetime, i.e., it is a developmental process)

545
Q

learning happens to…

A

an individual

546
Q

natural selection is…

A

phylogenetic (happens over long periods of time and across generations, i.e., it is an evolutionary process).

547
Q

natural selection happens to…

A

species

548
Q

selection by consequences elements found in connectionist theories

A
  1. Neural networks models make similar assumptions (pruning process are described).
  2. Genetic algorithms and genetic programming: They use explicitly biological terms such as “selection”, “mutation”, “crossover”, etc
549
Q

motivation is a…

A

hypothetical construct

550
Q

potential issues with Structuralism, Functionalism, Gestaltism

A

Potential issues: Introspection, lack of experimental rigour.

551
Q

radical behaviourism (skinnerian)

A

Potential issues: No hypothetical constructs or intervening variables tolerated (i.e., mediators, e.g., memory, emotions, motivation, the brain).

552
Q

radical behaviourism (skinnerian)

A

Potential issues: No hypothetical constructs or intervening variables tolerated (i.e., mediators, e.g., memory, emotions, motivation, the brain).

553
Q

Operational behaviourism (modern academic behaviourism).

A

Potential issues: Still focussed on conditioning to explain all behaviours; and focus on behaviour (as opposed to its underlying processes).

554
Q

Cognitivism: Focus on thought processes, awareness, consciousness, etc.

A

Potential issues: Poverty of the stimulus and representationalism.

555
Q

Connectionism

A
  • PDP models, a.k.a. neural network models
  • A form of neural associationism
  • Explains well perception, learning, memory, but not so well the higher cognitive processes.
556
Q

What is motivation? (4)

A
  1. Hypothetical (theoretical) construct
  2. Need for intervening variables » Hull » Operational behaviourism (Tolman)
  3. So important that Woodworth thought psychology should be called “Motivology”.
  4. But why did some behaviourists (mostly Hull and Tolman) see the need for motivational concepts?
    – The main reason: Explain how reinforcement (and punishment) work.
557
Q

Where is the evidence for motivation?

A

behavioural
neurophysiological

558
Q

behavioural evidence for motivation

A
  1. Behaviour is variable — some days you are horny, other days not — this changes how you feel
  2. Behaviour is persistent — why do we procrastinate? Can persist in good and and behaviours regardless of the outcome
559
Q

neurophysiological evidence for motivation

A
  1. The impact of physiological states relating to homeostasis (balance, a compensating system), allostasis (systems do not balance easily because there are many things going on in parallel — chaotic and paradoxical), metabolism
  2. Neurotransmitters: e.g., dopamine (one of the most important NT for motivation, that is why it has a direct impact on learning)
  3. Hormones: Mostly peptide and steroid hormones
560
Q

How was motivation historically defined?

A
  1. ethology/biology
  2. psychology
561
Q

Ethology/biology historical definition

A

drive as instincts — the hunting drive is not really a desire, it is more an incentive behaviour

562
Q

psychology historial definition

A
  1. Drive theory: a response to deficits (it is a “deficit theory”, focus on -R) — deficit theory, works well with homeostasis, weak theory in some ways
  2. Incentive theory: focus on +R, intrinsic motivation, etc — much more based on positive reinforcement and classical conditioning
563
Q

incentive theory motivation

A
  1. Incentive motivation: Motivation for instrumental behaviour created by anticipation of a positive reinforcer. Also called the rG-sG mechanism (from Hull & Spence)
    - incentive motivation is about how you acquire a reinforcer
  2. Incentive learning: A process by which organisms learn about the value of a specific reinforcer while they are in a particular motivational state
    - Learning of how to be motivated, you can teach an animal to be motivated — requires not always making things too predictable or too easy
564
Q

do you need reinforcement for learning?

A

Technically no. incidental learning (latent learning), implicit learning dont require reinforcement

565
Q

first demonstration of incidental learning

A

Tolman and latent learning

566
Q

are reinforcers necessary for learning

A

Reinforcers are not necessary for learning but are there to motivate behaviour and give purpose.
Reinforcers affect motivation, not learning.
Learning is not performance…
but motivation is performance.

567
Q

will learning always extinguish if not reinforced

A

Nope
Eventually extinction will happen in classical conditioning and operant learning
Habit learning: operant w a lot of repetition, everything gets autonomized and the acquisition of habits will not get extinguished

568
Q

Habit learning…

A
  • Habit learning was hard to situate in learning theory
  • Hull and Lashley (e.g., “Maze running habit”) used the term.
  • Or Tolman’s “habit slip”.
  • The modern use started with Hirsh (1974), then Miskin et al. (1984).
569
Q

concept of habit learning is very close to…

A

“skill learning” (both with a motor component) and both are seen overall as “implicit” or “non declarative” forms of learning

570
Q

habit learning and skill learning both seem to be controlled by

A

same parts of brain — basal ganglia, dopamine, subcortical
Idea you can go beyond conditioning and get really strong learning that is not much at the mercy of external reinforcement because it has become intrinsically motivated
Eg. Dogs in field don’t always need a treat, sometimes just motivated enough by finding a turtle
The more you do something, the more likely you are to develop intrinsic motivation
Difference between liking and wanting something

571
Q

types of declarative (Explicit) learning

A

facts and events

572
Q

types of nondeclarative (implict) learning

A

skills and habits, priming, simple classical conditoning, nonassociative learning

573
Q

Habit learning is:

A

Instrumental in nature, and very much associated with “motor learning” and the basal ganglia (as opposed to hippocampal-based learning).
- More subcortical than we would imagine

574
Q

Hippocampal (contextualizer) learning

A

Context learning (spatial and temporal). Defined as “rapid learning”.
We also know from Gray that the hippocampus is a “comparator”.
The hippocampus gives context (the what, the when and the where).
Change in context can make you lose some learning or ability to adjust in a certain environment

575
Q

Basal ganglia-based learning

A

Motor in nature. Defined as “slow learning”.
Now often defined within the realm of the “cortico-striatal loop” (two loops, one is motor, the other is motivational)
If a rat learns to run the maze, but can do it by swimming it is not just basal ganglia earning (memorizing specific motor sequences), but also hippocampal. They both work.

576
Q

The cortico-striatal loop (from Seger & Spiering, 2011)

A
  • Not just cortical or subcortical, but both
    -So motor-based learning (habit and skill learning) and motivation are linked in the brain.
    -Both are heavily driven by the dopaminergic system.
    -Remember: Both “motor” and “motivation” (even “emotion”) have the same etymological root.
    –> Why you give amphetamines to people with ADHA (ritalin)
577
Q

All parts of the brain are kinda connected…

A

if you look at the etymology but motor and motivation go together

578
Q

the evolution of the concept of motivation: Hull

A

drive theory

579
Q

drive theory

A
  • Hull, early work: Reinforcement » reduction of drive (e.g., hunger)
  • Behaviour strength = D x H
  • Behaviour strength = Drive (need) x Habit (learning)
  • Interaction between drive and habit
  • Behaviour strength = Need x Learning
580
Q

Resistance to extinction depends on:

A
  1. Degree of food deprivation (how hungry you are): Drive or D
  2. How much prior reinforcement (reinforcement history) you got: Habit or H
581
Q

Incentive theory: Hull take 2

A
  • The concept of incentive motivation: Hull & Spence
  • Behaviour strength = D x H x K
    Behaviour strength = Drive x Habit x Incentive
    Behaviour strength = Need x Learning x Incentive
    Realized the K explains better forms of learning — innate
582
Q

Incentive =

A

motivational effect of reward

583
Q

Hull’s theory (even the update with Spence) is seen as

A

outdated today

584
Q

Now we conceptualize motivation as

A

response to an anticipation of need

585
Q

Tolman’s view

A

Tolman: He argued that animals learn expectancies
Food rewards are confirmatory (or not).
Food = catalyst (Rescorla, 1985).

586
Q

for a strict, radical behaviourist

A
  • the operational terms “anticipation” and “expectancy” are problematic… but only theoretically.
  • So they prefer still the term “reinforcement”
  • Clearly the reward value in a learning task is contingent on how easy it is to get it
  • The way the see the reinforcer is as a confirmer (yea, I was right — like tap teaching in dance)
587
Q

Here we lose the radical behaviourists because

A

(anticipation is involved)
The physiological processes are also not a focus of radical behaviourism.
The processes are invisible to the naked eye… unless you have technology… and we do in neuroscience.
Although the process is classically conditioned…
Indeed, the motivational effects of reward come from classical conditioning (this is “incentive learning”).
This is what incentive learning really is

588
Q

The motivational effect of reward comes from

A

classical conditioning

589
Q

reinforcement affecting motivational state idea

A

Idea is simple: if you are hungry, you are learning about that state of being hungry. It is actually when you are hungry that you are open to the incentive value of the situation you are in. Only when you are hungry can you learn, because it sets you up for that anticipation. Anticipation itself becomes the reward (remember the clicker!)
Clicker announces food > clicker announces possibility of food > pigeon that is still pecking at same stimulus despite the fact there is no food
This occurs because there is still anticipation!
In a sense, motivational states are occasion setters

590
Q

incentive learning

A

-You learn in the hungry state about the incentive value
-So you learn about an incentive, and then:
-You link this to the instrumental action.
-If you experience the +R in a non-hungry state, then you will have no interest in the +R (e.g., food).
-Motivational states are occasion setters! (Davidson, 1993, 1998, etc.) or facilitators (Rescorla, 1985)
-Great review of these ideas in Dickinson & Balleine, 1994

591
Q

Interlude; what are occasion setters?

A

Safety signals are negative occasion setters:
“if this light is on, the shock won’t happen”
“if I have my pills, I won’t have a panic attack”
and even: “if I have my pill bottle – even if empty – I won’t have a panic attack”.

592
Q

We can have modulation in conditioning

A

Facilitation and facilitators; occasion setting and occasion setter

593
Q

occasion setters are…

A

environmental cues (CS’s) that “set the occasion” for conditioning to occur. They “facilitate” the conditioning.

594
Q

Occasion setting (Holland) =

A

facilitation (Rescorla)

595
Q

Occasion setters (Holland) =

A

facilitators (Rescorla) » motivational states

596
Q

An o.s. is a

A

CS that confirms a CS-US pairing.

597
Q

An o.s. is a

A

modulator of the conditioning between a CS and a US.
In itself, it becomes rewarding

598
Q

So we have three ingredients in classical conditioning modulation:

A
  1. The CS
  2. The US
  3. The modulator (second CS) = occasion setter = facilitator » motivational states
599
Q

what does the modulator add

A

This adds what we call “conditional relations” and “conditional control”.
It is not unlike basic discriminations in operant conditioning when we reinforce the S+ (the target stimulus or “good response”), and not the S- (the foil, distractor, or “wrong response”).

600
Q

pavlovian discrimination

A

Two ingredients:
1. The target stimulus: The CS that is present on every trial in a CS-US pairing. It is the target because that is the one that the subject responds to.
2. The feature stimulus: The CS that is present only in trials indicating if the US occurs or not (depending on the procedure).
On top of operant discrimination, you can add something from a specific situation using classical conditioning

601
Q

occasion setter facilitates…

A

a specific association between stimuli — facilitator, can be there or not but its presence facilities something

602
Q

So back to incentive motivation and learning:

A

-Hunger triggers the foraging behaviour (system).
-Motivation “energizes” the action.
-But you must learn (“know”) that the action leads to the reinforcer.
-Occasion-setters (or facilitators) help in doing so!
-You must learn that the reinforcer has a positive effect on the motivational state.
-The motivational state increases the desirability of the reinforcer

603
Q

“Learning will always extinguish if not reinforced”

A

Actually extended training makes behaviour less sensitive to its consequences » habit learning (e.g., Holland 2004).

604
Q

It is not about the need, or the reward…

A

but about the anticipation of the reward…

605
Q

acquired motivation

A

Motivation is not from drives/needs, but rather the anticipation of need…
Back to Tolman and latent learning…
1.Exploration of the maze. No reward.
2. Reward in the maze: They excel, with no previous reinforcement history!
So… rewards:
1.Reinforce behaviour
2.Motivate

606
Q

early behaviourists believed in

A

gradual learning, not one trial learnings. But if a rat can do it just by exploring and no reinforcement, there is something else going on. Rewards reinforce and motivate behaviour

607
Q

tolmans ultimate contribution

A

As Bouton (2016) says elegantly, “a reward at the end of the tunnel” goes a long way
Sometimes just the possibility of it is enough
Flaherty (1996) discuses at length these phenomena in “Incentive relativity”.

608
Q

crespi (1942)

A

Negative and positive. contrast effects

609
Q

crespi study

A

An increase in reward (1 to 16 pellets) results in an increase in running speed and elation = positive contrast effect.
– Acquire the task, at shift time there is no change. The learning is stabilized.
A decrease in reward (256 to 16 pellets) results in a decrease in running speed and depression = negative contrast effect.
– Huge crash in running speed, they adjust their behaviour based on what they are reinforced in
1 pellet to 16 pellets
– Huge jump in their learning! Post-shift is huge

610
Q

crespi study rationale

A

Those effects are caused by a change in the expected value of the reward. Not really the reward, but what you expect the reward to be. How much effort you are putting in might be modulated by the expected value of the reward.
Not doubt this is paradoxical

611
Q

A few words about paradoxical reward effects:

A

Rewards can sometimes weaken behaviours (in an instrumental context)
- -Eg. You get tired
Non-rewards can sometimes strengthen behaviours (in an instrumental context)
– Eg. Allude to incentive/intrinsic motivation
This includes the negative contrast effects mentioned above.

612
Q

Rewards may not always be as magical as people think, they often and quickly become irrelevant

A
  • Animal trainers seem to forget this — the importance to develop incentive motivation and that you do not need to always reward.
  • Gen-X thought always rewarding was a great idea — everyone gets a trophy. Sounds good but it creates more problems than anything else, namely increasing a dependency on extrinsic motivation.
613
Q

What about extinction and motivation?

A

Extinction has a lot to do with the magnitude of the reinforcement.
The bigger the reward, the bigger the extinction
The higher magnitude for a reward, the faster the extinction. (Hulse, 1958; Wagner, 1961)… UNLESS, you are dealing with HABIT LEARNING

614
Q

over learning extinction effect

A

Many rewarded trials can increase the rate of extinction (as opposed to fewer rewarded trials)

615
Q

what should you do with high value reinforcement

A

Maybe make it intermittent

616
Q

example of intermittent reinforcement

A

Slot machines are very careful in giving very big payments very rarely. The small payments keep you there, the inconsistent big prizes are the anticipation.

617
Q

Can rewards have negative effects?

A

Extrinsic rewards can affect performance negatively.
“Punished-by-reward” principle in humans

618
Q

“Punished-by-reward” principle in humans

A

People like this do not want to take responsibility and are also extremely sensitive to external rewards. When it is not rewarding anymore they disconnect. This is the danger of using extrinsic motivation too much — idea is trying to find ways of developing intrinsic motivation

619
Q

can verbal rewards be good?

A

For some people these are more important than money
Expectations, overall, are important.

620
Q

Partial Reinforcement Extinction Effect (PREE) and PERSISTENCE

A
  • Rescorla, 1999 (and a few other studies) — said connection between US/CS was really important
    He realized he might be wrong after Pearce/Hall and Wagner
    -Partial reinforcement reduces the associative change produced by non- reinforcement.
    Two groups of rats: CRF (continuous reinforcement) and PRF (partial reinforcement)
    In acquisition, CRF (100% of trials are reinforced): better (faster, stronger acquisition)
    -But: In extinction, PRF (e.g., 50% of trials are reinforced) » persistence
621
Q

running speed (Effort for motivation) study

A

Running speed (effort for motivation) — in acquisition trials the continuous reinforcement were doing better than the partial. Seeing a little less effort
In extinction: the continuous lose the behaviour much more quickly, running speed goes down dramatically. The partially enforced rats see effort go down over time, but it is much slower.

622
Q

working dogs hpoglycaemia struggle

A

Train dogs to recognize signs of anxiety in people with PTSD
Biomedical alert and the lack of maintenance training — people with hypoglycaemia in people with diabetes
This events are very raw, so it is hard to reinforce. Even when we have people that volunteer to produce those samples, well over have the time they forgot. When they get hypo they deal with that situation at hand, not about training the dog. The dogs then do not get reinforced. That is problematic unless if during the training they start using partial reinforcement. Real life for these dogs means that dogs will not always get reinforced, but they need to keep detecting it.
Performance collapse, in part because of rare events: Few chances for reinforcement… if the responses to the events are even reinforced when they occur!
Learned industriousness (Einsenberg, 1992)&raquo_space; this can transfer to other tasks.

623
Q

Sequential theory (Capaldi)

A

How to explain PREE? Two theories:
1. Frustration theory (e.g., Amsel): Not well supported
2. Sequential theory: Not about frustration
1. Extinction = many non-rewarded trials (by definition!)
2. No problem! Especially if this matches acquisition done in PRF!
3. It is the memory of the acquisition phase that “sets the tone” for what to expect.
Set expectations early and your future behaviour is dependent on this. This is why always reinforcing is a bad idea. What happens during the acquisition phase is what matters.

624
Q

the sequential theory, and small and large rewards: Capaldi and Capaldi (1970), Leonard (1969)

A

Sequences of:
non-reinforced trials (N)
reinforced trials with large rewards (R)
reinforced trials with small rewards (r).
Two acquisition types:
rNR
RNr
rNR is more resistant to extinction as suggested by the sequential theory
This is because they expect less initially

625
Q

is frustration theory very well supported?

A

no, not very well supported. Frustration does not seem necessary to justify motivation.
In fact, some cognitive factors seem at least as important:
- Memory
- Associative strength

626
Q

How foraging works example

A

Important point. Basically said, why is it that there is so much resistance from some behaviourists and some animal trainers on the idea that reinforcing all the time is not the best. They point out that if you actually look at how animals get reinforced in the wild, that reinforcement every time would collapse very quickly
Looking at success rate in predation: for most is very low
Yet they continue doing it
If they were right that you need to be rewarded every time you try to hunt, most lions after a few trials would quit
Says: hey, our brains are wired for not knowing and instead hoping. That incentive/intrinsic motivation and effect of anticipation and expectancy.