Unit 4: Operant Conditioning Flashcards

1
Q

What’s the difference between instrumental and classical conditioning?

A

classical: presence or absence of stimulus causes response
operant: behaviour causes presence or absence of stimulus (consequence)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
2
Q

Who did the study of instrumental conditioning start with and what was he interested in?

A

Edward Thorndike
interested in animal intelligence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
3
Q

How were Thorndike’s experiments generally structured?

A

hungry animals placed in puzzle boxes
food outside of boxes but in view
-> animals had to learn how to escape the box to obtain food

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
4
Q

How did the animal’s behaviours change in the puzzle box?

A

initially unable to escape
slow to make right response
continued to practice until latencies become shorter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
5
Q

How did the animals learn how to solve the puzzle box?

A

trial and error to discover behaviour required to escape
successful behaviours retained
useless behaviours eliminated

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
6
Q

How did Thorndike label the animals ability to learn how to escape a puzzle box?

A

animal intelligence

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
7
Q

Why is animal intelligence not an accurate term?

A

many behaviours seem unintelligent
initial presence of various responses typical for confined animal, with some leading to a desirable result
consequences reinforce the action
->cat doesnt understand how levers work, but presses it because it thinks it will be rewarded for it

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
8
Q

Law of Effect

A

If R in presence of S is followed by positive event -> becomes strengthened
If it isn’t followed by a positive event -> S-R association becomes weakened

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
9
Q

What do we measure in discrete-trial procedures?

A

rat runs down a maze to get reward
measures response latency (time it takes for the rat to leave the start of the box) and running speed (how fast it reaches the end)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
10
Q

What’s a T maze trial?

A

type of discrete-trial procedure
allows us to measure percentage of correct choices

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
11
Q

What are trials?

A

specific periods of time during which the animal can show instrumental responses
set by the experimenter

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
12
Q

Why didn’t Skinner use discrete-trial procedures?

A

behaviour is continuous (one leads to the next)
-> trials more natural if animals aren’t removed
behaviour can be broken down into measurable units: operants

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
13
Q

What is magazine training?

A

US paired with CS via classical conditioning
sound elicits sign-tracking response

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
14
Q

How does response shaping work? (example: rat in Skinner box)

A

after magazine training, rat can learn operant response
1. food given if rat goes on hind legs anywhere in chamber
2. food given if rat leans over lever
3. food given if rat goes up on hind legs and presses lever
=> sequence called shaping/ reinforcement of successive approximations

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
15
Q

What are operant responses in free-operant procedures measured as?

A

rates

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
16
Q

Response rate

A

frequency of instrumental behaviours occurring
high: high probability of behaviour occurring
low: low probability of behaviour occurring

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
17
Q

How can we differentiate outcomes?

A

appetitive vs aversive
positive vs negative

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
18
Q

Which components do all instrumental conditioning procedures involve?

A

instrumental response
outcome (reinforcement, punishment)
stimulus
association between response and outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
19
Q

positive reinforcement

A

behaviour produces (adds) appetitive outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
20
Q

negative reinforcement

A

behaviour produces absence of aversive stimulus

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
21
Q

positive punishment

A

behaviour produces aversive outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
22
Q

negative punishment

A

behaviour produces absence of appetitive outcome

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
23
Q

Can a behaviour always be reinforced (by anything)?

A

no, only if behaviour is naturally linked to reinforcement
e.g. cant reinforce yawning in cats with opening box, because yawning isn’t naturally linked with release from confinement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
24
Q

What does the presence of a stimulus activate?

A

behaviour system related to that stimulus
e.g. hunger (S) causes hamsters to start digging and scrabbling (behavioural system linked to hunger), while stopping self-care behaviour (behaviour doesn’t address hunger)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
25
Q

What does instrumental conditioning depend on with regards to the reinforcement?

A

quality and quantity of reinforcement
nature of reinforcement
previous reinforcements for same instrumental behaviour

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
26
Q

Behavioural contrast effect

A

big reward perceived as especially good after small reward and vice versa

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
27
Q

Which types of relationships between response and reinforcement are there?

A

temporal relationship: contiguity
causal relationship: contingency

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
28
Q

Are temporal and causal factors dependent on each other?

A

no

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
29
Q

What can we say about temporal relations?

A

immediate reinforcement is preferable to delayed reinforcement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
30
Q

credit assignment

A

if too much time passes, we won’t be able to link specific behaviours to the reinforcement
(credit assignment is the reason for this)

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
31
Q

What’s more important to create associations, contingency or contiguity?

A

contiguity

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
32
Q

The fact that a behaviour occurred just before the reinforcement was more important than whether it caused the reinforcement. What is this kind of reinforcement called?

A

adventitious/ accidental reinforcement

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
33
Q

learned helplessness

A

when experiencing a tense state repeatedly
-> feeling of being incapable to change the situation

How well did you know this?
1
Not at all
2
3
4
5
Perfectly
34
Q

Why does the reinforcer not occur after every response in instrumental conditioning procedures?

A

reflects nature of real world

35
Q

What’s a schedule of reinforcement?

A

rule that determines how and when a reinforcer follows a response

36
Q

Ratio schedules

A

reinforcer occurs after X amount of responses

37
Q

continuous reinforcement schedules

A

reinforcement delivered after every response
commonly used in drug abuse treatments

38
Q

Is continuous reinforcement common in real life?

A

no

39
Q

Token economies

A

reward system where tokens can be exchanged for bigger rewards
used to reduce disruptive behaviours

40
Q

Partial/ Intermittent reinforcement

A

reinforcement only occurs sometimes

41
Q

fixed-ratio schedules

A

number of reinforcers received per number of responses is fixed

42
Q

Are continuous reinforcement schedules a type of fixed-ratio reinforcements?

A

yes

43
Q

How strong is the rate of responding generated by continuous reinforcement?

A

steady and moderate

44
Q

What is steady responding PRECEDED by? (in fixed-ratio schedules)

A

a brief pause

45
Q

How/ where do we see rates of responding?

A

in cumulative records
-> show total number of responses during a period of time

46
Q

How do cumulative recordings work?

A

pen rests on piece of paper
moves up vertically after each response
-> time between responses (horizontal distance)
-> slope: rate of responding (number of responses per unit of time operants)

47
Q

What does an increase in fixed-ratio requirements tend to cause?

A

increase in post-reinforcement pause despite no change in rate of responding during ratio run

48
Q

What’s ratio strain and what causes it?

A

caused by dramatic increases in fixed-ratio requirements
-> periodic pauses during ratio run
-> in extreme cases responses stop completely

49
Q

Variable-ratio schedules

A

number of responses required to achieve reinforcement varies

50
Q

Do variable-ratio schedules cause post-reinforcement pauses?

A

no, they are less likely to cause pauses

51
Q

Why don’t variable-ratio schedules cause post-reinforcement pauses?

A

subject doesn’t know how many responses are required for reinforcement
-> maintains stable rate of responding

52
Q

Is fixed or variable-ratio reinforcement more effective for long term effects?

A

variable-ratio

53
Q

What’s the difference between variable reinforcement and intermittent reinforcement?

A

Intermittent reinforcement: broad category, includes variable and fixed ratio/ intervall reinforcement, rewards only given sometimes
variable reinforcement: reinforcement given after unpredictable amount of times, can’t be fixed number of responses/ time

54
Q

What can be said about the effects (rate of responding and resistence) of variable and intermittent reinforcement?

A

variable: higher rate of responding, less resistent to extinction
intermittent: lower rate of responding, more resistent to extinction

55
Q

What’s an interval schedule?

A

responses are only reinforced if the response occurs after a certain amount of time

56
Q

Fixed-interval schedules

A

amount of time that has to pass before response is reinforced is constant

57
Q

Do fixed interval schedules always lead to the reinforcement being delivered after a certain amount of time?

A

no, the interval only determines when a reward is available, not when it’s delivered
-> response still has to occur

58
Q

Variable interval schedule

A

amount of time that passes between a reference point and response that produces a reward varies

59
Q

Do fixed interval schedules also lead to post-reinforcement pauses (and a higher response rate before the delivery of the next reinforcement)?

A

yes

60
Q

Do variable-ratio and interval schedules cause the same effects?

A

no, variable-ratio leads to higher response rate while variable-interval leads to more steady response rate

61
Q

Why are the different effects of ratio and interval schedules relevant for us?

A

implications human behavior (motivation)

62
Q

What happens in concurrent schedule trials?

A

different responses can be associated with different reinforcement on different schedules of reinforcement
(your responses in a situation are rewarded at different rates depending on the behavior, so you have to choose which behavior to carry out)

63
Q

How do we measure choice behaviour?

A

assessing how responses are distributed across response alternatives

64
Q

How do you mathematically measure response rates of concurrent schedules?

A

Response A (total left responses) / Total responses (A&B / left & right)

65
Q

How do you mathematically measure the rate of reinforcement (concurrent schedules)?

A

Total reinforcements for option A / Total reinforcements (A&B)

66
Q

What does the matching law describe?

A

rate of responding on alternatives matches rate of reinforcement (40% of all reinforcements come from A -> responds to A 40% of the time)

67
Q

What are the two perspectives, with which you can answer the question as to what motivates instrumental behavior?

A

Associative structure of instrumental conditioning (molecular perspective): how stimuli, responses and outcomes are related
Response-allocation approach (molar perspective): how instrumental behavior relates to long-term goals

68
Q

Does instrumental responding only involve the response and a reinforcement? If no, what’s missing?

A

No
the environment

69
Q

What are the three events of instrumental responding?

A

Contextual Stimuli
instrumental response
response outcome (only weakens/ strengthens response)

70
Q

What are habits and how much of our behavior do they constitute?

A

things we do automatically
45% (estimate)

71
Q

Two-process theory

A

two types of learning: Pavlovian and instrumental
S-O association established through classical conditioning activates reward expectancy/ emotional state
-> depending on nature of emotional state (appetitive/ aversive): motivates instrumental behavior

72
Q

What does the associative analysis of instrumental conditioning view behavior as?

A

result of associations between stimuli, responses and outcomes

73
Q

What is instrumental behavior motivated by? (context: organism’s behavior)

A

restricting an organism’s natural behavior (only 1 response leads to reward, rest doesn’t)

74
Q

Consummatory-response theory

A

it isn’t the nature of the response that is reinforcing, but the behavior related to consuming the reward

75
Q

What does the Premack principle state?

A

high-probability behaviors can be effective reinforcements for low-probability behaviors

76
Q

How is a high probability of reinforcing behavior maintained in instrumental conditioning procedures?

A

restricting access to reinforcement

77
Q

Response deprivation hypothesis

A

restricted access to reinforcement is critical for motivating instrumental responding

78
Q

Instrumental contingency & behavioral bliss point

A

Quality of reward must reflect quality of response & vice versa
(time on facebook must equal time studying)
-> behavioral bliss point: perfect ratio

79
Q

What does behavioral economics study?

A

what alternative reinforcements are available
how are alternatives related to reinforcement in question
costs of obtaining alternatives
-> understand what motivates instrumental responding

80
Q

demand curve

A

how consumption of products is influenced by the price

81
Q

elasticity of demand

A

how much consumption can vary in relation to increasing costs
high elasticity = high variability

82
Q

Does elasticity of demand depend on the availability of other products?

A

yes
more options = higher elasticity

83
Q

What else influences how much an increase in price affects consumption? (not availability of alternatives)

A

income
instrumental contingencies