Mod27
Mod27
Skinner’s Experiments
Thorndike used a fish reward to entice cats to find their way out of a puzzle
box through a series of maneuvers. The cats’ performance tended to improve
with successive trials, illustrating Thorndike’s law of effect. (Data
from Thorndike, 1898.)
A Skinner box
Inside the box, the rat presses a bar for a food reward. Outside, measuring
devices (not shown here) record the animal’s accumulated responses.
Shaping Behavior
Imagine that you wanted to condition a hungry rat to press a bar. Like
Skinner, you could tease out this action with shaping, gradually guiding the
rat’s actions toward the desired behavior. First, you would watch how the
animal naturally behaves, so that you could build on its existing behaviors.
You might give the rat a bit of food each time it approaches the bar. Once the
rat is approaching regularly, you would give the food only when it moves
close to the bar, then closer still. Finally, you would require it to touch the bar
to get food. By rewarding successive approximations (as Sutherland did with
her husband), you reinforce responses that are ever closer to the final
desired behavior, and you ignore all other responses. By making rewards
contingent on desired behaviors, researchers and animal trainers gradually
shape complex behaviors.
Reinforcers vary with circumstances What is reinforcing (a heat lamp) to
one animal (a cold meerkat) may not be to another (an overheated child).
What is reinforcing in one situation (a cold snap at the Taronga Zoo in
Sydney, Australia) may not be in another (a sweltering summer day).
Shaping can also help us understand what nonverbal organisms can
perceive. Can a dog distinguish red and green? Can a baby hear the
difference between lower- and higher-pitched tones? If we can shape them to
respond to one stimulus and not to another, then we know they can perceive
the difference. Such experiments have even shown that some nonhuman
animals can form concepts. When experimenters reinforced pigeons for
pecking after seeing a human face, but not after seeing other images, the
pigeon’s behavior showed that it could recognize human faces (Herrnstein &
Loveland, 1964). In this experiment, the human face was a discriminative
stimulus. Like a green traffic light, discriminative stimuli signal that a
response will be reinforced (Figure 27.3). After being trained to discriminate
among classes of events or objects—flowers, people, cars, chairs—pigeons
can usually identify the category in which a new pictured object belongs
(Bhatt et al., 1988; Wasserman, 1993). They have even been trained to
discriminate between the music of Bach and Stravinsky (Porter & Neuringer,
1984).
Figure 27.3
After being rewarded with food when correctly spotting breast tumors,
pigeons became as skilled as humans at discriminating cancerous from
healthy tissue (Levenson et al., 2015). Other animals have been shaped to
sniff out land mines or locate people amid rubble (La Londe et al., 2015).
Skinner noted that we continually reinforce and shape others’ everyday
behaviors, though we may not mean to do so. Isaac’s whining annoys his
dad, for example, but consider how his dad typically responds:
Or consider a teacher who sticks gold stars on a wall chart beside the names
of children scoring 100 percent on spelling tests. As everyone can then see,
some children consistently do perfect work. The others, who may have
worked harder than the academic all-stars, get no rewards. The teacher
would be better advised to apply the principles of operant conditioning—to
reinforce all spellers for gradual improvements (successive approximations
toward perfect spelling of words they find challenging).
Types of Reinforcers
Reinforcement Schedules
In most of our examples, the desired response has been reinforced every
time it occurs. But reinforcement schedules vary. With continuous
reinforcement, learning occurs rapidly, which makes it the best choice for
mastering a behavior. But extinction also occurs rapidly. When reinforcement
stops—when we stop delivering food after the rat presses the bar—the
behavior soon stops (is extinguished). If a normally dependable candy
machine fails to deliver a chocolate bar twice in a row, we stop putting
money into it (although a week later we may exhibit spontaneous
recovery by trying again).
Figure 27.4
Fixed Variable
Punishment
“A pat on the back, though only a few vertebrae removed from a kick in the
pants, is miles ahead in results.”
In classrooms, too, teachers can give feedback by saying, “No, but try this
…” and “Yes, that’s it!” Such responses reduce unwanted behavior while
reinforcing more desirable alternatives. Remember: Punishment tells you
what not to do; reinforcement tells you what to do. Thus, punishment trains a
particular sort of morality—one focused on prohibition (what not to do) rather
than positive obligations (Sheikh & Janoff-Bulman, 2013).
What punishment often teaches, said Skinner, is how to avoid it. Most
psychologists now favor an emphasis on reinforcement: Notice people doing
something right and affirm them for it.
Skinner’s Legacy