Contemporary Views of Operant Conditioning

KEY THEME

In contrast to Skinner, today’s psychologists acknowledge the importance of both cognitive and evolutionary factors in operant conditioning.

KEY QUESTIONS

In our discussion of classical conditioning, we noted that contemporary psychologists acknowledge the important roles played by cognitive factors and biological predispositions in classical conditioning. The situation is much the same with operant conditioning. The basic principles of operant conditioning have been confirmed in thousands of studies. However, our understanding of operant conditioning has been broadened by the consideration of cognitive factors and the recognition of the importance of natural behavior patterns.

Cognitive Aspects of Operant Conditioning

RATS! I THOUGHT YOU HAD THE MAP!

In Skinner’s view, operant conditioning did not need to invoke cognitive factors to explain the acquisition of operant behaviors. Words such as expect, prefer, choose, and decide could not be used to explain how behaviors were acquired, maintained, or extinguished. Similarly, Thorndike and other early behaviorists believed that complex, active behaviors were no more than a chain of stimulus–response connections that had been “stamped in” by their effects.

210

image
Edward Chace Tolman (1898–1956) Although he looks rather solemn in this photo, Tolman was known for his openness to new ideas, energetic teaching style, and playful sense of humor. During an important speech, he showed a film of a rat in a maze with a short clip from a Mickey Mouse cartoon spliced in at the end (Gleitman, 1991). Tolman’s research demonstrated that cognitive processes are an important part of learning, even in the rat.
Archives of the History of American Psychology, The University of Akron

However, not all learning researchers agreed with Skinner and Thorndike. Edward C. Tolman firmly believed that cognitive processes played an important role in the learning of complex behaviors—even in the lowly laboratory rat. According to Tolman, although such cognitive processes could not be observed directly, they could still be experimentally verified and inferred by careful observation of outward behavior (Tolman, 1932).

Much of Tolman’s research involved rats in mazes. When Tolman began his research in the 1920s, many studies of rats in mazes had been done. In a typical experiment, a rat would be placed in the “start” box. A food reward would be put in the “goal” box at the end of the maze. The rat would initially make many mistakes in running the maze. After several trials, it would eventually learn to run the maze quickly and with very few errors.

But what had the rats learned? According to traditional behaviorists, the rats had learned a sequence of responses, such as “first corner—turn left; second corner—turn left; third corner—turn right,” and so on. Each response was associated with the “stimulus” of the rat’s position in the maze. And the entire sequence of responses was “stamped in” by the food reward at the end of the maze.

Tolman (1948) disagreed with that view. He noted that several investigators had reported as incidental findings that their maze-running rats had occasionally taken their own shortcuts to the food box. In one case, an enterprising rat had knocked the cover off the maze, climbed over the maze wall and out of the maze, and scampered directly to the food box (Lashley, 1929; Tolman & others, 1946). To Tolman, such reports indicated that the rats had learned more than simply the sequence of responses required to get to the food. Tolman believed instead that the rats eventually built up, through experience, a cognitive map of the maze—a mental representation of its layout.

image
“Well, you don’t look like an experimental psychologist to me.”
Sam Gross The New Yorker Collection/The Cartoon Bank

As an analogy, think of the route you typically take to get to your psychology classroom. If a hallway along the way were blocked off for repairs, you would use your cognitive map of the building to come up with an alternative route to class. Tolman showed experimentally that rats, like people, seem to form cognitive maps (Tolman, 1948). And, like us, rats can use their cognitive maps to come up with an alternative route to a goal when the customary route is blocked (Tolman & Honzik, 1930a).

Tolman challenged the prevailing behaviorist model on another important point. According to Thorndike, for example, learning would not occur unless the behavior was “strengthened,” or “stamped in,” by a rewarding consequence. But Tolman showed that this was not necessarily the case. In a classic experiment, three groups of rats were put in the same maze once a day for several days (Tolman & Honzik, 1930b). For group 1, a food reward awaited the rats at the end of the maze. Their performance in the maze steadily improved; the number of errors and the time it took the rats to reach the goal box showed a steady decline with each trial. The rats in group 2 were placed in the maze each day with no food reward. They consistently made many errors, and their performance showed only slight improvement. The performance of the rats in groups 1 and 2 was exactly what the traditional behaviorist model would have predicted.

Now consider the behavior of the rats in group 3. These rats were placed in the maze with no food reward for the first 10 days of the experiment. Like the rats in group 2, they made many errors as they wandered about the maze. But, beginning on day 11, they received a food reward at the end of the maze. As you can see in Figure 5.6, there was a dramatic improvement in group 3’s performance from day 11 to day 12. Once the rats had discovered that food awaited them at the end of the maze, they made a beeline for the goal. On day 12, the rats in group 3 ran the maze with very few errors, improving their performance to the level of the rats in group 1 that had been rewarded on every trial!

image
FIGURE 5.6 Latent Learning Beginning with day 1, the rats in group 1 received a food reward at the end of the maze, and the number of errors they made steadily decreased each day. The rats in group 2 never received a food reward; they made many errors as they wandered about in the maze. The rats in group 3 did not receive a food reward on days 1 through 10. Beginning on day 11, they received a food reward at the end of the maze. Notice the sharp decrease in errors on day 12 and thereafter. According to Tolman, the rats in group 3 had formed a cognitive MAP of the maze during the first 11 days of the experiment. Learning had taken place, but this learning was not demonstrated until reinforcement was present—a phenomenon that Tolman called latent learning.
Data from Tolman & Honzik (1930b).

211

Tolman concluded that reward—or reinforcement—is not necessary for learning to take place (Tolman & Honzik, 1930b). The rats in group 3 had learned the layout of the maze and formed a cognitive map of the maze simply by exploring it for 10 days. However, they had not been motivated to demonstrate that learning until a reward was introduced. Rewards, then, seem to affect the performance of what has been learned rather than learning itself. To describe learning that is not immediately demonstrated in overt behavior, Tolman used the term latent learning (Soderstrom & Bjork, 2015).

From these and other experiments, Tolman concluded that learning involves the acquisition of knowledge rather than simply changes in outward behavior. According to Tolman (1932), an organism essentially learns “what leads to what.” It learns to “expect” that a certain behavior will lead to a particular outcome in a specific situation.

Tolman is now recognized as an important forerunner of modern cognitive learning theorists (Gleitman, 1991; Olton, 1992). Many contemporary cognitive learning theorists follow Tolman in their belief that operant conditioning involves the cognitive representation of the relationship between a behavior and its consequence. Today, operant conditioning is seen as involving the cognitive expectancy that a given consequence will follow a given behavior (Bouton, 2007; Dickinson & Balleine, 2000).

Learned Helplessness

EXPECTATIONS OF FAILURE AND LEARNING TO QUIT

Cognitive factors, particularly the role of expectation, are involved in another learning phenomenon, called learned helplessness. Learned helplessness was discovered by accident. Psychologists were trying to find out if classically conditioned responses would affect the process of operant conditioning in dogs. The dogs were strapped into harnesses and then exposed to a tone (the neutral stimulus) paired with an unpleasant but harmless electric shock (the UCS), which elicited fear (the UCR). After conditioning, the tone alone—now a CS—elicited the conditioned response of fear.

212

In the classical conditioning setup, the dogs were unable to escape or avoid the shock. But the next part of the experiment involved an operant conditioning procedure in which the dogs could escape the shock. The dogs were transferred to a special kind of operant chamber called a shuttlebox, which has a low barrier in the middle that divides the chamber in half. In the operant conditioning setup, the floor on one side of the cage became electrified. To escape the shock, all the dogs had to do was learn a simple escape behavior: Jump over the barrier when the floor was electrified. Normally, dogs learn this simple operant very quickly.

image
Martin E. P. Seligman: From Learned Helplessness to Positive Psychology Seligman (b. 1942) began his research career by studying learned helplessness in dogs, and later, in humans. He applied his findings to psychological problems, including depression. Seligman developed techniques to teach people to overcome feelings of helplessness, habitual pessimism, and depression (Seligman & others, 2011; Seligman, 2005). Elected president of the American Psychological Association in 1996, Seligman helped launch positive psychology, which emphasized research on human strengths (see Chapter 1). As Seligman (2004) explained, “It became my mission in life to help create a positive psychology whose mission would be the understanding and building of positive emotion, of strength and virtue, and of positive institutions.”
Courtesy of The Positive Psychology Center

However, when the classically conditioned dogs were placed in the shuttlebox and one side became electrified, the dogs did not try to jump over the barrier. Rather than perform the operant to escape the shock, they just lay down and whined. Why?

To Steven F. Maier and Martin Seligman, two young psychology graduate students at the time, the explanation of the dogs’ passive behavior seemed obvious. During the tone–shock pairings in the classical conditioning setup, the dogs had learned that shocks were inescapable. No active behavior that they engaged in—whether whining, barking, or struggling in the harness—would allow them to avoid or escape the shock. In other words, the dogs had “learned” to be helpless: They had developed the cognitive expectation that their behavior would have no effect on the environment.

To test this idea, Seligman and Maier (1967) designed a simple experiment. Dogs were arranged in groups of three. The first dog received shocks that it could escape by pushing a panel with its nose. The second dog was “yoked” to the first and received the same number of shocks. However, nothing the second dog did could stop the shock—they stopped only if the first dog pushed the panel. The third dog was the control and got no shocks at all.

After this initial training, the dogs were transferred to the shuttlebox. As Seligman and Maier had predicted, the first and third dogs quickly learned to jump over the barrier when the floor became electrified. But the second dog, the one that had learned that nothing it did would stop the shock, made no effort to jump over the barrier. Because the dog had developed the cognitive expectation that its behavior would have no effect on the environment, it had become passive (Seligman & Maier, 1967). The name of this phenomenon is learned helplessness—a phenomenon in which exposure to inescapable and uncontrollable aversive events produces passive behavior (Maier & others, 1969).

Since these early experiments, learned helplessness has been demonstrated in many different species, including primates, cats, rats, and fish (LoLordo, 2001). Even cockroaches demonstrate learned helplessness in a cockroach-sized shuttlebox after being exposed to inescapable shock (G.E. Brown & others, 1999).

In humans, numerous studies have found that exposure to uncontrollable, aversive events can produce passivity and learned helplessness. For example, college students who have experienced failure in previous academic settings may feel that academic tasks and setbacks are beyond their control. Thus, when faced with the demands of exams, papers, and studying, rather than rising to the challenge, they may experience feelings of learned helplessness (Au & others, 2010). If a student believes that academic tasks are unpleasant, unavoidable, and beyond her control, even the slightest setback can trigger a sense of helpless passivity. Such students may be prone to engage in self-defeating responses, such as procrastinating or giving up prematurely.

How can learned helplessness be overcome? In their early experiments, Seligman and Maier discovered that if they forcibly dragged the dogs over the shuttlebox barrier when the floor on one side became electrified, the dogs would eventually overcome their passivity and begin to jump over the barrier on their own (LoLordo, 2001; Seligman, 1992). For students who experience academic learned helplessness, establishing a sense of control over their schoolwork is the first step. Seeking knowledge about course requirements and assignments and setting goals, however modest, that can be successfully met can help students begin to acquire a sense of mastery over environmental challenges (Glynn & others, 2005).

Since the early demonstrations of learned helplessness in dogs, the notion of learned helplessness has undergone several revisions and refinements (Abramson & others, 1978; Gillham & others, 2001). Learned helplessness has been shown to play a role in psychological disorders, particularly depression, and in the ways that people respond to stressful events. Learned helplessness has also been applied in such diverse fields as management, sales, and health psychology (Wise & Rosqvist, 2006). In Chapter 12 on stress, health, and coping, we will take up the topic of learned helplessness again.

213

image
Learned Helplessness on the Field In humans, learned helplessness can be produced when negative events are perceived as uncontrollable. Even highly trained athletes can succumb to feelings of learned helplessness in the face of persistent defeats. Athletes who believe that they have no control over the factors that led to their loss or poor performance are less likely to believe that they can succeed in the future (Coffee & others, 2009). They’re also less likely to persist in the face of failure (Le Foll & others, 2008).
Flying Colours Ltd/Getty Images

Operant Conditioning and Biological Predispositions

MISBEHAVING CHICKENS

Skinner and other behaviorists firmly believed that the general laws of operant conditioning applied to all animal species—whether they were pecking pigeons or bar-pressing rats. As Skinner (1956) wrote:

Pigeon, rat, monkey, which is which? It doesn’t matter. Of course, these species have behavioral repertoires which are as different as their anatomies. But once you have allowed for differences in the ways in which they make contact with the environment, and in the ways in which they act upon the environment, what remains of their behavior shows astonishingly similar properties.

However, psychologists studying operant conditioning, like those studying classical conditioning, found that an animal’s natural behavior patterns could influence the learning of new behaviors. Consider the experiences of Keller and Marian Breland, two of Skinner’s students at the University of Minnesota. The Brelands established a successful business training animals for television commercials, trade shows, fairs, and even displays in department stores (Breland & Breland, 1961). Using operant conditioning, the Brelands trained thousands of animals of many different species to perform all sorts of complex tricks (Bihm & others, 2010a, b).

But the Brelands weren’t always successful in training the animals. For example, they tried to train a chicken to play baseball. The chicken learned to pull a loop that activated a swinging bat. After hitting the ball, the chicken was supposed to run to first base. The chicken had little trouble learning to pull the loop, but instead of running to first base, the chicken would chase the ball.

image
Keller and Marian Breland’s “IQ Zoo” B.F. Skinner’s students Keller and Marian Breland opened their animal training business in Hot Springs, Arkansas, in 1951. By the 1960s, the Brelands’ “IQ Zoo” was one of the most popular roadside attractions in the United States. Among its stars were basketball-playing raccoons, reindeer who operated a printing press, ducks who played the piano, and chickens who danced, walked a tightrope, and played tic-tac-toe. Beyond entertainment, the Brelands were pioneers in the development of animal training and behavior modification techniques. Marian Breland was one of the first psychologists to use positive reinforcement to teach basic self-help skills to people with developmental disabilities and also helped train marine mammals for the U.S. Navy (Bihm & others, 2010a, b).
Robert E. Bailey

214

The Brelands also tried to train a raccoon to pick up two coins and deposit them into a metal box. The raccoon easily learned to pick up the coins but seemed to resist putting them into the box. Like a furry little miser, it would rub the coins together. And rather than dropping the coins in the box, it would dip the coins in the box and take them out again. As time went on, this behavior became more persistent, even though the raccoon was not being reinforced for it. In fact, the raccoon’s “misbehavior” was actually preventing it from getting reinforced for correct behavior (Bailey & Bailey, 1993).

The Brelands noted that such nonreinforced behaviors seemed to reflect innate, instinctive responses. The chicken chasing the ball was behaving like a chicken chasing a bug. Raccoons in the wild instinctively clean and moisten their food by dipping it in streams or rubbing it between their forepaws. These natural behaviors interfered with the operant behaviors the Brelands were attempting to condition—a phenomenon called instinctive drift (Bailey & Bailey, 1993).

The biological predisposition to perform such natural behaviors was strong enough to overcome the lack of reinforcement. These instinctual behaviors also prevented the animals from engaging in the learned behaviors that would result in reinforcement. Clearly, reinforcement is not the sole determinant of behavior. And, inborn or instinctive behavior patterns can interfere with the operant conditioning of arbitrary responses.

Before you go on to the next section, take a few minutes to review Table 5.5 and make sure you understand the differences between classical and operant conditioning.

image

Test your understanding of Operant Conditioning with image .