Problem Solving Decision Making and Professional Judgment A Guide For Lawyers and Policy Makers
Problem Solving Decision Making and Professional Judgment A Guide For Lawyers and Policy Makers
paul brest
linda hamilton krieger
1
1
Oxford University Press, Inc., publishes works that further Oxford University’s objective
of excellence in research, scholarship, and education.
With offices in
Argentina Austria Brazil Chile Czech Republic France Greece Guatemala Hungary
Italy Japan Poland Portugal Singapore South Korea Switzerland Thailand
Turkey Ukraine Vietnam
All rights reserved. No part of this publication may be reproduced, stored in a retrieval system,
or transmitted, in any form or by any means, electronic, mechanical, photocopying, recording,
or otherwise, without the prior permission of Oxford University Press, Inc.
_____________________________________________
Brest, Paul
Problem solving, decision making, and professional judgment : a guide for lawyers and
policymakers / Paul Brest, Linda Hamilton Krieger. — 1st ed.
p. cm.
Includes bibliographical references and index.
ISBN 978–0–19–536632–7 (pbk. : alk. paper)
1. Practice of law—Decision making. I. Krieger, Linda Hamilton, 1954– II. Title.
K346.B74 2010
320.6—dc22
2009052053
_____________________________________________
1 2 3 4 5 6 7 8 9
Note to Readers
This publication is designed to provide accurate and authoritative information in regard to the subject
matter covered. It is based upon sources believed to be accurate and reliable and is intended to be current
as of the time it was written. It is sold with the understanding that the publisher is not engaged in rendering
legal, accounting, or other professional services. If legal advice or other expert assistance is required, the
services of a competent professional person should be sought. Also, to confirm that the information has
not been affected or changed by recent developments, traditional legal research techniques should be used,
including checking primary sources where appropriate.
You may order this or any other Oxford University Press publication by
visiting the Oxford University Press website at www.oup.com
For Lee Ross, for his wisdom and generosity, his enormous contribution to social
psychology and the study of judgment and decision making, and his leadership in
demonstrating how insights from the empirical social sciences can be applied to
improve, enrich, and ennoble our societies and ourselves.
This page intentionally left blank
contents
Acknowledgments xxvii
Preface xxix
The title of this book uses the terms problem solving and decision making in
their conventional senses. It uses judgment in two quite different ways. In
common parlance, the term implies good judgment—the capacity to assess situ-
ations shrewdly and to draw sound conclusions.14 We hope that the book will
contribute to improving readers’ judgment in this sense. But we also draw heav-
ily on the field of social science known as “judgment and decision making”
(JDM), in which “judgment” refers mainly to the processes of empiricism—how
one ascertains facts and makes predictions about the physical and social world.
Much JDM research asks how people actually come to judgments and make
decisions; it focuses particularly on the systematic errors made by intuitive deci-
sion makers—all of us, much of the time. In addition to introducing basic ana-
lytic and quantitative tools of decision making, the book surveys the JDM
literature in the hope that understanding these errors can at least sometimes
help avoid them.
1. Anthony T. Kronman, The Lost Lawyer: Failing Ideals of the Legal Profession
72–73 (New York: Cambridge University Press, 1993).
xxx preface
A client with a problem consults a lawyer rather than, say, a psychologist, social
worker, or business advisor because he believes that his problem has a signifi-
cant legal dimension. But real-world problems seldom conform to the boundar-
ies that define and divide different disciplines, and it is a rare client who wants
his lawyer to confine herself strictly to “the law.” Rather, most clients expect their
lawyers to integrate legal considerations with other aspects of their problem.
Solutions are often constrained or facilitated by the law, but finding the best
solution—that is, a solution that addresses all of the client’s concerns—often
requires more than technical legal skills. Indeed, it often turns out that no
solution is ideal in all respects, and that analyzing trade-offs is itself an important
nonlegal problem-solving skill.
Reflecting this reality, an American Bar Association report on the ten “funda-
mental lawyering skills” that new lawyers should acquire places “problem solv-
ing” at the very top of the list—even before legal analysis.1 At their best, lawyers
serve as society’s general problem solvers, skilled in avoiding as well as resolving
disputes and in facilitating public and private ordering. They help clients
approach and solve problems flexibly and economically, not restricting them-
selves to the decision frames that “legal thinking” tends to impose on a client’s
needs. Good lawyers bring more to bear on a problem than legal knowledge and
lawyering skills. They bring creativity, common sense, practical wisdom, and
that most precious of all qualities, good judgment.
Designing and implementing public policy—whether done by lawyers or
people with other professional backgrounds—call for the same attributes. While
counseling and litigating focus on the individual client’s interests, policy making
is intrinsically concerned with many individuals and institutions with different
and often clashing interests. Understanding and accommodating competing,
even incommensurable, interests and designing policies that will change behav-
iors in desired ways are among the policy maker’s fundamental skills.
This chapter inquires into the nature of problem solving and decision making,
both in general and more particularly in lawyers’ work with individual clients
and policy makers’ work in government agencies and nonprofit organizations.
To illustrate problems in both domains, we begin with some vignettes from
a day in the professional life of two characters: Luis Trujillo, a partner at a mid-
sized law firm in Orange County, California; and Christine Lamm, the director
of a county environmental protection agency.
It is an ordinary, if busy, work day for Luis Trujillo. On arriving at his office, he
finds on his calendar an initial consultation with a long-standing client at 9:00
a.m., a noon meeting with Karen Moore, a friend from law school, and an after-
noon conference with one of the firm’s associates to discuss the strategy in a
breach of contract action, which the associate has been handling under Trujillo’s
mentorship.
Trujillo’s nine o’clock meeting is with Jack Serrano, owner of Terra Nueva
Properties, a real estate development company that builds and manages low
and moderate income rental housing projects in the Orange County area.
Serrano takes great pride in his company’s reputation for providing family-
friendly, affordable housing. Until now, the company has enjoyed good rela-
tionships with its tenants and with local, state, and federal government
agencies. In all of his many years in business, Serrano has never been involved
in litigation.
Serrano arrives for his meeting with Trujillo in a state of obvious distress. He
is carrying a copy of the local newspaper, with a front-page story about a wave of
illnesses suffered by his tenants, allegedly the result of a polyurethane foam
product used to insulate the apartments. The article quotes tenants as saying that
the walls of the apartments smell bad, “like chemicals,” and it is accompanied by
a photo of tenants holding a piece of insulation at arm’s length. The article also
contains graphic descriptions of the tenants’ physical ailments and is accompa-
nied by yet another photo—this one of a lawyer and grim-faced residents of Terra
Nueva, captioned “Foam Insulation Syndrome Downs Local Residents—Tenants
to File Class Action Lawsuit.” The article quotes a report of a consumer organiza-
tion saying that similar outbreaks of “foam insulation syndrome” have occurred
elsewhere in the country.2 We return to Trujillo’s meeting with Jack Serrano
later in this and subsequent chapters.
Trujillo does pro bono work for the Los Angeles Volunteer Legal Services
Association (VLSA). After finishing his meeting with Serrano, he turns to a
memorandum from VLSA concerning mass firing of employees, without notice,
when a small manufacturing plant decided to move its operations to Mexico.
2. For an example of how life immitates hypothetical problems, see Leslie Wayne,
Chinese Drywall Linked to Corrosion, New York Times, November 23, 2009, http://www.
nytimes.com/2009/11/24/business/energy-environment/24drywall.html?scp=1&sq=chi
nese%20drywall&st=cse.
problem-solving and decision-making processes 5
The terminations do not appear to violate either federal or state statutes, but
Trujillo has the germ of an idea of how to deal with this (to which we will return
in Chapter 3).
Trujillo’s musings are interrupted by a phone call from the front desk, alert-
ing him that Karen Moore has arrived and is headed back to his office. Moore is
a vice president for Big-Mart, a chain of discount department stores in the region.
Trujillo has helped negotiate many real estate contracts for Big-Mart, which has
grown quickly to have over thirty stores. Trujillo and Moore spend most of the
time discussing a complex deal involving a new location. But toward the end of
the lunch, Moore presents a quite different problem.
On Trujillo’s advice some years ago, Big-Mart has done regular internal audits
to ensure that it is in compliance with the law, rather than await regulatory
actions or litigation.3 She reports that the human resources director has taken an
extensive look at Big-Mart’s employment records, and has discovered an unset-
tling disparity between the salaries of male and female assistant managers. The
average male assistant manager makes $39,257 a year, while the average woman
makes $38,528—a disparity of $729.
Trujillo wonders whether the disparity might be due to other factors, such as
seniority or education. Resolving this question will require the statistical analysis
of Big-Mart’s employment data, something we will defer to Part 2 of the book.
Later in the day, Trujillo meets with Anna Wilkins, a recent law school gradu-
ate and new associate at the firm. Before turning to the main point of the meet-
ing, a breach of contract case, Trujillo mentions an incident in a trial in a tort
case—the first trial in which she ever acted as lead counsel. Anna had been about
to object to a question on the ground that it called for hearsay, when Trujillo
tugged at her sleeve and indicated that she should let it pass. Wilkins says
that she has since checked her recollection of the law. The response would
certainly have been inadmissible, and she wonders why Trujillo stopped her
from objecting. “You’re absolutely right on the law,” he says, “but we’re really
not contesting that particular factual issue. Moreover, we had been making quite
a few objections, and the judge was communicating her increasing irritation to
the jury.”
They then discuss the breach of contract case that Wilkins is handling for
the firm. The firm’s client, Clyde Evers, has sued Newport Records, a small
recording company. Newport refuses to pay for accounting software that Evers
customized and installed, saying that the software does not do what Evers said it
would do. The amount due is $600,000.
3. See Thomas D. Barton, Preventive Law and Problem Solving: Lawyering for
the Future (Lake Mary, FL: Vandeplas Publishing, 2009).
6 problem solving, decision making, and professional judgment
Yesterday, Wilkins received a phone call from Evers, who seemed upset that
nothing had happened since the case was filed quite a long time ago, and asked
her whether she couldn’t hasten its resolution.
Based on her knowledge of summary judgment from her Civil Procedure
class in law school, her reading of the contract (which disclaims any warranty of
performance), and her study of the relevant law concerning warranties, Wilkins
believes that Evers can win on summary judgment and proposes to file a motion
to that effect. After examining the case file, Trujillo introduces Wilkins to the
practical realities of summary judgment practice in the state courts.
Trujillo explains that even though a motion for summary judgment could
theoretically bring about a quick disposition of the case, it could have untoward
consequences. The judge before whom the motion is likely to be argued views
summary judgment—especially for plaintiffs—with considerable skepticism.
It is true that the contract disclaims any warranty of performance. But it
appears that Evers had made extravagant oral representations about what
the software would do. Even if those representations are not formally binding,
they may bias the judge further against summary judgment once he learns
of them.
Moreover, the law requires that warranty disclaimers be in a particular type-
face, which is somewhat different from the disclaimer in the contract with
Newport. The judge might regard the difference as inconsequential and grant
summary judgment; or he might have the jury determine whether or not the
defendant actually read and understood the disclaimer.
And there is yet another problem. The defendant claims that after a brief trial
period it stopped using Evers’ software and purchased an off-the-shelf product
instead. If the written disclaimer of warranty is ineffective, Newport may have a
claim against Evers for breach of warranty. But the statute of limitations on
this claim—which is much shorter than the statute of limitations governing
Evers’ claim—is about to run out. The defendant’s lawyer, a local sole practitioner
not known for high-quality work, probably hasn’t been focusing on the case; but
the motion may lead him to pay attention and file the claim.
What’s more, Trujillo explains, the cost of litigating the motion for summary
judgment will not be insubstantial. If the motion is denied, Evers’ costs would be
greatly increased. Even if the motion is granted, Newport Records will likely
appeal, with attendant costs and the possibility of still having to go to trial.
“Hmmm, I take your points,” says Wilkins. “We’ll just have to wait until the
case comes to trial.” But Trujillo responds, “Not so fast. Did Evers give you any
indication why he was upset that the case wasn’t progressing? Surely you dis-
cussed the time frame with him at an earlier point.” Wilkins replies that Evers
mentioned that he was anxious to have the funds from the judgment to invest in
a new venture.
Trujillo then asks Wilkins whether she can think of any available options
beside moving for summary judgment and waiting for trial, and eventually
problem-solving and decision-making processes 7
they discuss the pros and cons of approaching Newport Records’ lawyer with a
settlement offer.
Christine Lamm received a joint degree in law and public policy only ten years
ago. A deputy administrator in the county’s environmental protection agency,
she was catapulted into the role of acting administrator of the department upon
the sudden departure of her boss two years ago. Last year, with some misgivings
based on her lack of experience and her tendency to do everything “by the book”
in an overly deliberative manner that did not always take political realities into
account, the mayor formally appointed her head of the department. She serves
in that capacity “at the pleasure” of the mayor, meaning that she can be removed
by the mayor at will.
Lamm begins the day by continuing to work on a complex project involving
the siting of a wastewater treatment plant in Edenville. But the work is inter-
rupted (and will not be resumed until Chapter 4) by an urgent phone call from
Paula Henderson, the mayor’s chief of staff, about the Terra Nueva affair. The
mayor saw the same newspaper articles that brought Serrano to Luis Trujillo’s
office that morning, and he wants something done about the foam insulation
problem “ASAP.” Henderson asks Lamm to draft a set of proposed regulations
banning use of polyurethane foam insulation in new construction and renovation
projects in the county.
Lamm listens anxiously to Henderson’s request, mindful that her continued
employment turns on her ability to remain in the mayor’s good graces. But
Lamm doesn’t just have her job to worry about; she feels personally and profes-
sionally committed to approaching the foam insulation problem at Terra Nueva
in a manner consistent with principles of sound public policy making. Quickly
calculating how best to mediate between these two sets of concerns, Lamm
explains to Henderson that, under the state’s Administrative Procedures Act, any
regulatory initiative banning the foam insulation will have to be premised on
agency findings—arrived at using scientifically acceptable methods—that the
foam was in fact causing harm. After a rather pointed response, underscoring
the mayor’s desire to act decisively in response to requests from community
groups that had long supported him, Henderson agrees with Lamm’s suggestion
that she convene a working group to investigate the causal connection between
the insulation product and the symptoms experienced by Terra Nueva residents.
After getting off the phone with Henderson, Lamm decides to take a walk to
clear her head, and to give a maintenance worker access to her office to repair a
light above her desk. When she returns, she notices that her computer screen is
dark, and recalls that she had a number of word-processing and spreadsheet
documents open when she left. Grumbling to herself that the last thing she
8 problem solving, decision making, and professional judgment
needs on a stressful day like this is a bunch of lost work, she checks the electric
plug, thinking, “I’ll bet the maintenance guy accidentally unplugged it when he
moved my desk; that sort of thing has happened to me before.” She checks under
her desk where the electrical socket is located. “It’s plugged in, but it seems a bit
loose,” she thinks. “I’ll jiggle it.” In the process, the plug comes completely out
of the socket, and the background sound of the hard drive and fan, which she
had not noticed until then, suddenly goes silent. The computer had not been off
before, but it is now. On further inspection, Lamm discovers that the screen was
dark because the cable connecting the computer to the monitor had come loose.
Now she in fact has lost some work.
As these vignettes suggest, in their day-to-day work, lawyers and public policy
makers are constantly working to solve problems, either alone or in collaboration
with others. The qualities they need to do this well are sometimes defined in
terms of judgment or practical wisdom, the skills in terms of problem solving
and decision making.
The academic and professional literatures provide a variety of definitions of
the term problem. For example, Charles Kepner and Benjamin Tregoe define a
problem as a situation where “something has gone wrong.”4 This definition cap-
tures at least two of the situations described above. “Something has gone wrong”
at Jack Serrano’s Terra Nueva apartments, and something “went wrong” with
Christine Lamm’s computer.
More broadly, law professor Gerald Lopez defines a “problem” as a situation
in which “the world we would like varies from the world as it is.”5 Correspondingly,
Lopez defines problem solving as “trying to move the world in the desired
direction.”6 This definition aptly describes Christine Lamm’s project for siting
the wastewater treatment plant. It also describes the situation presented by
VLSA’s potential case on behalf of the laid-off woodworkers. From the laid-off
workers’ perspective, “the world they would like” is one with a legally enforce-
able right to reasonable notice of pending layoffs, if not protection from the lay-
offs themselves. However, “the world as it is” apparently provides no such claim.
Crafting a novel legal theory, and then persuading a judge to apply it and provide
a remedy, represents an effort to “move the world in the desired direction.”
Problems also include situations where nothing has gone wrong yet, but
where there is reason to believe that if some action is not taken, something may
4. Charles H. Kepner and Benjamin B. Tregoe, The New Rational Manager viii
(Princeton: Princeton, 1981).
5. Gerald P. Lopez, Lay Lawyering 32 UCLA Law Review, 2 (1984).
6. Id.
problem-solving and decision-making processes 9
go wrong in the future. Problem solving in these cases calls for the deployment
of strategies calculated to head off foreseeable future problems. Much of lawyers’
work in their roles as counselors, deal makers, estate planners, and legislative
drafters involves anticipating and avoiding problems that might arise. That’s
why Trujillo recommended that Big-Mart do a regular legal audit—the equiva-
lent of an annual physical exam. To take another example, the standard lease
agreement that Trujillo had drafted years ago for Jack Serrano tries to anticipate
various things that might “go wrong” with a particular tenancy—persistent loud
parties disturbing to neighbors, chronically late or unpaid rent, undesirable
subletting arrangements, to name a few. Part of Trujillo’s craft as a lawyer
involves his ability to anticipate problems of this sort and to work into the lease
agreement mechanisms through which Serrano can address them quickly,
effectively, and at the lowest possible cost. Unless anticipated ahead of time, a
“something may go wrong” problem can easily become a “something has gone
wrong” problem.
To accommodate problems of these various types, we adopt a more inclusive
definition of the term problem, similar to that suggested by Allen Newell and
Herbert Simon: a “problem” is any situation in which the state of affairs varies,
or may in the future vary, from the desired state, and where there is no obvious
way to reach the desired state.7 For example, we will see in Chapter 4 that there
is no single obvious solution to Christine Lamm’s problem of where to site the
wastewater treatment plant.
Newell and Simon define the conceptual area between the existing and desired
states of affairs as a problem space. To solve a problem is to navigate through the
problem space—through the virtual area between the actual or potential unsatis-
factory state and the desired state. We can represent this conception of a problem
and the problem solving process in the following way, as shown in Figure 1.1.
PRESENT DESIRED
STATE The Problem Space STATE
The various pathways through a problem space may not be equally satisfac-
tory. Some pathways that appear promising at the outset may ultimately prove to
be dead ends. Some are inferior because they demand the expenditure of exces-
sive resources, or create new problems even while solving the original one, or
because they compromise other objectives. Problem solving is thus subject to
what Newell and Simon term path constraints. A completely satisfactory solution
is a path that leads through the problem space and is consistent with all relevant
constraints.
The process described by Newell and Simon can be analogized to an expedi-
tion by explorers who must cross uncharted territory to get to their desired des-
tination. The explorers may take one path, only to have to backtrack after
discovering that it ends at a steep cliff. They may take another and encounter a
wide river. How to cross the river in effect poses an ancillary problem, which
must be solved to continue the journey. As for path constraints, the explorers
might be vegetarians, or might have mores that preclude them from traveling on
the Sabbath, which could hinder their progress toward the destination but nev-
ertheless must be taken into account.
If “problem solving” consists of “trying to move the world in the desired direc-
tion,” it must ultimately eventuate in a decision—a “commitment to a course of
problem-solving and decision-making processes 11
action that is intended to produce a satisfying state of affairs.”8 In terms of the pre-
ceding discussion, decision making involves choosing a particular pathway across
the problem space that lies between the actual and desired states of affairs. The
“best” solution to a problem is one that satisfies, to the greatest extent possible, the
broadest range of objectives, including constraints, implicated by the problem.
8. J. Frank Yates, Elizabeth S. Veinott, and Andrea L. Patalano, Hard Decisions, Bad
Decisions: On Decision Quality and Decision Aiding, in Emerging Perspectives on
Judgment and Decision Research 13–63 (Sandra L. Schneider and James Shanteau
eds., New York: Cambridge University Press, 2003).
12 problem solving, decision making, and professional judgment
Figure 1.4 illustrates how the model of deliberative problem solving just
described above combines elements of divergent and convergent thinking.
Early in the process, when a problem is being framed, when interests
and objectives are being identified, and when alternative solutions are being
generated, divergent thinking can bring a great deal of value to the problem-
solving endeavor. Divergent thinking enables us to conceptualize the problem
from a wide variety of perspectives, so as to permit consideration of the broadest
possible array of potential solutions. Divergent thinking helps identify the full
range of interests implicated by a particular decision. And divergent thinking
inspires innovation in coming up with solutions to the problem. Later in the
process, convergent thinking comes into play in analyzing causation, evaluating
options, choosing which course of action to implement, and implementing and
monitoring the choice.
9. Hubert Dreyfus and Stuart Dreyfus, Mind Over Machine: The Power of
Human Intuitive Expertise in the Era of the Computer 29 (New York: Free Press,
1986).
10. John Bargh and Tanya Chartrand, The Unbearable Automaticity of Being, 54
American Psychologist 462 (1999).
11. Kenneth Hammond, Human Judgment and Social Policy: Irreducible
Uncertainty, Inevitable Error, Unavoidable Injustice 60 (New York: Oxford
University Press, 1996).
problem-solving and decision-making processes 15
Divergent
thinking
Convergent
thinking
effort, and typically without conscious awareness. They involve little or no con-
scious deliberation.”12
The [chess] grand masters . . . use their intuition to recognize the promis-
ing moves that they should examine more closely. They shift to an analytic
mode by looking at the moves they will play out in the context of the
game, and rely on their ability to mentally simulate what will happen if
they play a move. In the course of these mental simulations, some of the
moves drop out because they are found to contain weaknesses. By the end
of the mental simulations, the grand masters are usually left with only a
single move they find playable.
— Gary Klein, Intuition at Work 75 (2003).
17. Gary Klein, A Case Study of Intuition, in Intuition at Work: Why Developing
Your Gut Instincts Will Make You Better at What You Do 3–9 (New York:
Doubleday, 2002).
18. Id. at 11, 12–13.
18 problem solving, decision making, and professional judgment
The heart has its reasons that reason does not understand.
—Blaise Pascal
Intuitive problem solving and decision making depends not only on the
essentially mental processes of recognizing patterns, but on affect as well. In
recent years, researchers have given increasing attention to the role of affect in
decision making—ranging from a “faint whisper of emotion to strong feelings
of fear and dread, to visceral drives such as hunger and sexual need.”26 Thomas
Hoving writes that the art historian, Bernard Berenson
27. Thomas Hoving, False Impressions: The Hunt for Big Time Art Fakes 19–20
(New York: Simon & Schuster, 1996). For a discussion of the role of intuition in the
discovery of another art fraud, see Malcolm Gladwell, Blink: The Power of Thinking
Without Thinking 3–17 (New York: Little Brown, 2005).
28. Benedict Carey, In Battle, Hunches Prove to be Valuable, New York Times, July 28,
2009, http://www.nytimes.com/2009/07/28/health/research/28brain.html?pagewanted=
2&_r=1&hp.
29. Antoine Bechara, Hanna Damasio, Daniel Tranel, Antonio R. Damasio, Deciding
Advantageously Before Knowing the Advantageous Strategy, 275 (5304) Science 1293–95
(Feb. 28, 1997).
30. Antonio Damasio, Descartes’ Error 200 (New York: G.P. Putnam’s Sons, 1994).
problem-solving and decision-making processes 21
skilled professional must be able to differentiate between her own emotions and
those of clients and others. This capacity is a component of so-called emotional
intelligence.31 Like many of the other skills of intuition, these capacities are best
improved through reflective experience and practice.
Process Characteristics
Automatic Controlled
Effortless Effortful
Associative Deductive
Rapid, parallel Slow, serial
Process opaque Self-aware
Skilled action Rule application
Content on which Processes Act
Affective Neutral
Causal properties Statistics
Concrete, specific Abstract
Prototypes Sets
31. See Daniel Goleman, Emotional Intelligence: Why It Can Matter More
than IQ (10th Anniversary Edition New York: Bantam, 2006); John D. Mayer, Peter
Salovey, and David Caruso, Models of Emotional Intelligence, in Handbook of Intelligence
396–420 (Robert Sternberg ed., 2d ed. New York: Cambridge University Press, 2000).
32. Daniel Kahneman and Shane Frederick, Representativeness Revisited: Attribute
Substitution in Intuitive Judgment, in Heuristics and Biases: The Psychology of
Intuitive Judgment, supra at 49.
22 problem solving, decision making, and professional judgment
courses of action (moving for summary judgment and doing nothing), and
concludes that one is superior to the other. While he reaches this conclusion
intuitively and quickly, he confirms his hypothesis more systematically in the
discussion with Wilkins. Then, recognizing that neither solution meets all of the
client’s objectives, Trujillo broadens the problem frame—a strategy characteris-
tic of more deliberative problem-solving processes.
As Hubert and Stuart Dreyfus note, “when time permits and much is at stake,
detached deliberative rationality . . . can enhance the performance of even the
intuitive expert. . . . [S]uch deliberation tests and improves whole intuitions.”40
Note also—and this is an important element of interaction between lawyers and
their clients as well as associates—that the very fact of discourse with Wilkins
forces Trujillo to articulate intuitive judgments, a process which tends to move
decision making toward analysis and deliberation.
Analysis and intuition work together in the human mind. Although intuition is the
final fruit of skill acquisition, analytic thinking is necessary for beginners learning a
40. Hubert Dreyfus and Stuart Dreyfus, Mind Over Machine: The Power of
Human Intuitive Expertise in the Era of the Computer 40 (New York: Free Press,
1986).
26 problem solving, decision making, and professional judgment
new skill. It is also useful at the highest levels of expertise, where it can sharpen and
clarify intuitive insights.
—Hubert and Stuart Dreyfus, Mind Over Machine41
What is the relationship between expertise and the two models of problem
solving outlined above? Experts and others with proficiency in a subject often
rely heavily on intuition, but their intuitions are usually informed and comple-
mented by deliberation.
Before turning to expertise, consider the various nonprofessional activities in
which you have know-how, competence, or proficiency—for example, crossing the
street, driving a car, cooking, or playing or watching a sport. Although expertise
denotes a proficiency (whether in medicine, law, or poker) that sets someone
apart from most fellow citizens, we develop competence in these activities of
everyday life much as an expert does in a professional domain—through obser-
vation, education, mentorship, and feedback from experience. Recall, if you can,
the transition from novice to proficient actor—when you learned how to drive,
for example—as you moved from approaching each activity deliberatively,
often with conscious reference to rules or procedures, to (for the most part) just
“doing it.”
Both everyday competences and expertise are domain-specific. You can be an
expert chef and have no idea what’s going on in a football game. In Life on the
Mississippi, Mark Twain writes:
Give a man a tolerably fair memory to start with, and piloting will develop it
into a very colossus of capability. But only in the matters it is daily drilled in.
A time would come when the man’s facilities could not help noticing land-
marks and soundings, and his memory could not help holding on to them
with grip of a vise; but if you asked the same man at noon what he had had for
breakfast, it would be ten chances to one that he could not tell you. Astonishing
things can be done with the human memory if you will devote it faithfully to
one particular line of business.42
In what may be one of the earliest texts to connect judgment and expertise,
Twain goes on to describe other qualities that an expert must have:
A pilot must have a memory; but there are two higher qualities which he
must also have. He must have good and quick judgment and decision, and a
cool, calm courage that no peril can shake. Give a man the merest trifle of
pluck to start with, and by the time he has become a pilot he cannot be
unmanned by any danger a steamboat can get into; but one cannot quite say
the same for judgment. Judgment is a matter of brains, and a man must start
with a good stock of that article or he will never succeed as a pilot.43
In a study of nurses’ development of expert judgment, Patricia Brenner quotes
this description of a novice nursing student:44
I gave instructions to the new graduate, very detailed and explicit instructions:
When you come in and first see the baby, you take the baby’s vital signs and
make the physical examination and you check the I.V. sites, and the ventilator
and make sure that it works, and you check the monitors and alarms. When I
would say this to them, they would do exactly what I told them to do, no
matter what else was going on. . . . They couldn’t choose which one was the
most important. . . . They couldn’t do for one baby the things that were most
important and then go to the other baby and do the things that were the most
important, and leave the things that weren’t as important until later on. . . .
If I said, you have to do these eight things, . . . they did these things, and
they didn’t care if their other kid was screaming his head off. When they did
realize, they would be like a mule between two piles of hay.
Hubert and Stuart Dreyfus note that the solution for the student is a set of
hierarchical decision-making rules, spelling out priorities. As she develops
expertise, however, these largely give way to an intuitive sense of the situation.
What does an expert professional, like Luis Trujillo, bring to problem solving
in a particular domain that a nonexpert would be unable to provide? Drawing on
Newell and Simon’s metaphor of problem solving as navigation across a virtual
problem space, an expert possesses special navigational knowledge or skill with
respect to problems of a particular type. Experts often know about pathways
through a problem space of which nonexperts are unaware. Though several
paths may look promising at the outset, an expert may be better able to predict
which ones will prove to be dead ends or worse.
Thus, a person might hire an expert to navigate a particular type of problem
space for the same reasons that a novice group of explorers would hire a guide
rather than set out entirely by themselves. An experienced guide has the ability to
“read” the terrain, finding likely routes and avoiding perils that would be invisible
to the novice. Indeed, in law, medicine, and some other domains, only a licensed
professional is permitted to guide another person across a problem space.
Consider Luis Trujillo’s expertise in serving as a guide for his client, Jack
Serrano, with respect to the complaints of the tenants at Terra Nueva. Trujillo
43. Id. at 118–19. We are grateful to James Shanteau, Mark Twain on Expertise, http://
mail.sjdm.org/pipermail/jdm-society/2007-August/003168.html, for this quotation.
44. Patricia Brenner, From Novice to Expert: Excellence and Power in Clinical
Nursing Practice 23 (Upper Saddle River, NJ: Pearson Higher Education, 1984), quoted
in Dreyfus and Dreyfus.
28 problem solving, decision making, and professional judgment
must know the substantive law and relevant procedural rules bearing on the
threatened suit whether or not it is filed. Such knowledge is necessary back-
ground to considering other means of resolving the problem. If Trujillo does not
know the law in this particular area, he has the expertise to research and learn it.
(This ability to do legal research is an area of expertise initially acquired at law
school and then honed on the job.)45
By the same token, in considering the siting of the wastewater treatment
plant, Christine Lamm will combine procedural knowledge (likely acquired in
public policy school) about how to structure decisions of this sort with substan-
tive knowledge (probably acquired on the job) of the factors relevant to the deci-
sion. And in addressing the Terra Nueva problem, it is possible that both Trujillo
and Lamm will need to acquire knowledge about medicine, epidemiology,
and the properties of polyurethane foam—to understand the connections
between exposure to the insulation and the claimed illnesses. They won’t have
time to—and won’t need to—acquire the extensive knowledge that would make
them either intuitive or sophisticated analytic problem solvers in these domains;
rather, they will have to rely on other experts.
Trujillo and Lamm and the scientific experts on whom they may call differ
from laypersons not merely in the quantity of domain-specific knowledge, but in
the quality of its cognitive organization. Experts possess large sets of particular
schemas, which describe the attributes of domain-relevant problems and con-
tain solutions to them.46 The development of professional expertise entails learn-
ing many increasingly nuanced expert schemas. Once learned, these schemas
enable the expert to recognize recurring patterns, rapidly draw inferences from
them, and execute effective responsive interventions.
Robin Hogarth asserts that the essential difference between experts and nov-
ices is that they process knowledge differently. Not surprisingly, his description
bears a strong resemblance to Gary Klein’s description of recognition-primed
decision making:47
First, experts acquire habits that help them process more information. They
learn to counteract limitations in short-term or working memory and are able
to “chunk” information more effectively. In other words, despite facing
normal limitations on memory, experts find ways of restructuring informa-
tion so that they can take more into account . . . For example, whereas a
medical student might see a patient’s symptoms as several different items of
45. See Gary Marchant and John Robinson, Is Knowing the Tax Code All It Takes to be a
Tax Expert? On the Development of Legal Expertise, in Tacit Knowledge in Professional
Practice 3–20 (Robert Sternberg and Joseph Horvath eds., Mahwah, NJ: Lawrence
Erlbaum, 1999).
46. Kurt Van Lehn, Problem Solving and Cognitive Skill Acquisition, in Foundations of
Cognitive Science 527, 545–46 (Michael I. Posner ed., Cambridge: MIT Press, 1989).
47. Hogarth, Educating Intuition, supra at 158.
problem-solving and decision-making processes 29
To understand just what sort of expertise lawyers and policy makers possess,
consider these examples of the kinds of tasks they are called upon to perform:
• An appellate lawyer seeks to influence the course of the law by persuading
a court that common-law doctrines prohibiting race discrimination by
common carriers should prohibit sexual orientation discrimination by
landlords.
• A public health administrator is asked by the mayor to develop a set of
initiatives to address the disproportionate rate of asthma-related deaths of
minority children in a large city in California.
• A company’s general counsel works with its president and chief operating
officer to develop a procedure for responding to sexual harassment claims
if any should arise; or responds to the crisis occasioned by a particular
accusation of harassment.
• A business lawyer identifies risks involved in a movie deal, including other
parties’ incentives to behave strategically, and structures the transaction so
as to mitigate the risks to his client.
30 problem solving, decision making, and professional judgment
1.8.1 Lawyers’ Knowledge About the Law, Legal Institutions, and Actors
In the first of these vignettes, the lawyer’s task of persuading an appellate court
to extend settled case law to a new area calls for creative problem solving with
respect to both analogical reasoning and advocacy. The task draws on the skills
of doctrinal analysis, legal research, writing, and advocacy—many of which play
a background role in the other tasks as well. The vignettes also highlight other
kinds of legal expertise, such as knowing how to influence judges, juries, admin-
istrative officers, and other actors in the legal system.
The frame in which one views a problem or decision is a function of the decision
context,2 which is determined both by the values, interests, or objectives at stake
and by the authority of the decision maker. People may address the same prob-
lem in different decision contexts. For Frank Serrano and his lawyer, Luis
Trujillo, litigation determines the initial context for addressing the problem at
Terra Nueva, but the broad range of possible stakeholders and interests suggest
the possibility of different or broader decision contexts as well.
1. John Dewey, Logic: The Theory of Inquiry (New York: Holt, Rinehart, and
Winston, 1938).
2. Ralph Keeney, Value-Focused Thinking: A Path to Creative Decisionmaking
30 (Cambridge: Harvard University Press, 1992).
34 problem solving, decision making, and professional judgment
The major constraints on creative problem solving do not arise from jurisdic-
tional limitations on decision contexts, but from human psychology. Chapter 1
introduced the concept of “schemas,” which structure people’s perception of the
environment. We schematize problems in the same ways that we schematize
everything else. Problem schemas, also referred to as “problem frames,” serve
important cognitive functions. They enable us to perform a wide range of
problem-solving tasks intuitively, with little expenditure of cognitive resources.
But like all knowledge structures, problem frames have a potential downside
3. Id. at 205.
framing problems and identifying causes 35
4. Varda Liberman, Steven M. Samuels, and Lee Ross, The Name of the Game: Predictive
Power of Reputations versus Situational Labels in Determining Prisoner’s Dilemma Game
Moves, 30 Personality and Social Psychology Bulletin 1175–85 (2004).
36 problem solving, decision making, and professional judgment
People often solve the wrong problem. They may mistake symptoms of a problem
for the problem itself, define the problem too narrowly, or define the problem in
terms of one particularly salient, but not necessarily optimal, solution. Problem
framing can occur automatically with little or no conscious, considered thought.
Experts, who “over-learn” reoccurring problem frames as part of their profes-
sional training, are particularly susceptible to automatic problem framing.
The most common problem-framing errors can be divided into three broad
groups:
1. defining the problem in terms of one salient potential solution;
2. mistaking a salient symptom of the problem for the deeper problem itself;
and
3. defining a multidimensional problem unidimensionally, often as a result
of “expert,” or otherwise automatic, problem-framing processes.
We consider each of these in turn.
6. Jacob Getzels, Problem Finding and the Invention of Solutions, 9 Journal of Creative
Behavior 12, 15–16 (1975).
38 problem solving, decision making, and professional judgment
the trust and tax issues are relatively minor compared to questions about how
the children will participate harmoniously in running the farm and share in its
profits (or losses). She knows from experience with other family businesses that
whatever stability in family relations may exist while the parent is actively
running the enterprise often dissolves upon the parent’s retirement or death.
The clients’ problem frame—“how do we set up a GRIT?”—failed to capture
important dimensions of their actual problem: How should we structure our
estate plan to best provide for our children? Helping her clients develop a better
problem frame was an important aspect of the lawyer’s job.
The most powerful protection against shallow, solution-based problem frames
is the deceivingly simple question, “Why?” For example, by asking the clients
what they want to accomplish, the lawyer leads her clients to reframe their prob-
lem as, “How can I best provide for the needs of our family after our deaths?”
Once the problem frame is expanded past the false boundaries drawn by the
presumed solution of drafting a particular instrument, the utility of other inter-
ventions becomes apparent. By asking “why” until a client’s deepest practical
goals and objectives are recognized, a lawyer can assist her client in solving the
right problem.
Ralph Keeney provides a good example from the realm of public policy, involv-
ing the decision of how to transport hazardous material to a distant waste dump:
One objective may be to minimize the distance the material is transported by
trucks. The question should be asked, “Why is this objective important?” The
answer may be that shorter distances would reduce both the chances of acci-
dents and the costs of transportation. However, it may turn out that shorter
transportation routes go through major cities, exposing more people to the
hazardous material, and this may be recognized as undesirable. Again, for
each objective concerning traffic accidents, costs, and exposure, the question
should be asked, “Why is this important?” For accidents, the response may be
that with fewer accidents there would be fewer highway fatalities and less
exposure of the public to hazardous material. And the answer to why it is
important to minimize exposure may be to reduce the health impacts of the
hazardous material. To the question, “why is it important to reduce health
impacts?” the response may be that it is simply important. This indicates that
the objective concerning impacts on public health is a candidate to be a
fundamental objective in the decision context.7
that lawyers and policy makers face are situations where something has “gone
wrong” rather than opportunities to keep things from going wrong or, better yet,
to improve the world. The uproar over Terra Nueva is the trigger that gets the
immediate attention of Luis Trujillo and Christine Lamm. Because the events
are vivid, consequential, and emotionally freighted, it is easy to experience the
trigger as the problem itself. But this is not necessarily so. The trigger may just
be the symptom of a deeper problem.
Of course, this does not mean that the symptom should be ignored. More
often than not, the issue that triggered the problem demands attention. The
mistake is in failing to identify the relationship between the trigger and the
deeper, often multidimensional, state of affairs that gave rise to the problem.
Consider the following example. In 1992, an undercover investigation by the
California Department of Consumer Affairs caught Sears Auto Centers system-
atically defrauding customers by selling them unnecessary repairs. A parallel
investigation in the State of New Jersey uncovered a similar pattern of fraudulent
sales activity by Sears Auto Centers located there. After initially denying that
anything improper had occurred, Sears eventually admitted that “mistakes had
occurred” and paid many millions of dollars to settle the two matters. If Sears’s
problem were defined as “resolving the enforcement actions at the lowest feasi-
ble cost,” the multimillion dollar settlements might well be viewed as a success-
ful solution.8
However, consumer advocates and eventually Sears itself defined the prob-
lem differently. The problem was not just the state enforcement actions or even
the fraudulent sales themselves. A deeper problem, and the cause of these symp-
toms, lay in Sears’s management practices, which encouraged dishonest behav-
ior. These practices included the imposition of mandatory repair-dollars-per-hour
quotas on mechanics, the use of a commission-based compensation system, and
the deployment of high-stakes contests—all designed to encourage employees to
maximize repair sales. Seen through this lens, the fraudulent behaviors and the
enforcement actions themselves were symptoms of a deeper problem involving
Sears’s management and compensation practices. Solving the problem required
changing these practices, as well as resolving the lawsuits.
As this example suggests, effective problem framing often requires careful
thinking about the causal antecedents that gave rise to a particular problem
trigger:
1. Generate an initial statement of the problem—an “initial frame
statement.”
2. Identify the “trigger” that instigated the problem-solving procedure in the
first place.
2.3.3 Automatic Problem Framing and the Pitfalls of Expert Problem Frames
Problem solving often occurs automatically, through the intuitive activation and
application of well-learned problem schemas. In these situations, some salient
aspect of the problem activates a stored representation of “problems-of-this-type”
residing in memory. Once activated, the schema directs the problem solver’s
attention to whatever information the schema contains. This may include a
ready-made problem frame, a set of plausible solutions, and an information
search blueprint, telling the schema-holder what type of unknown information
is relevant to solving the problem and how to go about obtaining and using that
information to select the most effective solution strategy.
Experts, including lawyers and specialized policy makers, are particularly sus-
ceptible to automatic problem framing. The process of developing professional
expertise entails learning sets of “expert frames.” These frames efficiently cap-
ture aspects of the situation that are relevant to the professional’s particular area
of expertise. But they are inevitably rather narrow, often failing to capture impor-
tant dimensions of the problem as it is actually experienced by clients, citizens,
and other stakeholders.
Lawyers tend to frame problems differently from the way their clients do
because they approach the problem with different schematic mind sets. The
lawyer hears the client’s story through the filters provided by the lawyer’s expert
schemas. These tell him which aspects of the client’s narrative are important and
which are not. They direct the lawyer to follow up on certain subjects, asking
additional questions and probing for detail, and to cut the client short when she
dwells too long on “irrelevant” matters. In this way, aspects of the situation that
the client subjectively experiences as important may become invisible to the
lawyer, and the entire matter may head in a direction that poorly serves the
client’s interests, broadly conceived.9
9. Conversely, the client may come to the lawyer wanting a legal solution, but having
a mistaken impression of the nature of applicable legal constraints. Because he is
framing problems and identifying causes 41
Consider the different ways in which the problem at Frank Serrano’s Terra
Nueva apartments might be framed. If Serrano had consulted a stereotypic
“hardball” litigator, the lawyer might have conceptualized the problem simply as
being “sued” by the tenants and the solution as winning a decisive victory in
court.
But the lawsuit may be only one aspect of a multidimensional problem.
Serrano has been the subject of a front page news story that, he worries, may be
followed by others. With his reputation at stake, Serrano does not just have a
lawsuit problem, but a public relations problem as well—a problem that could be
exacerbated by a no-holds-barred litigation defense strategy. And what about
Serrano’s relationships with his company’s various stakeholders and constituen-
cies: tenants, community activists, financial backers, regulators, and his contrac-
tors and suppliers?
Though successfully defending Serrano against the suit will constitute an
important element of a successful outcome, these other aspects of the situation
may also be important to Serrano. If he is to do his job well, Trujillo must act as
a good counselor as well as a skilled litigator. He must also help Serrano identify
nonlegal aspects of the situation that he may experience as important problems
now or in the future.
A good lawyer understands that, even though it was the client’s identification
of the problem as a “legal” one that brought him to the office, a purely legal
frame may be too narrow. But a lawyer who reflects on his or her own profes-
sional development will be aware of the forces that induce such framing.
Professional expertise often works like a kind of zoom lens, focusing in on one
small portion of a broad landscape, revealing its features in great detail. This
focus—an intentional myopia—facilitates the accurate, efficient, and effective
identification of problems and the rapid deployment of interventions designed
to solve them. But it can also lead the expert to miss important features of the
larger picture, with significant collateral consequences for what lies beyond his
field of vision.
The solution to this dilemma is not to abandon the schemas that lie at the
core of the lawyer’s professional expertise. After all, it was because of this exper-
tise that the client walked in the door. Optimal problem framing in lawyer-client
collaborations requires a “both/and” rather than an “either/or” approach—an
unfamiliar with the law and lacks a robust set of expert legal schemas, the client may in
fact misframe his problem. The client may dwell on material that he erroneously believes
to be pertinent to the legal frame. He may frame his problem in terms of an ineffective or
unavailable solution. He may resist his lawyer’s efforts to direct the conversation toward
subject matters that the lawyer knows are relevant to the proper framing of the legal
problem, or he may be reluctant to provide in sufficient detail information that the
lawyer knows is important.
42 problem solving, decision making, and professional judgment
approach that enables the lawyer-client team to view the problem from a variety
of different perspectives.
The idealized model of problem solving (set out in Chapter 1 and at the begin-
ning of this chapter) first defines the nature of the problem and then identifies
client objectives. But the problem definition is inextricably bound up with the
those objectives. Since problems often implicate a number of different objec-
tives, the best problem frame makes room for consideration of all of them.
In the following discussion, we use objectives and goals synonymously to refer
to relatively concrete aspects of the client’s or other stakeholders’ desired out-
comes. Interests are somewhat more general or abstract than objectives and
somewhat more concrete than values. These concepts are not separated by bright
lines, but lie on a continuum. It is useful to bear in mind that someone’s actual
experience of a solution as a “success” or a “failure” will be driven by both con-
crete and more intangible factors, ranging from monetary gains or losses on one
side of the spectrum to fidelity to philosophical, religious, or spiritual values on
the other.
objectives and interests unless the client does so first. But unless the lawyer
provides an opening for the consideration of these issues, the client is apt to
avoid them too, implicitly assuming that they are irrelevant or inappropriate.
Hidden or submerged objectives do not disappear simply because they were
never explicitly identified. But if they are not identified, they are less likely to be
achieved. Under these circumstances, the client may feel dissatisfied with an
outcome, even if it achieved all overtly specified goals.
Stakeholder analysis suggests that the client may himself comprise a variety
of different “stakeholders.” A client’s various “selves” may have differing, even
competing interests in a particular problem situation. Serrano’s social self, his
emotional self, his financial self, and his ethical/religious self may all have
“stakes,” or interests, in the Terra Nueva situation. These interests may compete
and even conflict with each other. For example, if Serrano were only concerned
with mitigating legal liability, his lawyer might take steps that would damage his
relationship with tenants, suppliers, regulators, and others. The conflicts may be
essentially intrapersonal as well as strategic. For example, strong empathy with
the tenants or a desire to apologize may compromise the possibility of an ulti-
mate victory in the courts. Likewise, a legal victory might leave Serrano wracked
with guilt or shame.
In summary, the optimal framing of a problem requires a thoroughgoing
identification of all the important objectives implicated by the situation. Here is
a checklist for going about this:
1. Construct a list of stakeholders and then view the problem situation from
each stakeholder’s perspective.
2. Identify the objectives pertaining to the client’s relationship with each
stakeholder.
3. Imagine the best outcome for each stakeholder and identify exactly what it
is about that outcome you would like to achieve.
4. Imagine the worst outcome for each stakeholder and identify exactly what
it is about that outcome you want to avoid. (People often find it difficult to
think about worst-case scenarios.10)
5. Watch out for “shallow” objectives, resulting from the confusion of means
with ends. Deepen these by asking, “Why?” until you cannot go back any
further (without becoming unconstructively existential or just annoying).
6. Allow for the surfacing of objectives that may be difficult to discuss in the
particular social context in which decision making is occurring. If the client
is an individual, consider the whole person, and make the lawyer-client
11. James March, Bounded Rationality, Ambiguity, and the Engineering of Choice, 9 Bell
Journal of Economics 598 (1978).
12. John W. Payne, James R. Bettman, and Eric J. Johnson, The Adaptive Decision
Maker 10 (New York: Cambridge University Press, 1993).
46 problem solving, decision making, and professional judgment
13. Charles H. Kepner and Benjamin B. Tregoe, The New Rational Manager
(2d ed. Princeton: Princeton Research Press, 1997). For a similar approach, see Dean K.
Gano, Apollo Root Cause Analysis (2d ed. Yakima, WA: Apollonian Publications, 2003).
framing problems and identifying causes 47
In any event, the article says that the UEW surveyed families in the new
buildings and that 40 percent have reported headaches, dizziness, or rashes
developing since they moved into their apartments. The union claims that
this is “overwhelming evidence” that the foam insulation is causing the sick-
ness. I should add that the article does not say whether the UEW had heard
complaints before it did the survey.
LAMM: By the way, didn’t I read in the documents we got from the mayor’s
office that all of the affected Terra Nueva residents live in units 5 and 6? What
can you tell me about those units?
YAMAMOTO: They were finished and people started moving into them a
year ago.
COOKE: I talked yesterday to one of Serrano’s people, who said that the manu-
facturer who supplied the insulation that they used for the older apartments
had raised his prices so much that the contractor bought the insulation for the
new units from another company. Apparently, it’s a completely new product.
Serrano was among the first developers in the county to use it. He told me
that it did have a sort of “chemical” smell when it was installed—lots of syn-
thetics do—but it wore off pretty quickly.
LAMM: Well, that’s interesting. Do you know if there are any other differences
in the materials used in the new apartments—for example, carpets or
paint?
COOKE: I’m pretty sure that there aren’t.
YAMAMOTO: Let me just say that if 40 percent of the Terra Nueva tenants are
sick, that’s a pretty strong argument in favor of there being a problem with
the insulation. I am also mindful that there have been similar reports of “foam
insulation syndrome” in other parts of the country.
COOKE: Well, hold on a minute, Marsha. My guess is that if you asked anyone
if they had had headaches, dizziness, or rashes at some point, 40 percent or
more would say yes. If there was something wrong with the foam insulation,
it should be making everyone in those units sick. And we don’t know anything
about the other units. I recall reading an article about people throughout
Europe getting sick from Perrier water after they learned that it contained a
tiny trace of some chemical. The company withdrew the so-called “contami-
nated” bottles, but the whole thing turned out to be mass hysteria. Clusters of
illnesses of this sort often turn out to be illusions.
In my view, there isn’t any solid evidence in the toxicology literature sup-
porting the theory that this particular product causes illness of this type.
Where I sit right now, particularly after doing some interviews last week, I
don’t think the problem is the foam insulation, or the union for that matter.
48 problem solving, decision making, and professional judgment
In my purely personal opinion, I wonder whether this whole thing may in fact
be a tactic by the tenants’ association. When I talked to Serrano last week, he
mentioned that, because his costs had risen, he raised rents at Terra Nueva a
couple of months ago. The new rents will go into effect the first of next year.
He told me that the tenants’ association presented tremendous opposition
to the rent hikes and that the association’s president is one of the named
plaintiffs in new lawsuit.
LAMM: Forty percent of the tenants may be a lot or a little. But we really don’t
know enough to come to any conclusions yet. We don’t even know whether
the apartments with the new product have the same insulation as the ones in
the reports of so-called “foam insulation syndrome”—whatever that is. These
are things we need to explore.
At this point, all we have is some different theories about what’s going on.
So let’s step back and walk through this systematically. I’m going to suggest
that we follow these steps:
1. Define the problem;
2. Specify the problem’s “what,” “where,” “when,” and “extent”;
3. Spot distinctions and identify changes responsible for the distinction;
4. Identify causal theories for further testing.
Is Is not
Where Units 5 & 6 (new units) [What about the other units?]
how many tenants were surveyed, or who they were, or how the survey was
conducted—let alone anything about the health of tenants in the other units.
YAMAMOTO: What difference does that make if this many people are sick?
LAMM: That’s a good question, Marsha. It’s easy to get fixated on where the
problem is. But you can’t really begin to unpack the causes of the problem
until you know where it isn’t. Suppose, just for the sake of argument, that lots
of the tenants in the old units, with older insulation, also report the symp-
toms. There could still be a problem. But it wouldn’t be the problem that the
tenants, the mayor, and now half the county thinks it is—the new insulation.
That’s why we need to learn more about how the tenants in the other units are
doing—and also to be sure we know what kinds of insulation their apart-
ments have.
Just to finish my chart, let’s talk about the when of the reports. We really
don’t know when any of the tenants first experienced these symptoms, but for
the moment, let’s say it’s a year ago—it doesn’t seem as if it could have been
any earlier than that. That will give us a baseline for some other “whens”: the
symptoms were reported after the new units were built, after United
Semiconductor employees moved in, and after the rent increase was announced.
We don’t know whether the first reports were made before the union organiz-
ing campaign, although it seems likely they were—otherwise, why would the
union have conducted the survey; still, we need to check this out.
50 problem solving, decision making, and professional judgment
2.5.3 Spot Distinctions and Identify Changes Responsible for the Distinction
LAMM (continued): Now let me add a column about anything that seems
distinctive about what is and is not and about anything else noteworthy:
Where Units 5 & 6 (new [What about (1) New insulation [though not
units) the other sure about whether some older
units?] units have the same insulation]
(2) plaintiffs live here
Extent 40% of those [60% did [How many people were surveyed?
surveyed in not report Who was surveyed? What about the
those units symptoms?] other units?]
With respect to the what, the reports of illness obviously are distinctive.
With respect to the where, the fact that the known complaints come from
people living in the new units is distinctive. The particular insulation used in
their apartments may be distinctive, but we really don’t know for sure. The
whens seem most interesting. The complaints only began after the new units
were built, after the United Semiconductor employees moved in, and after
the announcement of the rent increase. But it is at least noteworthy that,
though the rent increase affects all the tenants, the only people complaining
of illness are in the new units. We need to check out if the symptoms were
reported before the union-organizing campaign.
YAMAMOTO: If so, that eliminates the union as the cause of the whole mess.
framing problems and identifying causes 51
LAMM: Probably so—though we can’t discount the fact that the union may be
distorting or exaggerating the extent of the problem.
2.5.6 Problem
Toward the end of the discussion, Christine Lamm remarked, “There’s some-
thing bothering me that I can’t quite put my finger on.” It sounds as if she has
in mind yet another hypothesis for the sicknesses at Terra Nueva. What might it
be—and how would you go about testing it?
Terra Nueva and the examples given so far in this chapter demonstrate decision-
making strategies to use when something has gone wrong. But lawyers and
policy makers also have opportunities to be proactive in problem solving, to
move the world in a desired direction. Strategic planning—whether by state or
local governments, nonprofit organizations, or foundations—is a classic case of
forward-looking problem solving.
Here we provide an example of strategic planning by the William and Flora
Hewlett Foundation. Unlike governments, which often implement strategies
through their own agencies and personnel, grant-making foundations do so
through grants to nonprofit organizations. However, a foundation cannot know
what grants to make to what organizations until it has established its own goals
and determined strategies for achieving them.
Note to students: While the discussion of goals, barriers, outcomes, and targets fits
within the theme of Chapter 2, much of the rest of the materials concern the choice
among alternatives, and fits better with Chapter 4 or elsewhere. To present a coherent
story, we present all aspects of the strategic plan here rather than divide it among
several chapters.
Since its establishment in 1967, the Hewlett Foundation’s concerns have
included preserving the natural heritage of the American West. In 2009, the
Foundation’s Environment Program engaged in a strategic planning process for
this area of its grant-making.15 The Foundation was willing to devote approxi-
mately $125 million over five years to achieve its goals in the West, with the hope
that this would be matched by other donors interested in the region. The total
commitment, though fairly large by philanthropic standards, pales in compari-
son to the magnitude of the problems addressed. Hence the need for great clarity
and focus in goals and strategies.
Scope of work. The Foundation began by defining the geographic scope of its
work. Bounding its definition of the American West from “the uplift on the east-
ern edge of the Rocky Mountains to the Pacific coast” (as shown in Figure 2.2)
made it possible to establish clear goals for the region and analyze the barriers to
achieving them.
Goals. There are many possible goals related to protecting the natural heri-
tage of the American West, from preserving iconic landscapes to ensuring the
viability of farming and ranching communities. The Hewlett Foundation envi-
sioned “an ecologically vibrant West where the landscape is unspoiled and people
and wildlife thrive.” Since functioning natural systems underpin this vision, the
15. The planning was done by foundation staff members, led by Environment Program
Director Tom Steinbach, with the assistance of Ivan Barkhorn and staff from the Redstone
Strategy Group, LLC.
54 problem solving, decision making, and professional judgment
Anchorage
Anchorage
Calgary
Seattle
Denver
Los Angeles
Foundation made its ultimate goal ensuring the “ecological integrity” of the
West, meaning that natural systems would function close to the way they would
in the absence of human activity (see Figure 2.3).
Barriers. The next step in the Foundation’s strategic planning process was to
identify the barriers to achieving ecological integrity in the West. These barriers in
effect describe the problems the Foundation would need to solve to reach its goals.
The Foundation’s research and experience showed that current and emerging
human uses in the West, in addition to bringing economic benefits, can pose
Establish corridors
16. These targets were specified in much greater detail than appears in this overview.
56 problem solving, decision making, and professional judgment
into account that it is only one of many stakeholders concerned with the American
West. The frame through which the Foundation’s Environment Program views
the problems facing the West sometimes differs from that of state and federal
policy makers, businesses, citizens, interest groups, and other organizations.
A collaborative approach to strategic planning drew on the knowledge and
resources of these stakeholders and set the groundwork for future collaboration
with them.
Advocacy. Foundations and the nonprofit organizations they support have
various tools for achieving their goals, including providing direct services
(supporting public parks, for example), basic and applied research, and policy
advocacy. Advocacy, typically based on applied research, was a major strategy for
achieving the goals of the Foundation and its grantees in Western conserva-
tion.17 for example, improving land conservation requires policy changes that
create new protected areas and improve management and planning of existing
protected lands. The Foundation’s grantees, aware of the high cost of purchasing
land, found it important to advocate for federal and state funding and incentives
for private conservation.
Understanding the many stakeholders in the West is essential to building
constituencies supportive of Western conservation. Although other stakehold-
ers’ priorities do not necessarily overlap completely with the Foundation’s, the
Foundation was able to identify opportunities for collaboration with other
funders and interest groups. For example, hunters and anglers share the
Foundation’s interest in expanding land protections that help ensure both more
enjoyable sporting and greater ecological integrity.
Stakeholders include governments as well as private interests. After eight
years of stalled policy progress at the national level during the Bush administra-
tion, the political environment became more favorable to conservation in 2009.
The Foundation’s advocacy strategies appropriately changed to respond to
the Obama administration’s conservation-minded Secretaries of Agriculture,
Energy, and Interior, and to greater interest in conservation at local levels.
Comparing alternative courses of action. The next step in the Environment
Program’s strategic planning was to generate and compare different possible
courses of action to achieve the outcomes it desired. The Foundation’s analysis
of the problem led it to embrace a combination of broad West-wide policy
changes and work in specific regions, such as the Alberta Tar Sands, where the
extraction of oil is particularly damaging to the environment.
Within this broad approach, the Foundation sought the optimal way to maxi-
mize the impact of its grants in improving the ecological integrity of the West.
17. Federal law provides constraints on certain activities defined as “attempts to influ-
ence legislation,” but the kind of advocacy in which the foundation and its grantees
engaged is fully compatable with those constraints.
framing problems and identifying causes 57
Although we defer a full discussion of the concept of expected return until Part 3
of the book, we preview it here in order to complete this example of strategic
planning.
An expected return framework helps estimate the effectiveness of dollars
spent on particular grant-making strategies in achieving the Foundation’s goals.
Quantitative expected return analysis involved multiplying estimates of the
benefit an intervention would have if it succeeded by the likelihood that it will
succeed, and dividing by the costs of the intervention to the Foundation (see
Figure 2.4). The expected return for each promising project could thus be
compared, and the most cost-effective ones chosen.
Benefit is a monetizable measure of the progress a specific set of grants would
make toward the Foundation’s overall goal. In this case, the benefit was the
number of acres improved multiplied by what expert interviews and scientific
studies indicated would be the magnitude of improvement in ecological integrity
produced by the grant.
The likelihood of success is the estimated probability of a particular interven-
tion succeeding. Interviews with stakeholders and experts, and polls by the
League of Conservation Voters helped the Foundation calculate likelihoods of
success for particular policy changes in the various political jurisdictions encom-
passed by the West.
For the denominator, the Foundation considered the cost to the Foundation
of achieving the desired outcome. For a strategy to be cost-effective, the cost
must be less than the benefit times the likelihood of success—ideally a great deal
less. The higher the calculated expected return, the greater the Foundation’s
“bang for the buck.” The Foundation’s main costs were the amounts of its grants
and the staff resources needed to make and monitor the grants and facilitate
collaborations among stakeholders.
A rough comparison of the expected return of various strategies helped the
Foundation choose among different possible investments to achieve its conser-
vation goals. The expected return framework helped staff set clear targets and
make quantitative comparisons. Staff complemented the rough assumptions
necessary for the calculations with continual internal review and external feed-
back to select the combination of strategies that would most cost-effectively
achieve the Program’s desired outcomes.
Benefit Likelihood
(area-adjusted of success
Expected integrity improvement) × (%)
return =
Cost
($)
The Strategic Plan, or Logic Model. Figure 2.5 shows the core of the strategic
plan, sometimes called a logic model, which connects the specific activities the
Foundation supports to its outcomes and targets, and ultimately to the goal of
ensuring the ecological integrity of the West. As the preceding discussion
indicates, a strategic plan is developed by working backward from the general
goal to more specific outcomes and then to the activities necessary to produce
those outcomes. Of course, a strategy is only a working hypothesis that must
remain flexible to seize high-return opportunities and respond to unanticipated
threats to critical habitat or migration corridors. Although details of the strategy
were modified in the light of changing circumstances within a year of its adop-
tion, it continues to provide the framework for the Foundation’s work in the
American West.
Targets, grantmaking, monitoring, and evaluation. Foundation staff deter-
mined outcome metrics for each of the activities and outcomes—some of which
are measurable on an annual basis, others of which may take a number of years
to assess.
Having identified the activities necessary to work toward the goal, Foundation
staff then identify the organizations that can carry out the activities. Staff mem-
bers engage in due diligence, make grants, monitor organizations’ performance,
and evaluate progress toward their goals.
We will discuss the need for feedback in decision making later in the book.
For now, suffice it to say that feedback is vital to any decision maker’s long-term
effectiveness. Monitoring and evaluation will help the Foundation respond to
changing conditions and problems as they arise. The Foundation’s explicit met-
rics and targets are helpful in measuring the contribution of each grant to achiev-
ing its overall goal. A rigorous evaluation of the program as a whole will document
successes and lessons learned for use in the next round of strategic planning.
Returning to the overall topic of this chapter, the Hewlett Foundation’s stra-
tegic plan for the West provides an illustration of how analyzing the problems to
be solved, examining relevant interests, and comparing the consequences of dif-
ferent courses of action can guide proactive attempts to move the world in a
certain desired direction.
This page intentionally left blank
3. generating alternatives
Creativity in Legal and Policy Problem Solving
In Chapters 1 and 2, we explored the first three steps in a formal model of prob-
lem solving. These steps included problem framing, the specification and pri-
oritization of relevant interests and objectives, and the identification of major
uncertainties concerning the problem’s causes. We also saw in Chapter 1 that
not all problems occur in the present tense. Some will emerge in a future that we
can anticipate and plan for more or less effectively. With any problem, present or
future, the first three steps in a formal approach are followed by a fourth, involv-
ing the generation of potential solutions. In this chapter, we consider this fourth
step in more detail, first in the context of present problems, then in the context
of problems that may emerge in the future. Finally, we consider barriers to the
effective generation of alternative solutions and suggest some possible ways to
overcome them.
More than any other aspect of problem solving, the generation of alternatives
is associated in popular consciousness with the notion of creativity, which we
explore in this chapter. As will become clear, however, creativity is implicated at
almost every step of the problem-solving process. At its core, creativity is a stance.
It is about being a “chooser,” not merely a “picker.”
1. Barry Schwartz, The Paradox of Choice: Why More Is Less 75 (New York:
HarperCollins, 2004).
62 problem solving, decision making, and professional judgment
“Pickers,” on the other hand, are “relatively passive selectors from whatever
is available. With a world of choices rushing by like a music video,” Schwartz
observes, “all a picker can do is grab this or that and hope for the best.”2
As Schwartz’s description suggests, what separates choosers from pickers is
the ability to pause and reflect about a problem before taking action. Generally
speaking, people are uncomfortable with an unsolved problem and, as a result,
may reflexively seize upon the first plausible solution that comes to mind.
A chooser can forbear from this sort of reflexive “picking” long enough to think
expansively about the values, goals, and objectives implicated by a problem, a
process we considered in Chapter 2. She is able to do two other things described
in Chapter 2: she can view the problem from multiple perspectives, and she can
think systematically about its causes.
Choosers are able to resist the urge to jump at the first solution that comes to
mind. This is more difficult than it might seem. As we saw in Chapter 1, many
decisions result from the spontaneous application of problem schemas. Schemas
supply a standard problem frame, a set of apparently relevant interests and objec-
tives, and one or two stock solutions. In this sort of recognition-primed problem
solving, only those solutions that are schematically associated with the problem
are even considered. Over time, as a person responds to a particular type of prob-
lem with a stock solution, that response can become deeply habituated, even if it
fails to yield optimal solutions. And if this response pattern is common to a
community of experts—lawyers or administrative officials, for instance—the
response can come to be viewed as normative.
For these reasons, striking out beyond a preset solution to a schematized
problem can be unsettling. The same is true of novel problems, for which no
standard solution has yet been developed. In such situations, “pickers” may feel
confused and anxious, which may increase the tendency to seize at the first plau-
sible solution that presents itself.
Choosers, on the other hand, are able to cope with the indeterminacy accom-
panying an unsolved problem. In the empty space between problem recognition
and the selection of an alternative solution, a chooser can think more systemati-
cally, and ultimately decide more successfully, about how best to frame the prob-
lem, about the goals, values, and interests it implicates, about its causes, and
about its potential solutions. It is this last subject—the generation of alternative
courses of action—to which our attention now turns.
2. Id. at 224.
generating alternatives 63
3. Edward DeBono, Lateral Thinking: Creativity Step by Step (New York: Harper
& Row, 1970).
4. One note of caution is warranted here. Both the quality of decisions and decision
makers’ subjective sense of satisfaction with their decisions can be impaired by the desire
to “leave doors open,” that is, to avoid committing to a particular course of action. For a
discussion of this problem, see Jiwoong Shin and Dan Ariely, Keeping Doors Open: The
Effect of Unavailability on Incentives to Keep Options Viable, 50 Mgmt. Sci. 575 (2004);
Daniel T. Gilbert and Jane E. J. Ebert, Decisions and Revisions: The Affective Forecasting of
Changeable Outcomes, 82 J. Personality & Soc. Psychol. 503 (2002).
5. Thomas D. Barton, Conceiving the Lawyer as Creative Problem Solver, 34 Cal. W. L.
Rev. 267, 269 (1998).
64 problem solving, decision making, and professional judgment
6. See John Hammond, Ralph, Keeney, and Howard Raiffa, Smart Choices: A
Practical Guide to Making Better Decisions (Boston, MA: Harvard Business School
Press, 1999).
7. John Gardner, Grendel 159 (New York: Knopf, 1971).
8. Bruce Nussbaum, This is the IDEO Way, Business Week, May 17, 2008.
generating alternatives 65
solutions should look like, and whether they will pass the approval of
gatekeepers.
Think of some information-gathering alternatives that Luis Trujillo or
Christine Lamm might devise after being confronted with the problem at Terra
Nueva discussed in the previous chapters.
3.2.3.b Time-Buying Alternatives Especially when faced with a stress-inducing
problem, many of us are inclined to follow the adage, “Don’t just sit there, do
something!” and follow a “fire, ready, aim” process. But, some problems take
care of themselves over time. How often have you made a doctor’s appointment,
only to have the symptoms disappear before entering the waiting room? Even
those problems that are likely to persist may require time for analysis, informa-
tion gathering, consultation, and the generation and evaluation of an alternative
course of action. In these circumstances, consider first alternatives that will buy
you time—again, without (significantly) compromising ultimate solutions. What
might Luis Trujillo advise Jack Serrano to do in this respect in the face of public
controversy and the threat (or actuality) of suit?
One important way that people keep their options open over time is, well, to
buy options—not financial options (to purchase securities at a particular price),
but real options, where the underlying asset is not a financial instrument.
William Hamilton writes:9
An essential characteristic of all options is asymmetry in the distribution of
returns—greater upside potential than downside exposure. This results from
opportunities to terminate the investment or otherwise limit negative out-
comes while taking future actions to exploit positive outcomes fully . . . In
concept, real options are quite similar to financial options because they both
create the flexibility, but not the requirement for further action, as the future
unfolds.
If you have ever made a non-refundable deposit on any purchase, you have
purchased a real option.10
Cass Sunstein suggests that certain government policies can be understood to
create societal options against irreversible harms. Sunstein writes:
[W]here a decision problem is characterized by (1) uncertainty about future
costs and benefits of the alternatives, (2) prospects for resolving or reducing
the uncertainty with the passage of time, and (3) irreversibility of one or more
How did you go about trying to solve the problem? Most people try drawing
lines that pass through the dots and remain within the imaginary square they
constitute—and then they give up. They implicitly assume that the problem has
a path constraint that, in fact, it does not have—the rule, “Don’t go outside the
dots.” Who knows where we learned this tacit rule—perhaps from childhood
games of “connect the dots,” from the injunction to “color inside the lines,” or
from the fact that we are used to seeing objects in (more or less literal) frames.
In any event, it then becomes a limiting frame for solving the problem.15
For present purposes, we use the nine-dot problem to illustrate an important
point about the generation of alternative courses of action. It is often easier to
assume the presence of path constraints that limit a problem’s potential solutions
than it is to recognize their absence. Indeed, the very process of structuring a prob-
lem may create path constrains. We will say more about this later in the chapter.
14. We are not sure who created the Nine Dot Problem. We found it in James L.
Adams, Conceptual Blockbusting: A Guide to Better Ideas 24–25 (Reading MA: Basic
Books, 3rd ed., 1986).
15. You can find the “classic” solution (and several other wicked ones as well) in James
Adams’ book (supra) or on any number of Web sites.
16. U.S. v. McKinney, 919 F.2d 405, 421 (7th Cir. 1990) (Posner, J., concurring).
68 problem solving, decision making, and professional judgment
17. Carrie Menkel-Meadow, Aha? Is Creativity Possible in Legal Problem Solving and
Teachable in Legal Education?, 6 Harv. Negotiation L. Rev. 97–144, at 125 (2001).
18. This scenario is based on an analogous case litigated by one of the authors in
the early 1980s. Carson v. Atari, Inc., California Superior Court, County of Santa Clara,
No. 53073 (1982).
generating alternatives 69
What makes the solution effective is that it solves the problem. What makes the
solution surprising is that it differs from the stock alternative(s) spontaneously
triggered by the relevant problem schema. Most of the time, creative solutions
grow out of some established schema or combination of schemas. They involve
old things combined in new ways, or new things combined in old ways. A break
with the past makes the solution novel, but it is often continuity with the past
that makes it effective.
In this regard, notice that Luis Trujillo’s idea about a potential solution to the
Anaheim woodworkers’ problem did not exactly come out of “legal left field.”
The concept of “reasonable notice” that he proposed to introduce into employ-
ment law had deep roots in commercial law. This illustrates a number of impor-
tant points about how creativity operates in professional fields.
practice in both areas. For these lawyers, employment law and commercial law
represent discrete areas of specialization involving “habitually incompatible
frames of reference.” In such situations, the solution to a problem that might
seem intractable to an employment lawyer habituated to a doctrinal frame-
work in which “notice” has not been equated with “reasonable notice,” might
appear quite obvious to a lawyer habituated to doctrines in commercial law, in
which it has.
Many of the legal innovations we recognize as “creative” derived from this
kind of cross-fertilization between separate doctrinal or disciplinary domains.
Consider, for example, the warranty of habitability in landlord tenant law. The
concept of a “warranty” was originally unknown to the common law of real prop-
erty. In fact, the principle of caveat emptor stood in direct opposition to the notion
that an interest in real property might come with an implied warranty of any
kind.24 To solve a persistent social problem occasioned by informational asym-
metries between landlords and tenants, disparate levels of market sophistica-
tion, and other sources of bargaining inequality, lawyers eventually persuaded
courts to take the warranty concept from contract law and graft it onto landlord
tenant law.25
A similar process, through which a construct taken from one legal domain
is introduced to solve a problem in another, is illustrated by Charles Reich’s
24. See, e.g., Clifron v. Montague, 40 W. Va. 207 (1985); Walsh v. Schmidt, 206 Mass.
405 (1919); and Daly v. Wise, 132 N.Y. 306 (1892) (applying the principle of caveat
emptor).
25. See, e.g., Green v. Superior Court, 10 Cal. 3d 616 (1974); Javins v. First Nat’l Realty
Corp., 138 App. D.C. 369 (1970); Lemle v. Breedon, 51 Haw. 426 (1969) (abrogating the
doctrine of caveat emptor and imposing an implied warranty of habitability in landlord
tenant transactions).
72 problem solving, decision making, and professional judgment
landmark 1964 Yale Law Journal article, The New Property.26 In The New Property,
Reich urged that statutory entitlements to government benefits, such as Aid to
Families with Dependent Children or Medicaid, should be viewed as a species of
personal property, of which a person could not be deprived without due process
of law. Any first year law student who has read Goldberg v. Kelly in Civil Procedure
has seen the fruits of this innovation: Reich’s idea provided the basis for decades
of work protecting social welfare beneficiaries from the arbitrary, capricious, or
otherwise unfair denial or deprivation of public benefits.
Sometimes, a legal innovation can result from the cross-fertilization of
a particular legal subject matter area with an idea from an entirely different
discipline, such as economics. Consider in this regard the First Amendment
metaphor of the “marketplace of ideas,” emerging from Justice Oliver Wendell
Holmes’ dissent in Abrams v. United States. Although Holmes did not use this
phrase himself, Justice William Brennan later drew on the opinion to write:
[T]he ultimate good desired is better reached by free trade in ideas—that the
best test of truth is the power of the thought to get itself accepted in the
competition of the market . . . 27
Here, a principle drawing its persuasive appeal and normative force from the
domain of laissez faire economics was applied to the completely different domain
of free speech.
Law, public policy, and business contain many additional examples of innova-
tive solutions to seemingly intractable legal problems. Consider these, many
of which were self-consciously devised by lawyers, social advocates, and
entrepreneurs:28
• elimination of the privity requirement (that a manufacturer was only liable
to the purchaser of a product) in product liability;
• recovery of compensatory damages for emotional distress;
• the creation of a common law, and later of a constitutional “right to
privacy”;
• joint custody;
• same-sex marriage;
• condominiums, cooperatives, the corporation;
• pornography as a violation of civil rights;
• sexual harassment as a violation of sex discrimination laws;
• truth and reconciliation commissions;
• the joint venture;
• the “poison pill” as a response to a threatened hostile takeover; and
• the political action committee (PAC) and its successors.
26. Charles A. Reich, The New Property, 73 Yale L.J. 733, 785 (1964).
27. Abrams v. United States, 250 U.S. 616, 631 (Holmes, J. dissenting).
28. Menkel-Meadow, supra at 125.
generating alternatives 73
As Carrie Menkel-Meadow notes, many of these ideas have come about by the
creative reading or misreading of legal words and concepts—by “expanding,
aggregating, disaggregating, rearranging and altering existing ideas and con-
cepts, borrowing or translating ideas from one area of law to another or from
other disciplines, and . . . by use of redesign or architecture of words and con-
cepts to build both new legal theories at the abstract level and new institutions at
the practical level.”29
In making this observation, Menkel-Meadow’s view harmonizes with that of
psychologist Mihály Csikszentmihalyi, who suggests that creativity often involves
crossing the boundaries of different domains, combining elements of one with
elements of another to create something new. “An intellectual problem,” he sug-
gests, “is not restricted to a particular domain. Indeed, some of the most creative
breakthroughs occur when an idea that works well in one domain gets grafted
onto another and revitalizes it.”30 In legal problem solving, as in many other
domains, innovation entails a kind of conceptual blending, a process in which the
problem solver combines two or more organizing conceptual frames to create a
distinctly new frame that does not merely hybridize its constituent elements, but
also generates its own emergent structure.31
29. Id.
30. Mihaly Csikszentmihalyi, Creativity: Flow and the Psychology of Discovery
and Invention 88 (HarperPerennial 1996).
31. George H. Taylor, Law and Creativity, in Philosophy and American Law 81
(Francis J. Mootz III ed., New York: Cambridge University Press, 2009); Gilles Fauconnier
and Mark Turner, The Way We Think: Conceptual Blending and the Mind’s Hidden
Complexities (New York: Basic Books, 2002).
32. Id. at 255.
74 problem solving, decision making, and professional judgment
36. See generally Penelope Murray, Genius: The History of an Idea (Oxford, UK:
Blackwell, 1989).
37. Rudyard Kipling, Working Tools (1937), in The Creative Process: A Symposium
161–63 (Brewster Ghislin ed., Berkeley, CA: The University of California Press, 1985).
38. William James, Talks to Teachers on Psychology 64 (New York: Henry Holt,
1908).
39. William James, Great Men, Great Thoughts, and the Environment, 46 Atlantic
Monthly 441–59 at 456 (1880).
40. See, for example, Roger von Oech, A Whack on the Side of the Head: How
You Can Be More Creative (New York: Warner Books, 1983); Paul Sloane, The
Leader’s Guide to Lateral Thinking Skills: Unlocking the Creativity and
Innovation in You and Your Team (2d ed. Philadelphia, PA: Kogan Page, 2006).
76 problem solving, decision making, and professional judgment
creative work in an area unless they really love what they are doing and focus on
the work rather than the potential rewards.”47
Individual traits. Sternberg notes that some people find it easier than others
“to transit back and forth between conventional and unconventional thinking.”48
Certain personality traits may be positively correlated with creative production.49
These include a preference for moderate risk-seeking, willingness to tolerate
ambiguity, persistence, and self-efficacy, as well as an attraction to complexity
and aesthetic orientation. Bucking the crowd to “break script” requires self-
confidence, independence, and persistence, as well as communicative skill.
of Creativity Seen Through the Lives of Freud, Einstein, Picasso, Stravinsky, Eliot,
Graham, and Gandhi (New York: Basic, 1993).
47. Robert J. Sternberg, Wisdom, Intelligence, and Creativity Synthesized,
108 (New York: Cambridge University Press, 2003).
48. Id. at xv. By the same token, Joy Paul Guilford posits that what we define as “think-
ing” comprises five different mental operations (cognition, memory, divergent thinking,
convergent thinking, and evaluation), which act on four categories of content (figurative,
symbolic, semantic, behavioral) to produce intellectual products. Creativity depends on
cognitive flexibility—the thinker’s ability to break apart different configurations of preex-
isting products and combine them in original ways. See generally Joy Paul Guilford, The
Nature of Human Intelligence (New York: McGraw-Hill, 1969).
49. For a summary of the relevant research, see Timothy I. Lubart, Creativity, in
Thinking and Problem Solving 290–332 (Robert J. Sternberg ed., 1994); Robert J.
Sternberg and Timothy I. Lubart, Defying the Crowd: Cultivating Creativity in
a Culture of Conformity (New York: Free Press, 1995).
50. Howard Gardner, Intelligence Reframed: Multiple Intelligences for the
21st Century 116-17 (New York: Basic Books, 1999).
51. Emma Policastro and Howard Gardner, From Case Studies to Robust Generalizations:
An Approach to the Study of Creativity, in Handbook of Creativity, supra at 213–25.
78 problem solving, decision making, and professional judgment
52. For more on this point, see Robert J. Sternberg, Wisdom, Intelligence, and
Creativity Synthesized (2003), supra at 21, 101–09.
53. Robert J. Sternberg, A Propulsion Model of Types of Creative Contributions, 3 Review
of General Psychology, 83–100 (1999); Robert J. Sternberg, James C. Kaufman, and
Jean E. Pretz, The Creativity Conundrum: A Propulsion Model of Kinds of
Creative Contributions (New York: Psychology Press, 2002).
54. Robert J. Sternberg, Wisdom, Intelligence, and Creativity Synthesised,
supraat 138.
55. Joseph L. Sax, The Public Trust Doctrine in Natural Resource Law: Effective Judicial
Intervention, 68 Mich. L. Rev. 471 (1970); Joseph Sax, Takings and the Police Power, 74
Yale L. J. 36 (1964).
generating alternatives 79
56. Joseph L. Sax, Takings, Private Property, and Public Rights, 81 Yale L. J. 149, 151
(1971).
57. Sindell v. Abbott Laboratories, 163 Cal. Rptr. 132, 607 P.2d 924 (1980).
58. Brown v. Superior Court, 751 P.2d 470 (Cal. 1988); Hymowitz v. Eli Lilly and Co.,
539 N.E.2d 1069 (N.Y. 1989).
80 problem solving, decision making, and professional judgment
constitute the field. Once the new ideas are accepted, the domain preserves and
transmits them to other individuals.59
These relationships are illustrated by Figure 3.3.60
Culture
Domain
Selects Transmits
Novelty Information
Produces
Novelty
Field Individual
Stimulates
SOCIETY Novelty PERSONAL
BACKGROUND
So far, we have focused on solving problems that occur pretty much in the
present. But individuals, corporations, nonprofit organizations, and govern-
ments are often called upon to act today to solve problems that lie in the future.
As in the case of our farmer’s estate planning problem in Chapter 2, the future
59. Mihály Csikszentmihalyi, Society, Culture, and Person: A Systems View of Creativity
(1988), in The Nature of Creativity 325–39 (Robert J. Sternberg ed., New York:
Cambridge University Press , 1988).
60. Figure 3.3 is from Mihály Csikszentmihalyi, Implications of a Systems Perspective for
the Study of Creativity, in Handbook of Creativity, supra at 13–35, 315.
generating alternatives 81
61. Much of our discussion is based on the work of the Global Business Network
(GBN), a member of the Monitor Group.
62. Diana Scearce and Katherine Fulton, What If? The Art of Scenario
Thinking for Nonprofits 11 (Global Business Network, 2004).
63. Erik Smith, Chapter 5: Using a Scenario Approach: From Business to Regional Futures,
in Engaging the Future: Forecasts, Scenarios, Plans and Projects 85 (Cambridge:
Lincoln Institute of Land Policy, 2007).
82 problem solving, decision making, and professional judgment
Positive
External influences
cripples the Valley. Difficulty attracting Ethnic and class divisions deepen,
“clean” industry and talented workers stunts threatening the region. Increased poverty
regional growth and limits its attractiveness. among the “have-nots” leads to decreased
The public education system crumbles, and investment. Local tax revenues decline as its
civil unrest flares. State funding dwindles. residents migrate elsewhere. Nevertheless, a
Air and water quality worsen. Latino middle class slowly emerges.
Worsening
racial and economic divisions, with the goal of catalyzing regional change.69
Although it is too early to discern the effects of the exercise, GVC was sufficiently
optimistic to launch a new scenario initiative, “The Great Valley in 2020,” includ-
ing publication of a workbook for use by nonprofits and other organizations to
guide their scenario-planning efforts.70
Scenarios were also employed in a 2002 strategic planning effort by Casey
Family Programs, a nonprofit organization dedicated to providing and improv-
ing foster care. The organization was struggling to clarify its strategy and mis-
sion. A then-rising stock market had dramatically increased Casey’s endowment,
inspiring the organization to consider what success for the organization would
look like ten years down the road. Two critical uncertainties were considered:
• “How will U.S. society and government evolve over the next 10 years? Will
citizens become more/less actively involved in social issues and supportive
of an interventionist government?
• What role will technology play in fostering community?”7172
Scenario workshops yielded the results shown in Figure 3.5.
Interventionist
government
community
community
Promotes
Role of technology
Minimalist
government
As it turned out, key aspects of the Big Mother scenario did in fact unfold. This
scenario encompassed international geopolitical crises, increasing nationalism,
and an economic downturn—none of which was generally expected in 2002.
Chiemi Davis, Casey’s senior director of strategic planning and advocacy, reports
that “because of the scenario work . . . we now consider ourselves much better
prepared for the current economic and cultural realities.”73 GBN notes that the
Casey Foundation “saw the potential threat in a bleak scenario, like Big Mother,
and prepared accordingly.” 74
73. Id.
74. Id.
75. Id. 7.
76. George A. Goens, Beyond Data: The World of Scenario Planning. Constructing Plausible
Plots in a World that Defies Linear Logic, April 2001, http://www.aasa.org/publications/
saarticledetail.cfm?ItemNumber=3761.
77. Smith 83.
generating alternatives 85
• Thinking from the outside in: Rather than focusing on factors under our
control, scenario planning explores how events beyond our control might
affect the future.
• Including multiple perspectives: Scenario planning helps overcome myopic
visions of the future, and forces policy makers to think more openly about
what the future might bring.
In developing a scenario, planners must determine an appropriate time frame
based on when the most important effects of the decision will play out. The
relevant horizon for policy decisions may be as long as several decades. Planners
must also identify key future uncertainties, including external drivers—a process
that can be aided by considering emerging social, technical, economic, environ-
mental, and political (STEEP) trends.
After prioritizing key uncertainties, participants can begin to brainstorm future
scenarios. This is conceptually challenging. An almost infinite variety of scenarios
are possible—after all, almost anything could happen in the future. To remain
useful, the exercise must be bounded. A useful set of scenarios generally will
have quite different implications for decision making (though it would be note-
worthy and important if radically different scenarios led to identical decisions for
an organization or polity). In selecting scenarios for consideration, planners seek
out scenarios that are:
• Plausible. The scenario must be believable.
• Relevant to the key strategic issues and decisions at hand. If the scenario
would not cause a decision maker to act differently compared to another
scenario, there is little use in considering it.
• Challenging to today’s conventional wisdom. It should make one think
about different possibilities and options.
• Divergent from each other. Together, the scenarios should “stretch” the
thinking about the future environment, so that the decisions take account
of a wider range of issues.
• Balanced. It is useful to ensure that a group of scenarios strike a good
psychological balance between challenges and opportunities, risks, and
upside potential.78
Once a compelling set of scenarios is identified, it is useful to identify “lead-
ing indicators” to serve as warning signs that a given scenario is unfolding.
mislead participants about the probability that a particular scenario will occur.79
Scenario planners work to weave together vivid storylines; their storytelling often
entails “unpacking” a scenario into its component parts to illustrate how a given
set of events could converge to result in such an outcome. But vivid stories about
how a particular event might unfold make its occurrence seem more likely than
it may actually be.80 And the unpacking process can contribute to a logical error
in which participants’ assessments of the probability of events that are mutually
exclusive and collectively exhaustive add up to more than 100 percent.81 Scenario-
thinking thus may impair predictive accuracy by inspiring forecasters to “attach
excessive probabilities to too many possibilities.”82
Earlier in this chapter, we used the nine-dot problem as a way to illustrate one
common obstacle to the generation problem solutions—imagining nonexistent
path constraints.
Most people are not able to solve the nine-dot problem at all. By contrast, we
do come up with solutions to most of our personal and professional problems—
though we have a tendency to adopt the first plausible alternative that comes to
mind, so long as it seems “good enough.” Such “satisficing” often makes sense:
We have many little problems to solve in the course of a day. However, even
when a particular situation warrants giving more thought to possible solutions,
we tend not to systematically consider a range of alternatives. We are constrained
by cognitive, social and emotional, practical, and personal factors. We consider a
number of these here.
79. See Philip E. Tetlock, Expert Political Judgment: How Good Is It? How Can
We Know? The Limits of Open Mindedness 194, 196 (Princeton and Oxford: Princeton
University Press, 2005); Section 9.6, infra.
80. Daniel Reisberg, Cognition: Exploring the Science of the Mind 386–87
(New York: W.W. Norton Press, 2001).
81. Id. 197.
82. Id. 201.
generating alternatives 87
4''
Imagine that a steel pipe is imbedded in the concrete floor of a bare room. The
inside diameter of the tube is .06 inches larger than the diameter of a ping pong
ball, and, in fact, a ping pong ball is resting on the concrete floor at the bottom of
the pipe. You are one of a group of six people in the room, in which you also find
a carpenter’s hammer, a chisel, a box of cereal, a file, a wire coat hanger, a monkey
wrench, a light bulb, and one hundred feet of clothesline. You must find a way to
get the ping pong ball out of the pipe without damaging the ball, the tube, or the
floor.83
Despite its obvious effectiveness, Adams observes, few people presented with
this problem propose to solve it by having the people present pee into the pipe,
thus floating the ping pong ball to the top of the tube where it can easily
be retrieved. Few if any people would even think of this solution, let alone
propose it, because, as Adams notes, urination is a private activity, at least in our
society.84
Though somewhat atypical, Adams’ scenario illustrates another class of
impediments to the generation of a robust set of alternative solutions—social
conformity effects. We don’t want to make fools of ourselves or offend the sensi-
bilities of others in our social environment—and proposing untested ideas about
alternative approaches to a problem can be risky. Social norms tend to develop
around recurring problems, providing a relatively small set of “acceptable” can-
didate solutions. Fearing disapproval, or perhaps not wanting to offend some-
one—often a person in authority—who has already proposed a plausible, more
“traditional” solution, may keep us from even imagining, let alone suggesting a
new approach. Indeed, pursuing solutions that violate social norms may be
perceived as somehow “cheating,” even if the solution does not violate any
explicit rules applicable to the problem itself.
We do not know whether lawyers and policy makers are more or less suscep-
tible to social conformity constraints than anyone else. But lawyers are often
viewed—by themselves as well as by others—as conservative, risk-averse, and
precedent-bound. Deviation from the “tried and true” can easily slip into devia-
tion from the applicable “standard of practice.” Such deviations may, at least
under some circumstances, subject a lawyer not only to ridicule, but also to
malpractice liability or bar discipline. In like fashion, the public accountability of
policy makers and the constrained routines of bureaucrats often promote risk-
averse behavior and inhibit taking creative approaches to solving problems.
While some professional school courses—probably more in public policy
than in law—encourage students to be creative problem solvers, much of the
core curriculum trains students to be more comfortable criticizing ideas than
creating them. The tendency to criticize ideas prematurely inhibits the genera-
tion of a rich and varied array of potential solutions or alternative courses of
action. Risk aversion, fear of breaking the rules or making a mistake, and
an implicit belief that there is one right answer to every problem all combine
with our general discomfort with an unsolved problem to tend us toward quick
convergence on one, or at most a small set, of stock, safe solutions.
84. By the way, based on informal empirical observations, men are not more likely to
solve this problem than women.
generating alternatives 89
It also remains unclear whether the various techniques for enhancing creativity
hawked by any number of “how-to” books can systematically be employed to
improve the problem-solving abilities of lawyers and policy makers. But in sum-
mary, we offer a number of helpful ideas, gleaned from the trade literature and
the work of academic researchers.
Perhaps most significantly, if necessity is the mother of invention, overwork
makes for effective birth control, creating the ideational equivalent of China’s
“one child” policy. The generation of a broad range of potential solutions to any
given problem appears to benefit from:
• time to ruminate and let problems percolate in the unconscious mind;
• permission to sit for a while with an unsolved problem. In the quest for
creativity, idleness may prove a virtue.
• awareness of the constraints that schematic expectancies and other pre-
scripted responses implicitly impose on our thinking;
• the ability to go beyond the frames and schemas automatically triggered by
a problem of a particular sort; and
• willingness to at least temporarily eschew evaluation, reserve judgment,
and postpone evaluative scrutiny of potential alternative solutions.
Here are a few practical guidelines for applying these general principles to
specific problem-solving situations:
1. Focus individually on each of the objectives identified earlier in the
problem-solving process and, with respect to each one, ask, “How could it
be achieved?”
2. Generate at least three ideas for each identified objective before moving on
to the next.
3. Question the constraints you have imposed on possible solutions to the
problem.
4. Think about analogous problems in other fields or disciplines and ask
whether there is some device from those fields or disciplines that might be
adapted to solve the problem.
5. Consult with people with knowledge sets other than your own to get their
ideas about potential solutions.
6. Generate a number of potential solutions before critiquing or choosing any
one of them.
We close the chapter with a problem, which you can use to stretch your think-
ing about the topics we have covered here.
Luis Trujillo and his client, Jack Serrano, do not yet have enough information
to have any confidence in the causes of the problems at Terra Nueva. Yet, they
need to act quickly. If the newspaper report is accurate, a class action suit has
already been filed against Serrano. Remember that defending the suit is not
necessarily Serrano’s only interest. He may also be concerned with his reputa-
tion; with his relationships with government regulators, financial backers, pres-
ent and potential business partners, current tenants, and prospective tenants;
with dealing with his and his family’s psychological stress; and with “doing
right” by anyone who has actually been injured by materials used in the homes.
Generate a list of actions that Serrano might take within the next days or
weeks to address these or any other interests you think he might have.
4. choosing among alternatives
Having thought through a variety of alternative strategies that may achieve your
objectives, it’s time to make the decision. Sometimes, the process up to this
point will lead inexorably to a single choice. Often, however, you may be faced
with competing interests or objectives or competing strategies for achieving
them—as Jack Serrano is, for example, in trying to decide how to respond to his
tenants’ complaints.
The first step toward choosing among several plausible alternatives is to be clear
about the consequences of choosing each one. The alternatives facing Serrano
are fraught with uncertainties, including how others will react to any actions he
takes. Before tackling such complexities, consider how you might assess the
consequences of alternatives in a simpler case—perhaps one closer to your
own experience. Suppose that you are considering renting an apartment, and
that you have four major criteria:
• Cost. The lower the rent, the better.
• Size. You need at least two rooms, and three (one of which would serve as
a guest room) would be ideal.
• Proximity to the campus. The closer the better.
• Neighborhood. You’d like to live in a safe, attractive neighborhood.
You have now seen a number of apartments that seem plausible candidates.
How might you describe them in terms of these criteria?
Cost is the simplest and most readily quantifiable. When you think about
size, you realize it’s not just the number of rooms, but their dimensions. The
guest room is really tiny—all right for an occasional out-of-town friend, but your
mother might give you a hard time when she visits. (On the other hand, that may
keep her visits reasonably short.)
With respect to proximity, the apartment is not really within walking distance
of the campus, but there’s a bus stop just in front of your door; also it seems like
a fairly short bike ride. But does the bus go directly to campus, or do you have to
transfer? Would biking require you to ride on a busy commercial street with no
bike lanes? The neighborhood looks pretty nice during the daytime, but you
haven’t been there at night when the jazz nightclub across the street is open.
Also, now that you think about it, you haven’t seen a supermarket anywhere
nearby.
92 problem solving, decision making, and professional judgment
This little scenario suggests a number of general guidelines for specifying the
consequences of a potential decision.1
• Describe the consequences of the highest-ranked alternatives in terms of
each of your interests or objectives.
• Wherever possible, describe the consequences quantitatively; this will facil-
itate the comparison with other (existing or possible) alternatives. Make
sure that the quantity captures all the dimensions of an interest (e.g., size
of the rooms as well as number; commute time to campus as well as
distance.)
• Don’t disregard consequences that can’t be quantified, such as the quality
of the neighborhood or the safety of the bike ride. As Einstein famously
said: “Not everything that counts can be counted, and not everything that
can be counted counts.”
• Gather data about the consequences. If you’ve only seen the neighbor-
hood during the daytime or on weekends, go there on an evening or
weekday.
• Pretend that you have already chosen an alternative, and imagine a day
in your life from morning to night. Engage in such “mental simulation”
with a number of alternatives, not just your favorite. Otherwise, you may
become too committed to one choice and have difficulty imagining the
others.
• Be open to new interests or objectives coming to mind. Even when you
follow a deliberative model, the processes are recursive, with subsequent
steps reminding you of lacunae in earlier ones.
After engaging in the process with all plausible alternatives, one of them
may so obviously dominate others—that is, be better in some respects and at
least as good in all others—that the decision is inescapable. (Also, an alterna-
tive may be obviously dominated by others, so you can eliminate it from consid-
eration.) The interesting decisions, though, are ones where there is no one
obvious choice.
Note that this little exercise makes at least two implicit simplifying assump-
tions: (1) that there are no significant risks or uncertainties in how the future will
unfold, and (2) that you are pretty good at predicting how you will feel about the
decision once you actually begin living with it. In fact, real-world decision making
tends to be more complex in both of these respects. But we’ll leave that for part
3 of the book.
1. We are indebted to John Hammond, Ralph, Keeney, and Howard Raiffa, Smart
Choices: A Practical Guide to Making Better Decisions (1999) for some of the
guidelines.
choosing among alternatives 93
odor might evoke images or feelings of illness. Whether or not this is generally
accurate, it would be best to check it out with some conscious System II analysis
before making an important decision based on the linkage.
Routines. Routines are patterns of behavior that one follows regularly and
without thought: for example, swallowing vitamin pills with one’s morning
coffee or taking a particular route to work. Most bureaucracies and other organi-
zations depend on routinized ways of doing things—new employees are taught
to do it that way because “that’s the way it is done.” A routine may be the result
of past deliberation, but present actors do not necessarily advert to its rationale
and may not even know what it is.
Rules. A rule specifies the outcome of a decision given the existence of certain
facts. Rules have low decision-making costs and constrain discretion. Many gov-
ernment regulations take the form of rules: “Maximum speed limit, 65 miles per
hour” and “No bicycle riding in the playground.” Rules typically are both over-
and under-inclusive and therefore entail some individually suboptimal deci-
sions—traffic conditions may be such on a particular stretch of highway that
75 miles per hour is perfectly safe, but don’t waste your time arguing this to the
highway patrol officer or the judge. Self-imposed rules—for example, “never
have second helpings”—are often designed to avoid succumbing to temptations.
Even though a rule does not call for deliberation by the immediate decision
maker, it does require that someone has deliberated in advance to create
the rule.
Checklists
Checklists, of the sort used by airplane pilots, are similar to rules or presump-
tions. In 1935, the so-called “Flying Fortress” often crashed on take-off until the
military realized that this four-engine aircraft was so complex that pilots needed
to rely on a checklist rather than their intuition. Recent studies suggest that using
a checklist can make a huge difference in critical medical care. For example, by
checking off five simple steps in putting catheters into patients—from washing
hands with soap to putting a sterile dressing over the site—physicians and nurses
could dramatically reduce the number of infections. Atul Gawande, Annals of
Medicine: The Checklist, NEW YORKER, December 10, 2007.
your host. A common example from law practice is: “Never ask a witness a ques-
tion on cross-examination unless you know the answer yourself.” An environ-
mental regulation might prohibit discharging more than a certain amount of a
pollutant unless the business can prove that it isn’t feasible to keep the discharge
below the limit.
Presumptions of this sort permit quick, nondiscretionary decisions. They can
also allow for a gradual learning curve. People new to an organization or field of
In the course of researching this chapter, we came across these examples from a
book on business rules of thumb, modestly subtitled, Words To Live By And Learn
From That Touch Every Aspect Of Your Career:
• Throw a tantrum within the first ten minutes of the negotiating (from a
sports consultant).
• Never hire a salesman you’d want your daughter to marry.
• Don’t order drinks served with a paper umbrella.
SETH GODIN AND CHIP CONLEY, BUSINESS RULES OF THUMB (Warner Books 1987).
practice may adhere to them strictly and loosen the constraints as they develop
more experience.
Like a rule, a presumption may be the result of someone’s deliberation before-
hand. Unlike a rule, a presumption calls for a degree of deliberation at the time
of its application—enough deliberation to determine whether the circumstances
justify rebutting the presumption.
Standards. A standard leaves it to the individual decision maker and those
who will review her judgment to apply a norm such as “reasonableness,” “exces-
siveness,” or “competency” to a particular situation. For example, in addition to
setting particular speed limits, California law provides that “No person shall
drive a vehicle upon a highway at a speed greater than is reasonable or prudent
having due regard for weather, visibility, the traffic on, and the surface and width
of, the highway.”6 Depending on whether the norms of reasonableness, etc., are
well defined by practice and precedent, a standard may require more or less
deliberation.
Small steps. A decision maker may proceed by taking small, reversible steps
and observing their consequences, rather than making a big decision in one fell
swoop. This is akin to the use of time-buying options, discussed in Section 3.2.3.
6. People v. Huffman, 88 Cal. App. 4th Supp. 1, 106 Cal. Rptr.2d 820 (2000).
96 problem solving, decision making, and professional judgment
Sunstein and Ullman-Margalit give the examples of Jane living with Robert
before deciding whether to marry him, and observing of the Anglo-American
common law tradition of deciding cases on as narrow a ground as possible.7
Delegation. Rather than make the decision yourself, you can delegate the
responsibility to another individual or agency. A client may ask her lawyer to
make a decision on her behalf; an administrative official may be required to
delegate a decision to an agency.
Random processes. In addition to providing a fair way to resolve some inter-
personal disputes, a random process like flipping a coin can quickly resolve
intrapersonal disputes when you are faced with two equally attractive options
and the choice doesn’t justify further analysis.
To understand how EBA works and see its limitations, let’s consider the hous-
ing example in greater detail. Suppose that you are considering thirteen rental
apartments, and that you rank the four major attributes in this order: cost, size,
proximity to the campus, and neighborhood, with cost being the most impor-
tant. You then determine a specific acceptability threshold for each attribute. For
example, you cannot spend more than $1500 a month, and you need at least
600 square feet of space. Under EBA, you eliminate, say, eight apartments that
have monthly rents greater than $1500. Of the five apartments remaining, you
eliminate two studios with 575 square feet of floor space.
Each of the three remaining apartments costs about $1300 a month and has
about 800 square feet of space. One is located in the more cosmopolitan down-
town center—an eight-minute commute to campus; the two others are virtually
adjacent to one another in a smaller, more suburban neighborhood north of the
town—also an eight-minute commute to campus. Though the atmospheres and
amenities of the two neighborhoods differ, each is pleasant, safe, and attractive
in its own way.
EBA helped whittle thirteen choices down to three, all of which are satisfac-
tory. But notice that the procedure can lead to a suboptimal outcome if the thresh-
olds for the first several attributes were based on preferences rather than absolute
necessities. Perhaps you eliminated an apartment that was a bit too expensive
but that was so great in all other respects that it would have justified stretching
your budget. This is an inherent limitation of any noncompensatory procedure.
Along similar lines, EBA may allow others to manipulate your rankings for
their own interests. Amos Tversky provides this parodic example from an actual
television commercial:
“There are more than two dozen companies in the San Francisco area which
offer training in computer programming.” The announcer puts some two
dozen eggs and one walnut on the table to represent the alternatives, and
continues: “Let us examine the facts. How many of these schools have on-line
computer facilities for training?” The announcer removes several eggs. “How
many of these schools have placement services that would help you find a
job?” The announcer removes some more eggs. “How many of these schools
are approved for veteran’s benefits?” This continues until the walnut alone
remains. The announcer cracks the nutshell, which reveals the name of the
company and concludes: “That’s all you need to know in a nutshell.”8
Suppose that you can attend the classes in person and would prefer to do so,
that you are seeking training to advance in your current job, that you are not
a veteran, etc. Perhaps the best instructors teach at places that lack all of the
attributes touted by the commercial. In such a case, the EBA methodology out-
lined above would lead you far astray from your ideal outcome. EBA is only
Gerd Gigerenzer and his colleagues argue that, in making factual judgments,
people use “fast and frugal heuristics” similar to EBA. For example, in estimating
the relative size of two cities, if we recognize the name of one but not the other,
we will employ the “recognition heuristic” to conclude that if we recognize one
city but not another, the recognized city is larger. Gigerenzer asserts that, in gen-
eral, we tend to make judgments based on a single cue and only look to another
as a tie-breaker if it doesn’t produce a result. More generally, under the take the
best heuristic, the decision maker tries cues in order, one at a time, searching for
a cue that discriminates between the two objects in question. For example, “when
inferring two professors’ salaries, the [professorial] rank cue might be tried first.
If both professors are of the same rank (say both associate professors), then the
gender cue might be tried. If one of the professors is a woman and the other is a
man, then we might say that the gender cue ‘discriminates.’ Once a discriminat-
ing cue is found, it serves as the basis for an inference, and all other cues are
ignored. For instance, if gender discriminates between two professors, the infer-
ence is made that the male earns a higher salary, and no other information about
years of experience or highest degree earned is considered.”9
There is little doubt that people, in fact, often use simplifying decision proce-
dures in making factual judgments as well as decisions. But there is serious
controversy whether we employ only one cue at a time and, more fundamentally,
whether fast and frugal heuristics tend to produce accurate judgments.10
9. Gerd Gigerenzer, Peter M. Todd, and the ABC Research Group, Simple
Heuristics that Make Us Smart 99 (New York: Oxford University Press, 1999).
10. See, e.g., Daniel M. Oppenheimer, Not So Fast! (and Not So Frugal!): Rethinking the
Recognition Heuristic, 90 Cognition B1 (2003); Mark G. Kelman, The Heuristics
Debate: Its Nature and Its Implications for Law and Policy (Oxford University Press,
forthcoming 2010).
choosing among alternatives 99
position on this issue but who hold positions that you agree with on many
other issues. Also, a noncompensatory procedure does not account well for
your intensity of preferences. Unless gun control matters to you more than all
other issues together, you may not be choosing the candidate who best reflects
your interests.
In fact, most decisions invite making trade-offs among competing interests
or objectives. For example, you might trade off the candidate’s position on gun
control with her position on other social and economic issues that you deem
important. In choosing the apartment, you might be willing to trade off the
rental price for proximity. One can approach making trade-offs by (1) giving
reasons for one’s choices or (2) assigning quantitative values to different attri-
butes of the things being chosen.
You are about to take the bar examination. If you pass the exam, you will celebrate
with a week’s vacation in Hawaii. If you fail the exam, you will console yourself
with a week’s vacation in Hawaii. Your travel agent says that you can lock in low
airline fares now, and that the fares are sure to rise if you wait. Will you book the
vacation now, or wait until you receive the results of the exam? In a study by Eldar
Shafir, Itamar Simonson, and Amos Tversky, most students chose to pay the
additional cost and wait. They wanted to choose the vacation for the “right”
reason, which couldn’t be known in advance.13
12. Note, by the way, the role of images, scenarios, and affect in Darwin’s list-making.
Recall Antonio Damasio’s description of the role of affect in decision making in Section
1.5.3. As it turned out, Darwin married his cousin, Emma Wedgewood, in 1839, and
they had ten children, many of them notable in their own right. Twenty years later, he
published On the Origin of Species by Means of Natural Selection.
13. Eldar Shafir, Itamar Simonson, and Amos Tversky, Reason-Based Choice, 49
Cognition (1993).
14. Benjamin Franklin to Joseph Priestly, September 19, 1772. At the time, Priestly
was happily preaching and conducting scientific experiments. When offered the position
of librarian to Lord Shelburne, he asked Franklin for advice. Rather than advise on the
merits, Franklin offered a method of decision making.
choosing among alternatives 101
of the different motives, that at different times occur to me, for or against the
measure.
When I have thus got them all together in one view, I endeavor to estimate
their respective weights; and where I find two, one on each side, that seem
equal, I strike them both out. If I find a reason pro equal to two reasons con,
I strike out the three. If I judge some two reasons con, equal to some three
reasons pro, I strike out the five; and thus proceeding I find at length where
the balance lies; and if, after a day or two of further consideration, nothing
new that is of importance occurs on either side, I come to a determination
accordingly.
And, though the weight of reasons cannot be taken with the precision of
algebraic quantities, yet when each is thus considered, separately and com-
paratively, and the whole lies before me, I think I can judge better, and am
less liable to make a rash step, and in fact I have found great advantage from
this kind of equation, in what may be called moral or prudential algebra.
Franklin’s procedure of canceling the pros and cons of a decision is a rough
approximation of quantifying the strength and importance of the attributes
involved in the decision. Franklin seemed to have in mind a yes-no decision
about whether to take a particular action. But you could extend his procedure to
deciding among multiple alternatives, such as the choice among apartments.
If you are choosing between two apartments, write down the pros and cons of
each and do the cancellation. If you are choosing among more than two, you
might begin with what appear to be relatively strong and weak candidates, then
see if you can eliminate the weak one, and put another in its place.15
Almost surely, Darwin and Franklin did not think their processes would
determine their decisions rather than inform their intuitions. But this is by
no means a useless project. Indeed, as we suggest below, even when one uses
a more quantitative process, it is always useful to perform an intuitive “reality
check” at the end.
15. In SMART CHOICES, supra, Hammond, Keeney, and Raiffa describe how one can
extend this procedure to handle multiple alternatives simultaneously.
16. Ralph L. Keeney, Value-Focused Thinking: A Path to Creative Decisionmaking
6 (Cambridge: Harvard University Press, 1992).
102 problem solving, decision making, and professional judgment
17. Ralph L. Keeney and Howard Raiffa, Decisions with Multiple Objectives:
Preferences and Value Tradeoffs 1–2 (New York: Cambridge University Press, 1993).
18. See Keeney and Raiffa; Keeney, Value-Focused Thinking.
choosing among alternatives 103
• Site A. Airport site. This site, directly adjacent to the Edenville Municipal
Airport, is in the outer, western section of Edenville. This site would be
subject to height restrictions imposed by flight paths maintained by the
airport and would necessitate a contract with the airport. Initial cleanup of
land and pipelines to the north would be required.
• Site B. Northeast suburban site. This site is directly adjacent to a river that
feeds into a lake that is used for recreational activities. It is in the heart of
the quickly growing northern suburban land.
• Site C. Outermost north rural site. This site features a pristine piece of
large land in the northernmost section of Edenville. This area mostly con-
tains farmlands and is not yet impacted by development pressures.
• Site D. Expansion/upgrade of existing site. This major expansion option
would transport sewage from the growing population center in outer
Edenville to the existing, large wastewater treatment plant in south central
Edenville by laying pipelines through neighborhoods. The existing plant
would be greatly expanded and equipped with new technology to make it
more efficient.
4.5.3 Evaluate (Objectively or Subjectively, as the Case May Be) the Attributes
for Each Alternative
Lamm considers each alternative in turn:
• Site A. The airport area, which is not residential but industrial, presents a
site compatible with the plant. The flight paths would likely be easy to
accommodate. However, this site, while politically optimal, requires a large
economic investment in the cleanup of the land due to illegal dumping and
existing structures. Because of its location, this site would also require
costly pipelines to the north.
• Site B. This site is located optimally for proximity to influent sources, since
it is at the heart of the growing residential community. It will reduce the
number of pump stations involved. However, it is also greatly impacted by
development pressures, making any land expensive and limiting the size
of any facility. Similarly, the recreational character of the adjacent area
makes industrial use of this site problematic. Also, the downstream recre-
ational uses would require that the quality of the effluent from the waste
treatment plant be fairly high.
• Site C. This site provides the most space, with the capacity to screen the
plant. It sits high on hill, however, and would affect the view for the popula-
tion in north Edenville. Also, the site is the habitat of some small animals,
an issue which may be raised by environmental organizations.
• Site D. This expansion/upgrade option includes a lower capital cost for treat-
ment facilities and avoids the need to site new treatment facilities. This
option is attractive because most community members prefer the maximum
108 problem solving, decision making, and professional judgment
use of the existing wastewater treatment plant before siting a new one.
With investment in new technology, the upgraded plant could make use
of existing infrastructure for space-efficient increase of treatment capacity
and effluent quality. However, the disadvantages are the costly pipelines,
technology investment, size limits, and environmental justice issues.
Lamm first creates a chart using qualitative words to describe the values of the
attributes for each possible site as shown in Table 4.2.
She then sums up the score of the attributes for each site, as shown in
Table 4.4
A 80 90 85 60 75 90 480
B 75 60 60 50 100 60 405
C 100 50 50 95 90 65 450
D 50 85 80 100 50 70 435
4.5.5 Aggregate the Evaluations for Each Alternative and Compare the
Alternatives
Site A has the highest score, followed at a distance by Site C.
20. See, e.g., Paul Kleindorfer, Howard Kunreuther, and Paul Schoemaker,
Decision Sciences: An Integrative Perspective 137–39 (New York: Cambridge
University Press, 1993).
21. There are multi-attribute decision approaches that are designed to reflect the fact
that some of our preferences are nonlinear. Indeed, through a process known as conjoint
analysis, it is possible to assess an individual’s utility for two or more attributes and deter-
mine trade-offs without engaging in the rather abstract weighting process that Christine
Lamm engaged in. See Jonathan Baron, Thinking and Deciding 340 (3rd ed. New York:
Cambridge University Press, 2000).
110 problem solving, decision making, and professional judgment
assign percentages to each one so that percentages all add up to 100 percent,
with (say) these results:
Size 20%
Enviro Impacts 15%
Traffic Impacts 15%
Land Cost 5%
Pipeline Lengths 20%
Compatible 25%
At this point, Lamm would multiply each attribute’s normalized value by its
weight (step 4 above), as shown in Table 4.5.
Score Score Score Score Score Score Score Score Score Score Score Score
x x x x x x
weight weight weight weight weight weight
B 75 15 60 9 60 9 50 2.5 100 20 60 15
She would then sum up and compare the final scores for each firm, as shown
in Table 4.6.
table 4.6 wastewater treatment plant: aggregate weighted scores
Site Size (20%) Enviro Traffic Land Cost Pipeline Compatible Total
Impacts Impacts (5%) Lengths (25%)
(15%) (15%) (20%)
B 15 9 9 2.5 20 15 70.5
Even with this weighted adjustment, Site A and Site C still lead the choices,
confirming the earlier unweighted findings.
Is the additional step worthwhile? The calculations don’t take much effort,
but if you put yourself in Lamm’s shoes or, more accurately, in her head, you
may find it difficult to assign weights to the attributes and may be left feeling that
the weightings are pretty arbitrary. In fact, there are some useful strategies for
determining a decision maker’s weightings.22 But it turns out that a simple non-
weighted process works pretty well, even for complex decisions.23
In any event, a subjective linear model has clear advantages over noncompen-
satory procedures for making decisions with multiple alternatives and attributes.
Though it may be tedious, it is computationally simple. It is especially useful
when a decision must be based on a number of different people’s input, because
it tends to make the potential differences in their desires or values explicit.
22. For example, under the process of conjoint analysis mentioned in a previous foot-
note, one would determine Lamm’s utility function by interrogating her about many
hypothetical trade-offs.
23. See Section 10.5.1; Robyn Dawes, The Robust Beauty of Improper Linear Models in
Decision Making, 39 American Psychologist 571–82 (1979).
24. For a description of decision opportunities, see Keeney, Value-Focused Thinking,
241–67.
112 problem solving, decision making, and professional judgment
Table 4.7 summarizes the salient features of some of the different decision-
making strategies outlined above. “Transparency” refers to how clearly one can
articulate and explain the basis for the decision—which are predicates for
discussing it with others and justifying it to them.25
Most decisions about which lawyers are asked to counsel clients are not of the
repetitive sort that would permit using rules. And although clients may some-
times ask a lawyer for her “gut reaction,” good counseling usually calls for a
transparent process in which the client’s interests and the extent to which alter-
natives satisfy them are articulated and discussed.
The range of public policy decisions is huge, with many being rule-based or
amenable to formal decision-making processes and others calling for on-the-
spot assessments that preclude formalization. The administrative law scholar
Thomas O. McGarity writes of two different ways of thinking about administra-
tive decision making:26
Although the goal of rational agency decisionmaking seems unexceptional,
its proponents had in mind a very ambitious agenda. They meant to interject
a new and very different way of thinking into a firmly entrenched bureau-
cratic culture. Following the conventional nomenclature, we may label this
new kind of thinking “comprehensive analytical rationality.” The term “com-
prehensive” suggests that this kind of thinking ideally explores all possible
routes to the solution of a problem. The term “analytical” implies that it
25. The table is based on J. Edward Russo and Paul J.J. Schoemaker, Winning
Decisions 155 (New York: Doubleday, 2002).
26. Thomas O. McGarity, Reinventing Rationality—The Role of Regulatory
Analysis in the Federal Bureaucracy 6–7 (New York: Cambridge University Press,
1991).
choosing among alternatives 113
attempts to sort out, break down, and analyze (quantitatively, if possible) all of
the relevant components of a problem and its possible solutions. The term
“rationality” captures the pride that its proponents take in its objectivity and
the dispassion with which it debates the pros and cons of alternative solutions
without regard to whose ox is being gored. In practice, comprehensive
analytical rationality has been dominated by the paradigms of neoclassical
micro-economics.
This kind of rationality contrasts sharply with the thinking that has tradi-
tionally dominated the rulemaking process in most regulatory agencies,
which I shall refer to as “techno-bureaucratic rationality.” I use the term
techno-bureaucratic to distinguish the thinking that dominates highly
technical rulemaking activities that must grapple with highly complex (and
often unresolvable) issues of science, engineering, and public policy. Some of
the existing models of bureaucratic thinking, such as Lindblom’s perceptive
“muddling through” model, are relevant to bureaucratic programs that have
highly technical, scientific, and engineering components. I use the word
“rationality” because, unlike many students of regulation, I do not believe that
this kind of thinking is irrational per se. Techno-bureaucratic rationality is
a rationality built on a unique understanding of the regulatory universe that
is born out of frustrating hands-on experience with unanswerable questions
of extraordinary complexity. It is, in a sense, a “second best” rationality
that recognizes the limitations that inadequate data, unquantifiable values,
mixed societal goals, and political realities place on the capacity of structured
rational thinking, and it does the best that it can with what it has.
All things being equal—that is, given sufficient time, cognitive capacity, and
patience—value-based decision making is more transparent and seems more
likely to satisfy stakeholders’ long-run needs than its alternatives. But all things
are seldom equal. People appropriately tend to reserve full-fledged decision
value-based models (of which we have provided the most simple example) for a
very limited number of important decisions, where it is plausible to quantify or
at least ordinally rank the attributes of each alternative.
This page intentionally left blank
part two
making sense of an uncertain
world
Most issues that lawyers and policy makers encounter involve uncertainties.
We may be uncertain about what happened in the past: what caused the tenants’
rashes and other symptoms at Terra Nueva? And we are inevitably uncertain
about what will happen in the future: what is the likelihood that a particular
approach to eradicating malaria will actually work? How much do we need to
reduce greenhouse gas emissions to prevent the earth’s temperature from rising
more than 2 degrees, and what will happen if it does rise more?
Part 2 is devoted to judgment in the empirical sense of how one ascertains or
predicts facts about the physical and social worlds in situations of uncertainty. It
focuses mainly on the past—on whether a particular event or phenomenon
occurred and on understanding its nature. Often—as in the case of assessing
liability for a tort or crime—the past is of legal interest in its own right. And
almost always understanding what occurred in the past is essential to making
forward-looking decisions. For example, the decision whether to ban the use of
a putatively toxic product will be based on evidence about what happened to
people who were (or were not) exposed to it in the past.
that happen. Consider the not uncommon situated where a teacher believes
that a student will succeed or fail and, as a result of the “Pygmalion effect,”
helps make the prediction true. A famous series of experiments on the
determinants of worker productivity at the Western Electric Company’s
Hawthorne plant showed that productivity was increased by virtually
any intervention that made the workers feel important and valued—the
so-called Hawthorne effect.2
The “lens” model, first proposed by Egon Brunswik, and later developed by
Kenneth Hammond, Robin Hogarth, and others,3 provides a particularly helpful
way of conceptualizing the tasks of empirical inquiry under conditions of uncer-
tainty. The model assumes that there is a “real world” out there—we’ll call it the
environment—and that it is the observer’s task to use available evidence, or indi-
cators, to discover it.4 We offer our own, slightly revised version as shown in
Figure A.
Environment Observer
Lenses
2. Richard Herbert Franke and James D. Kaul, The Hawthorne Experiments: First
Statistical Interpretation, 43 American Sociological Review 623–643 (1978).
3. Kenneth R Hammond, Human Judgment and Social Policy: Irreducible
Uncertainty, Inevitable Error, Unavoidable Injustice (New York: Oxford University
Press, 2000). Robin Hogarth, Human Judgement and Social Choice (2d ed. New York:
John Wiley & Sons, 1987).
4. While this is inconstant with the wildest version of postmodernism, it is entirely
consistent with the notion that reality is “socially constructed”—that is, that there is no
one right or universal way to describe things, events, or phenomena. One need only look
at the contemporary debates over abortion to see two very different views about what it
means to be a “person,” with vastly different moral and legal implications for terminating
its existence. In the end, though, the question whether someone did in fact terminate its
existence (whatever the disagreement about the nature of the “it”) is a fact subject to
empirical study—a fact that can be known with some degree of certitude.
118 problem solving, decision making, and professional judgment
On the left side is the environment, or real world. One can think of the
lines emanating from the environment as indicators of its actual features or
characteristics. A fundamental characteristic of the environment is that it is
probabilistic—that is, the real world is accessible only through what Hammond
terms multiple fallible indicators: The relationship between the real world and
its indicators cannot be represented by strict functional rules, but only as
probabilities.5
On the right side is the observer’s perception of reality, or judgment, as he or
she seeks to integrate those multiple fallible indicators into a coherent under-
standing of the real world. But perception inevitably takes place through lenses,
depicted in the center—lenses that can be more or less distorted.
To use a simple perceptual example: You’re waiting on a corner to meet your
friend Karen for lunch. You see an object in motion on a sidewalk. The light
reflecting off the object (the left side of the lens diagram) shows its features, its
gait, and its apparel. As it gets closer, you discern a human being . . . a woman . . .
Karen. You wave to her. But as she gets still closer, you realize that it’s not Karen
after all, but a stranger. What happened? The characteristics were ambiguous—-
multiple fallible indicators. Moreover, your perception of those indicators was
not “neutral”—for example, it incorporated whatever schemas were active in
your mind at the time: you were supposed to meet Karen at this very time and
place and therefore expected that the person would be her.
In many legal applications, the person making the ultimate judgment—let’s
say a judge hearing a civil case—will not have direct access to the event but must
rely on reports of others’ perceptions. For example, suppose that a judge is trying
to determine whether Donna sideswiped Patricia’s car while it was parked at the
supermarket. Donna denies being at the scene. An eyewitness testifies that she
saw Donna graze the car and drive off. An expert testifies that paint scratched
onto Patricia’s car matches that of Donna’s. The lens model applies not only to
each witness’s recollection of what he or she observed, but also to the judge’s
perception of the statements of each witness. Consider the many ambiguities
in the evidence and the opportunities for distortion both in people’s initial per-
ception and in their recollection. (One might think of the adversary model,
5. Hogarth, supra at 10. For all practical purposes, it does not matter whether the
uncertainty is inherent in the environment or only reflects our limited access to the
environment (as it really is). In either case, we are saddled with the lack of certainty.
Hammond notes that the indicators can be fallible on two dimensions: the accuracy of
information or the accuracy of the cue or indicator reflecting the information. Using the
term “ecological” to refer to its relationship to the environment, or real world, he writes:
“Accuracy of information can be described in terms of its ecological reliability; the accuracy
of a cue or indicator is described as its ecological validity.” For example, while wind direc-
tion may have a high ecological validity for predicting rain, holding up one’s finger to the
wind is unreliable compared to a wind gauge. Hammond, supra at 170.
making sense of an uncertain world 119
One kind of question asks about the manner in which a decision or conclu-
sion was reached; the other inquires whether a given decision or conclusion
is justifiable. . . .
Consider: . . . A scientist who has discovered a vaccine which purportedly
provides complete immunization against cancer informs the scientific commu-
nity that he hit upon this particular chemical combination in the following
manner. He wrote down 1,000 possible chemical combinations on separate
pieces of paper, put them all into a big hat, and pulled out the pieces at random.
. . . The scientist has announced how he arrived at the conclusion that this
chemical formula might immunize against cancer, but of course he has not
answered the question of whether the vaccine will in fact immunize. . . .
Whether this formula is an effective vaccine is a quite different one.
Furthermore, if ex hypothesi the vaccine were effective, it certainly would not
be rejected because of the way in which the scientist selected it for testing. . .
The example is not purely hypothetical. Friedrich August Kekulé von
Stradonitz, who is credited with discovering the structure of benzene, said that it
came to him in a dream of a snake eating its own tail.
Intuition is critical to the second of the four stages of the process, and it also
plays roles in observation, hypothesis testing, and generalization.9
Part 2 considers essentially three questions. First, what are, in effect, “best
practices” for engaging in these processes in a formal, analytic way? Second, in
what ways do people systematically deviate from these practices, and with what
consequences? And finally, what, if any, useful generalizations can one make
about the roles of intuition and analysis in empirical judgment?
the analogy between these two processes and an appellate judge’s developing intuitions
about the correct decision in a case from justifying the decision. The former is analogous
to the speculative process of developing a hypothesis under the scientific method, while
the latter is analogous to the process of testing and generalizing a hypothesis. He writes
that “to insist—as many legal philosophers appear to have done—that a judicial opinion
is an accurate description of the decision process there employed if and only if it faithfully
describes the procedure of discovery is to help guarantee that the opinion will be found
wanting. But if the opinion is construed to be a report of the justificatory procedure
employed by the judge, then the not infrequent reliance on such things as rules of law and
rules of logic seems more plausible.”
9. The process is often recursive, with testing and generalizing leading to new observa-
tions and hypotheses. Hence Michael Polanyi’s description of a mathematician’s work as
an alternation between intuition and analysis. Michael Polanyi, Personal Knowledge:
Towards a Post-Critical Philosophy 130–31 (New York: Harper & Row, 1964).
making sense of an uncertain world 121
I keep saying that the sexy job in the next 10 years will be statisticians. And I’m not
kidding.
—Hal Varian, chief economist, Google11
The “best practices” for testing empirical hypotheses are rooted in statistics.
Examples of statistics abound in legal disputes, some of which involve major
issues of public policy. The evidence in Anne Anderson v. W.R. Grace & Co., made
famous by A Civil Action, involved alleged correlations between drinking water
from contaminated wells and childhood leukemia in Woburn, Massachusetts.
Statistical evidence has played a central role in litigation concerning asbestos,
silicone breast implants, electromagnetic radiation from cell phones, and autism
putatively caused by the mercury preservative in childhood vaccines. In Castenada
v. Partida,12 the Supreme Court relied on statistical analysis of the correlation
between Mexican Americans in the population and on grand juries to conclude
that Mexican Americans were systematically excluded from grand jury service.
In Hazelwood Independent School District v. United States,13 the Court applied a
similar analysis to determine whether the school district discriminated in the
hiring of African American teachers.
Beyond their professional identities, lawyers and policy makers are citizens,
consumers, and community leaders. If nothing else, basic concepts of probabil-
ity and statistics will improve their understanding of newspaper articles on
health, the environment, and many other subjects and give them a healthy skep-
ticism about claims made by advocates, businesses, and governments.
This hardly entails that every lawyer, policy maker, and citizen must be a stat-
istician. But it does suggest that you should grasp the fundamental concepts of
probability and statistics—to understand how a statistician thinks about
the empirical world. Possessing this basic knowledge will alert you to issues
you might otherwise miss and, just as important, provide you with an
informed skepticism about causal claims. Such knowledge will help you know
10. Oscar Wilde went even further, to write that “it is the mark of an educated man to
be moved by statistics.” But see Joseph Stalin: “The death of one man is a tragedy; the
death of a million is a statistic.”
11. Steve Lohr, For Today’s Graduate, Just One Word: Statistics, New York Times,
Aug. 5, 2009.
12. 430 U.S. 482 (1977).
13. 33 U.S. 299 (1977).
122 problem solving, decision making, and professional judgment
when to seek expert advice, and will enable you to communicate with experts.
Moreover—referring back to the “two systems” of judgment described in
Section1.6.1—there is reason to believe that a grasp of probability and statistics
will reduce the “intuitive” statistician’s errors. The evidence indicates that “even
very brief” statistical training has “profound effects on people’s reasoning in their
everyday lives14 and mitigates the biases of unmediated System I processes.15
The following chapters are premised on the belief that understanding the
fundamental concepts is more important than being able to do complex calcula-
tions. Our goal, essentially, is to help you become a savvy consumer of statistics
and strengthen your empirical intuitions. If nothing else, the chapters should
demystify some statistical concepts and terms that you encounter in reading
scholarly articles, legal opinions, and newspapers.
14. Richard Nisbettt, David Krantz, Christopher Jepson, and Ziva Kunda, The Use of
Statistic Heuristics in Everyday Inductive Reasoning, in Thomas Gilovich, Dale Griffin, and
Daniel Kahneman, Heuristics and Biases: The Psychology of Intuitive Judgment
520, 528, 529 (New York: Cambridge University Press, 2002)
15. Franca Agnoli, “Development of Judgmental Heuristics and Logical Reasoning:
Training Counteracts the Representativeness Heuristic” 6 Cognitive Development 195 (1991);
Franca Agnoli & David H. Krantz, “Suppressing Natural Heuristics by Formal Instruction:
The Case of the Conjunction Fallacy,” 21 Cognitive Psych. 515 (1989). See also Peters, et al.,
Numeracy and Decision Making, 17 Psychological Science 407 (2006).
5. introduction to statistics and probability
This chapter introduces statistical thinking and its application to law and public
policy. At the end of the chapter, you can find a glossary with definitions of the
numerous terms we discuss. Our main example will involve the tenants’ com-
plaints of illnesses possibly caused by foam insulation in the apartments at Terra
Nueva (discussed in Chapter 1). Before turning to this example, though, we
provide a brief introduction to probability.
5.1 probability
A natural way of expressing beliefs about events that may or may not occur is
with ideas of chance. You might say that the odds of a certain candidate winning
an election are 3 to 1, or that he or she has a 75 percent chance of winning. In
making these statements, you are expressing your beliefs in terms of the proba-
bility (or chance) of a particular event. The language of probability is an effective
way to communicate the strength of beliefs in cases where we are not sure about
an outcome but have some intuition on which to base our thoughts. In these
situations of uncertainty, as well as situations in which the involvement of chance
is more explicit, we use probabilistic notions to express our knowledge and
beliefs.
If we tossed a coin ten times, we might see four heads, five tails, and then
another heads: HHHHTTTTTH, for example. These results are called realized
values, realizations, or instances of the random variable coin toss.
1. This example assumes that heads and tails are not only mutually exclusive but jointly
exhaustive: i.e., one and only one of them must occur. If we wanted to factor in the possi-
bility that a coin could land and remain on its edge, the formula would be somewhat more
complicated: P(heads or tails or edge) = 1. Therefore, P(tails) = 1 – P(heads) – P (edge).
introduction to statistics and probability 125
5.1.2.a Odds We just said that “the odds of a candidate winning an election
are 3:1” is equivalent to “the candidate has a 75 percent chance (or a probability
of 0.75) of winning.” Odds of 3 to 1 means that the event is likely to occur about
3 times for every time it does not, that is, 3 out of 4 times, which can be repre-
sented as ¾ = 0.75. In terms of probabilities, the odds of an event are the ratio of
the probability that it will happen to the probability that it will not:
Odds of event A = P(A)/P(∼A).
We know from the complement rule that P(∼A) = 1 - P(A), so
Odds of event A = P(A)/(1 - P(A)).
In this case, the ratio of the probability of the event—P(win) = 0.75—to the
probability that it does not occur—P(lose) = 0.25—is 3:1.
5.1.2.b Meaning of Probabilities: Frequentism vs. Subjectivism What does it
mean to say, “The probability of the coin landing on heads is 0.5?” There are two
ways to define probabilities, frequentism and subjectivism. A frequentist would
define the probability of heads as the relative frequency of heads if the coin were
tossed infinitely many times. After many many tosses, 50 percent of them will
have landed on heads and 50 percent on tails, so each probability is 0.5.
On the other hand, a subjectivist would say that the probability of a coin land-
ing on heads represents how likely this event seems, possibly based on some
evidence. For example, she might examine the coin and conclude that there is no
evidence that it is weighted to make heads more likely than tails or vice versa, so
the two events must have equal probabilities.
From the axioms in 5.1.2, we saw that the probabilities of mutually exclusive and
jointly exhaustive events sum to 1. In flipping a coin, P(heads) + P(tails) = 1.
The same holds with more than two mutually exclusive events. When tossing a
die, we can consider the random variable of what number lands facing up, with
possible values 1, 2, 3, 4, 5, and 6. Then P(1) + P(2) + P(3) + P(4) + P(5) +
P(6) = 1.
With any random variable, a total probability of 1 is distributed among all pos-
sible values. If we represented the total probability of 1 with a dollar in coins,
those coins would be distributed over the possible values, with, 50c/ on heads and
50c/ on tails. In the die toss, 1/6, or about 17 cents, would be on each number
1–6. In these examples, the probability happens to be distributed evenly, but this
doesn’t have to be the case: imagine a die weighted so that the “6” came up more
than 1/6 of the time.
Like a set amount of money ($1) or a set number of die faces, there is a set
amount of probability (1), and it is distributed (not always evenly) over the mutu-
ally exclusive events. How it is distributed over the events is called the probability
126 problem solving, decision making, and professional judgment
Side Up Frequency
1 11
2 19
3 22
4 19
5 11
6 18
Dividing these by the total number of tosses, 100, we obtain the relative fre-
quencies, which are displayed in Figure 5.1(b). The relative frequency of an event
is the proportion of total opportunities on which it actually occurs, i.e., the ratio
of the number of times the event occurs (its frequency) to the total number of
times it could have occurred. In this case, for example, the die came up a 4 in 19
times out of a possible 100, so the relative frequency of the event that it comes up
a 4 is 19/100 = 0.19.
The difference between Figure 5.1(a) and 5.1(b) is key to statistics. Figure
5.1(a) represents the probability distribution, which is the same for every die
toss. Figure 5.1(b), on the other hand, results from just one experiment of one
hundred tosses. Each of these one hundred realizations came from one die toss,
with the probability distribution shown in Figure 5.1(a). For another experiment
of one hundred tosses, the result would be slightly different, because it depends
on chance.
introduction to statistics and probability 127
0.00 0.05 0.10 0.15 0.20 0.25 Die Toss Probability Distribution Die Toss Relative Frequencies
1 2 3 4 5 6 1 2 3 4 5 6
(a) Number facing up (b) Number facing up
It is worth pausing for a moment to emphasize this point. Recall the lens
model discussed in the introduction to Part 2: you can never know the real world
directly, but can only estimate it through multiple fallible indicators. In this
sense, the die toss random variable is completely atypical because (assuming that
it’s a perfect die) we know that its population distribution is 1/6 for each possible
value.
In summary then:
• Population distribution: The true distribution of the population will never
be known. We estimate its characteristics by studying our sample, but
these estimates will probably be a little bit off the truth—a different random
sample of the same size will usually have resulted in slightly different esti-
mates.
• Empirical distribution: The distribution of events within our random
sample from the population is called the empirical distribution.
Recall the Terra Nueva problem discussed in Chapter 1. The survey to determine
the incidence of rashes included 453 residents. We know the empirical distribu-
tion of the incidence of rashes of those who live in apartments with and without
foam insulation. But these are fallible indicators of the population distribution,
which cannot be known directly.3 The task of statistics is to inform us how good
an estimate of the population we can make based on the sample. The purpose of
statistical inference is not to determine whether the foam insulation caused a
rash in these particular 453 individuals; it is to use the sample to infer whether
foam insulation is related to rash in the general population of Terra Nueva.
3. In this particular instance, it might have been feasible to sample the entire actual
population of Terra Nueva. However, in most real-world situations, from environmental
hazards to public opinion polls, this isn’t possible or cost-effective. In any event, as a con-
ceptual matter, statistics assumes that the actual population is infinite—all possible past,
future, and present tenants of Terra Nueva.
introduction to statistics and probability 129
The typical format for depicting two categorical variables is called a contin-
gency table because, as we will see later, it can be used to determine whether one
variable is contingent on the other. In Table 5.1, in each of four cells is a
frequency, the number of members of the sample that fall into that cell’s joint
categories.
Present Absent
Not Foam not foam & rash not foam & no rash
Probabilities Rash
In some situations, one might want to display the data as proportions of the
total—that is, as relative frequencies. Table 5.5 displays the same information as
Table 5.4, but in relative frequencies. (Note that the format of the table resembles
that of the (unknowable) probability distribution of the population from which it
sampled. Table 5.3 is analogous to Figure 5.1(a), while Table 5.5 is analogous to
Figure 5.1(b).)
(A different sample of tenants would almost surely have produced slightly
different numbers than the sample represented in Tables 5.4 and 5.5).
Our goal is to use the frequencies we observe in this sample of residents to esti-
mate the probabilities, i.e., the relative frequencies in the total population. For
example, in our sample of 453 residents, we see that 120 had both a rash and
foam insulation, giving a relative frequency of 0.26. But a sample is always a fal-
lible representation of the population. That is, an estimate is almost inevitably
somewhat off from the true probability. How can we know how closely (or not)
the relative frequencies of the sample approximate it?
We begin with a discussion of the concept of sampling, and then turn to
the concepts of conditional probability and independence. Finally we dis-
cuss the matter of hypothesis testing—and in particular ascertaining the
independence (or not) of having a rash and living in a foam-insulated
apartment.
5.5.1 Sampling
Like pollsters before an election, lawyers and policy makers often are interested
in the characteristics of a large population but lack the time or resources to mea-
sure or survey every member of that population. This leads us to choose a smaller
random sample from the population and take the relevant measurements of
those in the sample.
Often, investigators take a simple random sample, which means that each
population member has an equal chance of being in the sample. From the infor-
mation on the sample, we use statistics to infer a conclusion about the entire
population.
Population -> Sampling -> Sample -> Inference -> Conclusions about population.
For pollsters, the population of interest consists of all voters in the region
of interest. A simple poll might survey a random sample of, say, 1000 likely
voters from the population of all voters in the region, then use statistical
inference to predict the outcome of the election. However, if the sample was
not chosen at random and therefore is not representative of the population as
a whole, the conclusions may be incorrect. This is a problem called selec-
tion bias or sampling bias; examples are in the “Surveys” box. (In reality,
pollsters do not always use simple random sampling and often instead
use complex formulas based on knowledge about party registration, past
elections, etc.)
You may have noticed that such predictions come with a “margin of error.”
That’s because the sample does not exactly reflect the entire population from
which it is drawn. In the next chapter, we’ll discuss the characteristics of the
sample that determine its margin of error.
introduction to statistics and probability 133
Surveys
HANS ZEISEL AND DAVID KAYE, PROVE IT WITH FIGURES: EMPIRICAL METHODS IN LAW AND
LITIGATION 103–04 (New York: Springer-Verlag, 1997); Evaluation of Edison/
Mitofsky Election System 2004 prepared by Edison Media Research and Mitofsky
International for the National Election Pool (NEP), Jan. 19, 2005, http://www.
exit-poll.net/faq.html. See generally Fritz Scheuren, “What is a Survey,” http://
www.whatisasurvey.info/.
In Section 5.4.1, we discussed marginal and joint probabilities. Recall that mar-
ginal probabilities are those (in the margins of the contingency table) that give
the probability for a single variable, regardless of what the other variable equals,
such as P(foam) = 0.6 and P(rash) = 0.4, while joint probabilities are those in the
center of the table, such as P(foam & rash) = 0.25.
In this section, we introduce conditional probability. A conditional probability
is the probability that event A will occur if we know that B has occurred. It is
written
P(A|B)
and is stated: “the probability of A given B.”
For example, suppose you are flying from San Francisco to Cedar Rapids via
Chicago. We can talk about the probability of your catching the flight from
134 problem solving, decision making, and professional judgment
Chicago to Cedar Rapids given that your flight from San Francisco leaves on
time—or given that it is thirty minutes late.
Conditional probabilities tell us the probability distribution of one variable
when another variable is known. In the Terra Nueva case, one might want to
know the probability of a tenant’s having a rash, given that she lives in an
apartment with foam insulation. We represent this conditional probability
formally by
P(rash| foam)
As noted above, the “|” is read “given,” so this is stated as: “the probability that
the tenant has a rash, given that she lives in an apartment with foam insulation.”
Although this probability does not appear in Table 5.3, it can be calculated
from those that are. Because we know the insulation is foam, we focus only on
the top row. The conditional probability is that fraction of the total probability of
the top row (0.60) for which a rash is present. That is,
P(rash| foam) = 0.25 / 0.60 = 0.42
One way to think about this is in terms of frequencies: Imagine that there are
1000 people total. Six hundred of them have foam insulation. And of those 600,
250 have a rash. So, of the group with foam insulation, the fraction with a rash is
250/600 = 0.42—that is, 42 percent of those with foam insulation have a rash.
More generally, we can write this calculation as:
P(rash present| foam) = P(foam & rash)/P(foam)
Is P(rash|foam) the same as P(foam|rash)? Or to put the question more gener-
ally, is P(A|B) the same as P(B|A)? Absolutely not! Consider the difference
between P(pregnant|woman) and P(woman|pregnant).
Correction: The Mail of January 3rd contained the incorrect statistic that
four-fifths of Bush voters identified moral values as the most important
factor in their decision. In fact, four-fifths of those identifying moral values
as the most important factor of their decision were Bush voters. New
Yorker, Feb. 14–21, 2005, 38.
5.6.1 Independence
Conditional probabilities are used to express information about how knowledge
of one variable influences our knowledge of another. Often, we want to know if
introduction to statistics and probability 135
two variables are, or are not, related to each other. In the language of probability
theory, we can ask whether or not two variables are independent.
Formally, we say that two events are independent of each other if P(A|B) =
P(A)—which also implies that P(B|A) = P(B). This means that the probability of
A happening does not change, regardless of whether B happens or not, and vice
versa. This matches up well with the intuitive notion of independence: If we say
that two things are independent, we usually mean that they have no effect on one
another. In the example above, is the probability of your catching the flight from
Chicago to Cedar Rapids independent of the fact that your flight from San
Francisco left on time? Is it independent of the fact that the flight number is the
same as the month and day of your birth?
In the Terra Nueva case, if insulation and rash were independent, then being
in an apartment with foam would not make it any more or less likely that a resi-
dent would develop a rash. In terms of conditional probabilities, the probability
of having a rash given that one lives in a foam-insulated apartment would be
identical to the probability of having a rash given that one lives in any apartment,
or P(rash|foam) = P(rash). Based on the calculations in the previous subsection,
does this seem to be true?4 We’ll find out in a while.
In gambling, events are supposed to be statistically independent. For exam-
ple, subsequent spins of a roulette wheel should give numbers that are com-
pletely unrelated to previous spins. If this were not the case, a player could use
the result of previous spins to improve his or her chances of winning, to the
detriment of the casino. By the same token, subsequent tosses of a coin are also
independent: the information that the last toss produced a head tells you nothing
about the expected outcome of the next toss.
A convenient consequence of independence is that when two variables are
independent, the probability that an event involving one and an event involving
the other both occur is the product of the marginal probabilities. If the first spin
and the second spin of a roulette wheel are independent, then the probability
that the ball lands on red on the first and on black on the second is
P(first spin = red & second spin = black) = P(first spin = red) * P(second spin
= black)
Similarly, if getting a rash were independent of one’s type of insulation, then
the joint probability that a Terra Nueva resident has foam insulation and a rash
would be
P(rash & foam) = P(rash) * P(foam)
4. Without knowing a tenant’s insulation type, her chance of having a rash is the mar-
ginal probability, 0.40. When we learn that her apartment has foam insulation, however,
the chance increases to 0.42. So insulation and rash are not independent.
136 problem solving, decision making, and professional judgment
In our case, in particular, this would be 0.6*0.4 = 0.24. If rash and insulation
were independent, and we selected a Terra Nueva resident at random, there
would be a 24 percent chance that the resident had both foam insulation and
a rash.
In general, if events A and B are independent,
P(A & B) = P(A) * P(B)
This is called the multiplication rule for independent events.
Problem: When rolling a pair of dice, what is the probability of rolling a
double six?
The idea of independence is essential to statistics. The existence of a depen-
dency between two variables means that those variables are related. Much of the
practice of statistics involves attempting to find relationships among variables,
and to quantify the strength of those relationships. These tasks correspond to
testing whether variables are independent and measuring the extent to which
they deviate from independence.
Schemas are an essential part of the human situation. “We are predisposed to see order,
pattern, and meaning in the world, and we find randomness, chaos, and meaning-
lessness unsatisfying. Human nature abhors a lack of predictability and the absence of
meaning.
—Thomas Gilovich.5
Much psychological research has examined how good people are at detecting
relationships or their absence. It turns out that people often neglect the implica-
tions of statistical independence when evaluating the probability of events being
produced by a random process.
5. Thomas Gilovich, How We Know What Isn’t So: The Fallibility of Human
Reason in Everyday Life, 9 (New York: The Free Press, 1991).
6. Donald A. Redelmeier and Amos Tversky, On the Belief that Arthritis Pain Is Related
to the Weather, 93 Proceedings of the National Academy of Sciences 2895 (1999).
introduction to statistics and probability 137
The winner of the Spanish lottery, who bought a lottery ticket ending in 48, when asked
why he chose that number, explained that for 7 nights in a row he dreamed of the number
7, and since 7 times 7 equal 48, he knew that would be the lucky number.9
Related to illusory correlation, people also manifest what Ellen Langer has
termed the illusion of control—acting as if they exercise control over events that
are in fact matters of chance.10 For example, some gamblers throw the dice softly
to get low numbers and hard to get high numbers; people playing a game of
chance will bet more against an opponent who looks like a “schnook” than one
who seems competent; and people who choose a lottery ticket demand a higher
price to sell it than those given a lottery ticket at random. In some of these cases,
a misunderstanding of chance may be combined with a desire or need for per-
sonal efficacy—the drive to master one’s environment begins early in infancy,
and the sense of not being in control can be stressful and debilitating—or with a
tendency to believe that if we desired an outcome and it occurred, we must have
been responsible for bringing it about.11
7. Lauren J. Chapman and Jean P. Chapman, Genesis of Popular But Erroneous Psycho-
Diagnostic Observations, 72 Journal of Abnormal Psychology 193–204 (1967); Lauren J.
Chapman and Jean P. Chapman, Illusory Correlation as an Obstacle to the Use of Valid
Diagnostic Signs, 74 Journal of Abnormal Psychology 271–80 (1969).
8. Richard Nisbett and Lee Ross, Human Inference: Strategies and
Shortcomings of Social Judgment 97 (Englewood Cliffs, NJ: Prentice Hall, 1980).
9. Stanley Meisler (1977), cited in Reid Hastie and Robyn Dawes, Rational Choice
in an Uncertain World 154 (Thousand Oaks, CA: Sage Publishing, 2001).
10. Ellen J. Langer, The Illusion of Control, 32 Journal of Personality and Social
Psychology 311 (1975).
11. Ellen. J. Langer and Jane Roth, Heads I Win, Tails It’s Chance: The Illusion of Control
as a Function of the Sequence of Outcomes in a Purely Chance Task, 34 Journal of Personality
and Social Psychology 191–98 (1975); Ellen J. Langer, The Illusion of Control, in Judgment
138 problem solving, decision making, and professional judgment
There are thirty-nine categories of work that are prohibited on the Sabbath.
Category No. 27, Kindling a Fire, also rules out the “kindling” of anything electri-
cal, including a television. It was Friday, May 6, 1994: Game Four of the Stanley
Cup conference semifinals between the Rangers and the Washington Capitals
was scheduled for that Saturday afternoon. I had decided to switch on the televi-
sion Friday afternoon—before Sabbath began, at sundown—and just leave it on
until Sabbath ended, twenty-five hours later. This wasn’t, technically, “being in
the spirit of the Sabbath,” but it wasn’t technically a sin, and the Rangers were
Under Uncertainty: Heuristics and Biases (Daniel Kahneman, Paul Slovic, and Amos
Tversky eds., New York: Cambridge University Press, 1982). See also Paul Presson and
Victor Benassi, Illusion of Control: A Meta-Analytic Review, 11 Journal of Social Behavior
and Personality (1996); Shelley Taylor, Positive Illusions 26 (1989).
12. See Suzanne Thompson, Wade Armstrong, and Craig Thomas, Illusions of Control,
Underestimations, and Accuracy: A Control Heuristic Explanation, 123 Psychology Bulletin
113 (1998); Nisbett and Ross, supra at 136–37. In Section 13.5, we will discuss research
indicating that positive illusions, such as the illusion of control, are adaptive because they
increase motivation and persistence. http://www.answers.com/topic/illusion-of-control.
13. Emily Pronin et al., Everyday Magical Powers: The Role of Apparent Mental Causation
in the Overestimation of Personal Influence, 91 Journal of Personality and Social
Psychology 218 (2006).
introduction to statistics and probability 139
very likely nine victories away from winning the Stanley Cup for the first time in
fifty-four years.
. . . I switched on the television, turned down the volume, and draped a bath
towel over the screen to hide from the neighbors the flickering blue light of our
moral weakness.
“Do you really think that if you turn on the TV on Sabbath God will make the
Rangers lose?” Orli asked.
Her naiveté astounded me.
“I don’t think He will. I know He will.”
Shalom Auslander, Personal History, NEW YORKER, Jan. 15, 2007, PG38.
14. Rachel Croson and James Sundali, The Gambler’s Fallacy and the Hot Hand:
Empirical Data from the Casinos, 30 Journal of Risk and Uncertainty 195 (2005).
140 problem solving, decision making, and professional judgment
sequence more representative of a fair coin than seeing another tail, so people
give it a higher probability.15
The other side of the gambler’s fallacy coin, so to speak, is the belief that an
individual is sometimes on a winning streak, so that random sequences are
interpreted as other than random. For example, the roulette study mentioned
above indicated that players made more “outside” bets (bets on red/black, odd/
even, or groups of numbers) when they had previously won on outside bets and
more inside bets (bets on particular numbers) when they had previously won on
inside bets.16 The notion that athletes exhibit periods of unusually good play—
the “hot hand,” “streak shooting,” or being “in the zone”—may be an example of
this fallacy.
There is a considerable debate about whether and when players actually have
a hot hand. The psychologists Thomas Gilovich, Robert Vallone, and Amos
Tversky recorded a professional basketball teams’ statistics for an entire season
and concluded that, in fact, players’ shots were independent.17 The researchers
did not argue that an individual can never be “in the zone”; only that their study
indicated that claims of this sort were unfounded. Sports fans, including schol-
ars of judgment and decision making as well as our own students, strongly resist
this finding. For example, an article on Bowlers’ Hot Hands criticized the basket-
ball study on the ground that it did not control for several confounding influ-
ences and presented evidence that most bowlers actually have a higher proportion
of strikes after bowling a certain number of consecutive strikes than after con-
secutive nonstrikes.18
Our apparently poor intuitions about randomness may also help provide
insight into another psychological phenomenon—the idea of meaningful coinci-
dence. Seemingly unlikely events catch our attention and often appear in newspa-
pers. Scott Plous gives an example of two men sharing the same surname, the
same model of car and even exactly the same car keys, both being in a shopping
center on the same time on April Fool’s Day, and managing to confuse their cars.19
15. This argument was made by Amos Tversky and Daniel Kahneman, Belief in the
Law of Small Numbers, 76 Psychological Bulletin 105–10 (1971).
16. An inside bet is made within the 36 numbers of the roulette table, for example, on
a single number or on two or four adjacent numbers. An outside bet is, say, red or black,
odd or first 12. The correlations were small but statistically significant at p < 0.05.
17. Thomas Gilovich, Robert Vallone, and Amos Tversky, The Hot Hand in Basketball:
On the Misperception of Random Sequences, 17 Journal of Personality and Social
Psychology 295–314 (1985).
18. Reid Dorsey-Palmateer and Gary Smith, Bowlers’ Hot Hands, 58 American
Statistician 38 (2004).
19. Scott Plous, The Psychology of Judgment and Decision Making 154 (New
York: McGraw Hill, 1993).
introduction to statistics and probability 141
two variables are related. In the Terra Nueva case, the alternative hypothe-
sis is that insulation and rash are not independent but rather are related
such that rash is more likely given the presence of insulation.
We test the hypotheses by comparing the observed relative frequencies (Table
5.5) to the relative frequencies that we would expect if insulation and rash were
independent. We can calculate the latter using the multiplication rule, as shown
below.
The question is, how far apart do those two tables have to be to convince us
that the null hypothesis is false?
The two tables will almost certainly be somewhat different. But is the differ-
ence small enough to be chalked up to chance or is the difference statistically
significant? A hypothesis test allows us to reach a conclusion about the entire
population based only on our sample data.
Step 2) Set the level of the test.
Since we have only a sample with which to test our hypothesis, it is possible
that we will make an error. In fact, we can distinguish two kinds of errors that
can result from performing a hypothesis test. There are two possible states of the
world: a relationship between two variables exists, or it does not. And there are
two possible conclusions we can draw: that a relationship exists, or that it does
not. In the case where no relationship exists, but we conclude that one does, we
have made a Type I error. In the case where a relationship exists but we conclude
that it does not, we have made a Type II error.
A good way to think about the difference between Type I and Type II errors is
in the context of medical diagnosis. A medical test can give two different kinds
of erroneous results: it can return a positive result when the patient does not
have the condition tested for (a false positive), or it can return a negative result
when a patient actually has the condition tested for (a miss). Both of these errors
can have important consequences for the patients involved, and medical research-
ers try to develop tests that avoid making them. In statistics, false positives are
called Type I errors and misses are called Type II errors. (See Table 5.6.)
table 5.6 type i and type ii errors
The null hypothesis is that insulation and rash are independent. Supposing
independence is true, and using the observed marginal relative frequencies
as estimates of the marginal probabilities, the multiplication rule gives us the
estimate:
P(foam & rash) = P(foam) * P(rash) = 0.59 * 0.40 = 0.24
We can do the same calculation for the other combinations of insulation and
rash: foam ¬ rash, not foam & rash, and not foam & not rash, and present them
in a new table of the relative frequencies expected under the null hypothesis. (See
Table 5.7.)
Step 4) Calculate how far the data are from what would be expected under the
null hypothesis.
introduction to statistics and probability 145
Let us compare the observed frequencies with the frequencies expected under
the null hypothesis of independence, as shown in Table 5.9.
Observed Rash
Present Absent
Insulation Foam 120 149
Not Foam 63 121
Expected Rash
Present Absent
Insulation Foam 109 160
Not Foam 74 110
In our sample, the upper left and lower right cells contain frequencies (of 120
and 121) higher than their expected frequencies (of 109 and 110). Consequently,
in the lower left and upper right cells, the observed frequencies are lower than
what is expected under independence. The further from the expected frequen-
cies the observed frequencies are, the further from independent the variables
appear to be. Here are the differences in frequencies between observed and
expected, as shown in Table 5.10.
Obs–Exp Rash
Present Absent
In a 2-by-2 table, the four differences will always have the same magnitude, with
the cells along one diagonal negative and the cells along the other diagonal posi-
tive. This is because the four marginal frequencies are the same for both observed
and expected tables, so each row and column of Table 5.10 has to sum to 0.
So the difference from what’s expected under independence is 11 in each
cell, not 0. But is 11 large enough to allow us to conclude a relationship in the
population?
To determine whether the data are extreme enough to justify the conclusion
that there is a relationship, we calculate the probability, if independence were
true, that the difference would be 11 or greater. This probability is called the
p-value of the test. If the p-value is small (specifically, smaller than our chosen
significance level of 0.05), then these data are quite extreme to be due only to
chance—quite improbable if independence were true—and we conclude that a
relationship exists, i.e., that the data are statistically significant.
The p-value can be calculated by a computer program or found in a table.
While we’re not going to go into the details here, it is calculated using the chi-
squared test (pronounced “kai squared”) for a relationship between two qualita-
tive variables.
For our data, with a sample of 453 and a difference in each cell of 11, the
p-value is 0.03. That is, if insulation and rash were independent, there would be
a 3 percent chance that the difference in each cell were 11 or greater. This is quite
improbable—0.03 is less than the level we set of 0.05—allowing us to conclude
that a relationship between insulation and rash exists.
Step 5) State your conclusion.
Because this result would be very unlikely if insulation and rash were inde-
pendence true, we are justified in rejecting the null hypothesis and concluding
that there likely is a relationship between insulation type and the absence or
presence of a rash.
22. See Jacob Cohen, The Earth is Round (p <.05), 49 Am. Psychol. 997 (1994).
23. In Section 8, we will examine Bayes Theorem, which in principle can convert P
(data| H0) into P(H0|data). But it requires assigning a value P(H0), the probability of the
null hypothesis prior to the experiment, which is often not known.
introduction to statistics and probability 147
understandable reason: we really want to know the answer to the second ques-
tion, which tells us how big a deal we can make of the data.
In general, this does not pose a problem when the results of a study turn out
to be statistically significant. The concern, rather, is that NHST may produce too
many misses (Type II errors), where a real effect goes unnoticed. We will return
to this issue in Section 7.6.
Rash
Present Absent
Insulation Foam A B
Not Foam C D
• Cell A. To what extent do the putative cause and effect happen together?
This can be phrased either as (1) when the putative cause occurs, to what
extent does the effect also occur?, or (2) when the effect occurs, to what
extent did the putative cause occur? Most people find it intuitively obvious
that this cell is necessary.
• Cell B. To what extent does the putative cause occur when the effect does
not? Consider the possibility that many Terra Nueva residents who live in
apartments with foam insulation do not have a rash.
• Cell C. To what extent does the effect occur even when the putative cause is
absent? Consider the possibility that many Terra Nueva residents have a
rash even though they do not live in apartments with foam insulation.
• Cell D. To what extent did neither the putative cause nor the effect occur?
Though less obvious, this cell is needed as well, as we describe immedi-
ately below.
148 problem solving, decision making, and professional judgment
Consider these two possible scenarios, in which cells A, B, and C are identical,
as shown in Table 5.12.
Rash
Present Absent
Foam 100 12
Insulation
Not Foam 20 0
Rash
Present Absent
Foam 100 12
Insulation
Not Foam 20 100
A week after he was featured on the cover of the very first issue of Sports
Illustrated, Major League Baseball player Eddie Mathews, suffered an
injury that forced him to miss seven games, giving rise to the notion that
being on the cover “jinxed” an athlete. In 2002, the editors reviewed all
2456 magazine covers to find that 37 percent of the featured athletes suf-
fered “measurable and fairly immediate” negative consequences. Alexander
Wolf, That Old Black Magic, Sports Illustrated, Jan. 21, 2002, 50–61.
How would you determine whether there was actually a Sports
Illustrated cover jinx?
introduction to statistics and probability 149
24. Laypersons may have a different view than experts. See Phil Brown, Popular
Epistemology and Toxic Waste Contamination: Law and Professional Ways of Knowing, 33
Health and Social Behaviour 267 (1992).
25. Ravi Maheswaran and Anthony Staines, Cancer Clusters and Their Origins, 7
Chemistry and Industry 254–256 (1997).
150 problem solving, decision making, and professional judgment
the cancer rates in other places with similar drinking water contamination. Or
it might involve controlled experiments with laboratory animals—also essen-
tially statistical. Finally, it might involve learning how the chemical alters genes
to produce the cancer.
Multiple testing, along with various biases, can affect the power, and even
validity, of studies that have statistically significant results. In the provocatively
titled, Why Most Published Research Findings Are False,26 John Ioannidis notes
that the hotter a scientific field (with more scientific teams involved), the less
likely the research findings are to be true. He argues that the possibilities for
improvement lie mainly in taking account of the prestudy odds of the truth of
the relationship that the study is testing. We will consider the notion of prior
probability in the introduction to Bayesian statistics in Section 8.5. But consider
these possible determinants:
• data from previous experimental studies;
• data from previous epidemiological or other observational studies;
Legal Problem
A plaintiffs’ law firm that specializes in toxic and related torts has a particular
interest in cancers putatively caused by strong electromagnetic fields. It has been
monitoring one hundred neighborhoods near large transformer stations. One of
those neighborhoods recently reported what appears to be a high incidence of
brain tumors in children. After informing several of the unfortunate families of
their legal rights and being retained by them, the firm undertook a well-designed
survey, which showed a statistically significant incidence of brain tumors. Does
this raise any concerns for you as a statistically astute observer of the legal
system?
26. John P. A. Ioannidis, Why Most Published Research Findings Are False, 2 PLOS MED
e124 (2005).
152 problem solving, decision making, and professional judgment
5.10 glossary
Conditional probability, The probability that A will occur given than B has
P(A|B) occurred
Sampling bias (selection bias) The amount by which an estimate is off from the
population parameter it is estimating due to the
way the sample was chosen from the population
entry mistook the four for a nine in that survey answer, the mean would change
from $39,558 to $49,558, a $10,000 increase just as a result of the change in
that single salary. Or his salary might be atypically high because he was the
boss’s son.
In cases in which it is important to avoid the effects of outliers, the median is
a better measure of central tendency. The median is the number that falls in the
middle of the observations, when they are ordered by magnitude. (If there are an
even number of observations, the median is the average of the two observations
that fall in the middle when the scores are ordered.) So, for our five salaries, the
median is $39,774. You can see why the median will not be strongly affected by
outliers—the values of the extreme scores are not involved in the calculation.
Thus, the change in one salary that would lead to a $10,000 increase in the mean
would have no effect on the median. The median is often used in reports of
housing prices, since, in a neighborhood of mostly homes worth less than
$500,000, just one multimillion dollar home could increase the mean by
thousands.
In some cases it is useful to consider the mode of a set of data. The mode is
the value that occurs the most times. The term has the same origin as “mode” in
fashion—the most popular number.
When the mean is used to measure central tendency, the standard deviation
(SD) is usually used to measure the dispersion. The standard deviation of a pop-
ulation is an indication of the amount by which the numbers differ from the
mean.
How do we determine the standard deviation? We want to know how far the
five salaries deviate from the mean salary. Since we do not know the population
mean,1 we estimate it with the mean of our sample of five, $39,558. For example,
by subtracting $39,558 from each of our five salaries, on the high side we have
what is shown in Table 6.1(a).
table 6.1(a) deviation of high salaries from the mean
Because by definition the mean is in the middle, the average of the deviations
themselves will always be zero. To get a useful measure of the amount of disper-
sion in the sample we need to make the numbers positive and give extra weight
to those observations further from the mean. To do so, we square each observa-
tion’s deviation from the mean and then take the average of all those deviations,
which is 5,888,960.2
This gives the variance, which is a calculation of how far the data points
deviate from the mean. But because we squared each deviation, the value of the
0.00020
0.00015
0.00010
Density
0.00005
0.00000
Density
25000 30000 35000 40000 45000 50000 55000 25000 30000 35000 40000 45000 50000 55000
(a) Salary (b) Salary
figure 6.3 (a) and (b) distributions of salaries having the same mean
but different standard deviations.
158 problem solving, decision making, and professional judgment
managers (the same as Figure 6.2 but graphed on a different horizontal scale),
and panel (b) displays the probability distribution of salary among Bulls-eye
assistant managers. Bulls-eye assistant managers have the same mean salary as
do Big-Mart assistant managers. Which population distribution has the greater
standard deviation?
The standard deviation tells how useful the mean of the sample of salaries is
in estimating the mean of the population of assistant managers’ salaries. If we
could look at a second number (after the sample mean) describing the distribu-
tion of salaries, the standard deviation would be a good choice.
You can visualize the bell shape of a normally distributed curve as an upside-
down bowl. Since the inside of the bowl is facing down, we say that there the
curve is concave downward. On the left and right sides of the curve, it is the
shape of a part of a right-side-up bowl. The inside of the bowl faces up, and this
part of the curve is called concave upward. An inflection point of a curve is a
point at which the curve switches from being concave up to concave down or vice
versa. In a normal distribution, there are two inflection points, and they are each
exactly one SD away from the mean. Using this fact is a good way to get a rough
idea of the SD when you have only a graph of the probability distribution.
0.00020
0.00015
Density
0.00010
0.00005
0.00000
the mean of the next one hundred salaries chosen at random? In the mean of
100, an extreme score is likely to be canceled out when combined with the other
99. On the other hand, it is quite possible that one randomly chosen salary will
be extreme.
So how far from the population mean is any sample mean likely to be? Rather
than take a hundred different samples, you can just use this formula to calculate
its standard error (SE)—
SD
SE =
N
—where SD is the standard deviation of the sample and N is the size of the
sample. The smaller the SE, the better an estimate our sample mean is of the
mean of the population.
Looking at the denominator, you’ll see that the larger the sample, the smaller
the standard error—though the SE only decreases by the square root of the size
of the sample. As we obtain more and more observations, the distribution of
the sample mean becomes tighter and tighter about the population mean, the
standard error decreases, and the precision with which a sample mean estimates
the population mean increases.
160 problem solving, decision making, and professional judgment
Why is the mean of a larger sample a more precise estimate of the population
mean than the mean of a small sample? Let’s go back to the discussion of cate-
gorical variables early in the previous chapter. Suppose that you didn’t know
anything about the mechanics of coin tosses or dice rolls: how much informa-
tion does one toss of a coin that lands on heads, or one roll of a die that lands on
four, provide about the chances of coins coming up heads or dice landing on
four? What about two tosses or rolls? What about a thousand? So, too, the mean
of the sample of many salaries provides a better estimate than the mean of a
sample of one, or even five.
Looking at the numerator, you’ll see that the smaller the standard deviation of
the sample, the smaller the SE. Recall that the standard deviation of a sample is
a measure of the dispersion of the items in the sample, in this case the disper-
sion of salaries. Which is more dispersed: a sample of the population of Big-Mart
assistant managers or a sample of Big-Mart employees (including everyone from
cleaning staff to executives)? And which population mean would you be better
able to estimate from their respective samples of one hundred: the population of
Big-Mart assistant managers or the population of all Big-Mart employees?
Notice that the formula for determining SE does not take account of the size
of the population from which the sample is taken. Our confidence in the rela-
tionship between the sample mean and the population mean is based solely on
the size and standard deviation of the sample and not on the proportion of the
sample to the population as a whole. If this seems counterintuitive, consider: as
long as a pot of soup is well stirred, one sip is all you need to taste the soup—no
matter what the size of the pot.
In our example, using the formula above, the SE of the sample mean is
$1820/√100 = $182.
Another common way to indicate our confidence in the estimate is to report
a range, called a confidence interval, around it that is likely (often with 95 percent
probability) to cover the true mean. (The confidence interval is the same as the
margin of error in a public opinion poll.)
Because the sampling distribution of the sample mean is normally distrib-
uted, we are 68 percent confident that the sample mean is within one SE of the
population mean, and 95 percent confident that it is within two SEs. The smaller
the SE, the smaller the confidence interval.
Thus, using our estimated SE of $182, we can be about 95 percent confident
that the mean score for the population is within $364 of our estimate:
$38,627 ± 2 * $182,
i.e., between $38,263 and $38,991. In other words, the estimate of $38,627 has
a margin of error of 2 * $182 = $364.3
3. More precisely, there is a 95 percent chance that the sample mean is within 1.96
standard errors of the population mean. A more precise estimate of the 95 percent
confidence interval for the population mean is therefore $38,627 ± 1.96 * $182.
scores, dollars, and other quantitative variables 161
Though it is generally true that one never knows the population mean and must
estimate it from the sample, Sears, Roebuck, and Co. vs. City of Inglewood (Los
Angeles Superior Court, 1955) provides a rare counterexample. The City of
Inglewood imposed a half-percent sales tax on sales made by stores to resi-
dents living within the city limits. Sears sued to recover taxes erroneously
imposed on nonresidents. Sears retained a statistician, who randomly sampled
33 out of the 826 business days during the period at issue, and examined sales
slips from each of those 33 days. Based on the mean overpayments on these
days, the expert estimated that the total overpayment was $28,250 with a 95
percent confidence interval between $24,000 and $32,400. The judge refused to
accept the estimate and required Sears to prove its refund claim on each indi-
vidual transaction rather than on the sample information. Subsequently, a
complete audit of all transactions over the entire 826 days was performed and
the actual refund amount was determined to be $26,750.22—well within the
confidence interval.
R. Clay Sprowls, The Admissibility of Sample Data Into a Court of Law: A Case
History, 4 UCLA L. REV. 222 (1957).
about this in the terms that we presented above: for each day, we compute an
estimate of the probability of a boy being born. This estimate is in fact the mean
of a set of observations, where each observation corresponds to a birth and is
either 1 (if a boy was born) or 0 (if a girl was born). The standard error of the mean
of 45 observations is less than that of the mean of 15 observations, so the confi-
dence interval on the percentage of boys born is tighter for the large hospital. Hence
we would expect more days with 60 percent boys from the smaller hospital.
The students’ responses are shown in Table 6.2. Only 20 percent of the
respondents answered correctly:
table 6.2 estimates of boy babies born in large and small hospitals
Response Frequency
Larger hospital 12
Smaller hospital 10
About the same (within 5%) 28
5. Eugene Borgida and Richard E. Nisbett, The Differential Impact of Abstract vs.
Concrete Information on Decisions, 7 J. Applied Social Psychology 258 (1977).
scores, dollars, and other quantitative variables 163
Null Hypothesis
0.00020
Density
0.00010
0.00000
Alternative Hypothesis
0.00014
Females Males
Density
0.00008
0.00002
6. The t-test considers the difference of $729 relative to the variance of the data, and
compares it to the t-distribution, a probability distribution like the normal distribution but
that varies according to the size of the sample.
scores, dollars, and other quantitative variables 165
words, under the null hypothesis of no salary difference in the population, there
would be a 3 percent chance that, in samples like ours, the mean salaries of the
two groups would differ by at least $729.
Step 5) State your conclusion. If the p-value is less than the alpha level, we
reject the null hypothesis in favor of the alternative. In our case, the p-value is
0.03. This means that, under the null hypothesis of no relationship between sex
and salary, only 3 percent of samples would be more extreme than ours. Since
this is less than 5 percent (the level of Type I error we were willing to tolerate) we
can reject the null hypothesis. We conclude that there is a relationship between
sex and salary.
So we have concluded that there is a statistically significant difference between
the salaries of male and female assistant managers at Big-Mart. But can we con-
clude that Big-Mart is discriminating based on sex in setting salaries? Perhaps
there is another variable that explains the difference. We explore this possibility
in Section 6.2.7.b below.
6.2 regression
7. Each state’s NAEP scores aggregated from individual student scores, using sam-
pling methods discussed in the preceding chapters. Tests are administered to a random
sample of students in schools that themselves are chosen to be representative of schools
and students in their states selected to participate. NAEP typically selects 3000 students in
approximately 100 schools in each state for each grade and subject.
166 problem solving, decision making, and professional judgment
235
MA
205 210 215 220 225 230
Average 4th Grade Reading Scores
NH VT
VA DE CT
ND MT MN ME
CO
WA WY PA NJ
SD OH NY
ID
UT IAKS NE
MO WI
KY
FL MD
TX IN MI
NCAR OR IL RI
GA WV
OKTN SC
AK
LA HI
AL
AZ NV NM
CA
MS
200
5 6 7 8 9 10 11 12 13 14 15
Per-Pupil Spending in $1000s
Nonlinear Relationships
Nonlinear relationships are also common. For example, the number and severity
of traffic accidents increases with the speed at which vehicles are traveling, but
the increase is geometric (i.e., based on a power of the speed)—rather than arith-
metic. In this case, the relationship would not be shown by a line, but by an
upward-sloping curve. Or consider the relationship between room temperature
and productivity in the workplace. A scatterplot of points with horizontal position
indicating temperature and vertical position indicating productivity might have
a ∩ (inverted U) shape, with low productivity values at both extremes and high
productivity in the middle.
In cases where the relationships between the variables are monotonic—that is,
the slope of the curve does not change direction—the variables sometimes can
be transformed to make their relationship linear. For example, the radius of a
circle and its area are not linearly related, but the radius is linearly related to the
square root of the area. Other functions commonly used to linearize a relation-
ship are the square, the logarithm, and the reciprocal.
235
MA
230
Average 4th Grade Reading Scores
NH VT
VA DE CT
225
ND MT MN ME
CO
WA WY PA NJ
SD OH NY
ID
UT IAKS NE
MO WI
220
KY
FL MD
TX IN MI
NCAR OR IL RI
215
GA WV
OKTN SC
AK
210
LA HI
AL
AZ NV NM
CA
205
MS
200
5 6 7 8 9 10 11 12 13 14 15
Per-Pupil Spending in $1000s
know that a state spends $10,000 per pupil, our best guess is that its students
would have an average NAEP score of about 220.
Often the response (such as achievement) is represented by the letter y and the
predictor (such as spending) by x. The line of best fit is then represented by the
regression equation, which is also the general equation for a line:
y = a + bx.
Because we can use this line to predict the response when we know the pre-
dictor, it is also called the prediction equation. In our example:
Predicted achievement = a + b * spending.
• b is the slope of the line—the amount the line goes up for each unit it goes
right. b is also called the coefficient of x. 1.5 is the slope of this regression
line. With an increase in spending of $1000, the predicted average score
increases by 1.5.
• x is the predictor, in this case spending.
• y is the predicted outcome, in this case achievement.
While x and y vary from state to state, a and b represent constant numbers.
If b is positive, the line slopes up, so the outcome increases when the predictor
increases. If b is negative, the line slopes down, so the outcome decreases when the
predictor increases. a and b can be calculated by any statistical computer program.
If the slope (b) equaled zero, we would have a horizontal line, which would
correspond to no relationship between the predictor and the outcome. In our
example, no matter how much was spent, we would not change our prediction of
achievement.
In our example, if spending increases by $1000, the predicted outcome y (achieve-
ment) increases by 1.5 points on the NAEP test. So our regression equation is
Predicted achievement = 205 + 1.5 * spending (in thousands).
For example, if a state spends $8000 per pupil, our best guess is that its fourth
graders will on average score 217 on the reading test. You can observe this by
looking at the regression line in Figure 6.7.
The difference between a data point’s observed outcome and its predicted
outcome is called the error of the prediction. In Figure 6.7, we can see that
California spends $8,417 per pupil, and its students score 206.5 on average. But
the line of best fit predicts that students in a state with $8,417 of spending per
pupil will achieve scores of 217.7 on average. The difference of 11.2 is the error
of that observation.
The line of best fit is the line that minimizes the sum of the squared errors
over all observations in the sample. (Squaring each error ensures that it is posi-
tive and weights those observations that are further away.)
6.2.2 Correlation
Another way of saying that two variables are linearly related is to say that they are
correlated. The correlation coefficient (or correlation), denoted r, of two variables
indicates the strength of the linear relationship.
r has the same sign (positive, negative, or zero) as b, so, like b, it is positive
when the line of best fit slopes up (when the outcome increases with an increase
in the predictor) and is negative when the line slopes down (when the outcome
decreases with an increase in the predictor). However, the magnitudes of b and
r are generally not equal.
• b is the amount by which y increases for a one-unit increase in x. Two
states that differ in spending by $1000 differ in predicted achievement by b.
170 problem solving, decision making, and professional judgment
235 NE OK
PA
MD DE
UT
FL
230
Average 4th Grade Reading Scores
RI
AK MA
LA CT
NM SC
VT
225
NJ
IL KS
NC
NV SD
220
CO AR
AZ
WV VA
NH
215
KY IN
MN TN
WA ND OR
WI
210
HI TX
NY OH
AL WY
MI MT
IA
205
CA MO
ID
GA
MS
200
ME
5 6 7 8 9 10 11 12 13 14 15
Per-Pupil Spending in $1000s
NJ
NY
CT
VT
RI
DE MA
ME
AK
230
Average 4th Grade Reading Scores
MDPA
WY
WI
OHMINH
IL
225
HI WV
INNE
MN
VA
MT
KS
GA
220
IA CA
NMOR
MO
CO
215
WA
ND
SCLA
TX
SD
210
KY
NC FLAR
AL
NVTN
205
MS
ID
OK
200
AZ
UT
5 6 7 8 9 10 11 12 13 14 15
Per-Pupil Spending in $1000s
of men in successive generations, Galton found that while tall fathers tended to
have reasonably tall sons, the sons were seldom as tall as their fathers. Likewise,
very short fathers had reasonably short sons, but the sons were unlikely to be as
short as their fathers. Because it seemed like the children’s heights were moving
away from the heights of their fathers, and towards the mean height, Galton
dubbed this phenomenon regression to the mean.
6.2.3.a Regression and Chance Galton came to understand that this process
is not due to a characteristic unique to heredity, but to the involvement of chance.
Regression to the mean is a consequence of the fact that some chance is involved
in determining the height of an individual, i.e., that r is not equal to 1 or -1. By
chance in this context, we mean anything that is not explained by the predictor
(father’s height), for example, mother’s height or nutrition—that is, chance com-
prises both factors that we don’t know about and factors that we know about but
are not considering.
To see how the involvement of chance produces regression to the mean, con-
sider a case in which chance alone determines the outcome. Imagine that you
telephone two numbers selected from the phone book at random, and ask the
people at the other end to tell you their height. The first person you speak to
reports his height as 6’2”. Do you expect the next person to be of similar height,
or closer to the average height of people in the phone book?
When chance is the only process involved, it is intuitive that an extreme
event—like a very tall person—is likely to be followed by something less extreme.
Moderate outcomes are just more probable. But the same phenomenon holds
even when there is a relationship between pairs of events, as there is between the
heights of fathers and sons. While heredity contributes to determining a son’s
height, so do several other factors. Having a tall father pulls the son’s height
toward the tall end, but chance pulls it toward the middle. Consequently, regres-
sion to the mean takes place.
Just as factors other than a father’s height determine the height of his son, so do
many factors besides a state’s per-pupil expenditures determine fourth grade read-
ing scores. Consider that states may differ in families’ average education and aver-
age income, and that these factors may also play a role in student achievement.
The winner’s curse refers to the phenomenon that, in an auction with many
bidders and high variance in their estimates of the value of the commodity, the
winning bidder will overestimate the value of the commodity and pay too much
for it. This may occur, for example, when oil companies bid for exploration
rights, or baseball teams bid for free agents, or companies compete to acquire
another company.
Richard Harrison and Max Bazerman link the winner’s curse to regression to
the mean.11 Imagine two normally distributed curves, one showing bidders’ esti-
mates of the value of the auctioned item and the other showing their actual bids.
Assume that the midpoint of the estimate curve is the actual value of the item.
The average bid will be lower than the average estimate, since people will bid less
rather than more than their estimate of the value of the item. The winning bid
will, by definition, be above the mean of bids, and will reflect one of the highest
estimations of the item’s value.This concept is illustrated in Figure 6.11.12
– –
B E
Variables Assumptions
–
E = Estimates 1. True value ≈ E
B = Bid 2. True value will be equal for
D = (Amount of discounting) all bidders
=E–B
11. J. Ricardo Harrison and Max H. Bazerman, Regression to the Mean, Expectation
Inflation, and the Winner’s Curse in Organizational Contexts, in Negotiation as a Social
Process 69 (Roderick Moreland Kramer and David M. Messick eds., 1995). They draw on
an earlier paper by Bazerman and William E. Samuelson, I Won the Auction But Don’t
Want the Prize, 27 Journal of Conflict Resolution 618–634 (1983).
12. Id.
scores, dollars, and other quantitative variables 175
the mean achievement, we average the squared errors, where the errors are the
vertical distances of the observations from the line of best fit; that is, the differ-
ences between each sample member’s (state’s) observed outcome and its pre-
dicted outcome based on its spending (see Figure 6.8). These errors are generally
smaller than the errors from the sample mean, so the average of their squares
will be smaller. In other words, the variance around the line of best fit is smaller
than the variance around the mean. This is because the regression line captures
some of the systematic changes in one variable as a function of the other.
The difference between these two variances is referred to as the variance
accounted for by or explained by the predictor, here spending. In other words, the
total variance of the outcome consists of two parts:
This is called the analysis of variance (ANOVA) equation. The better the pre-
dictor, the larger the proportion of the total variance it explains. This proportion
is therefore used as a measure of the usefulness of a predictor and is called the
coefficient of determination:
r r2
Aspirin and reduced risk of heart attack 0.02 0.0004
Chemotherapy and surviving breast cancer 0.03 0.0009
Calcium intake and bone mass in post-menopausal women 0.08 0.0064
Ever smoking and subsequent incidence of lung cancer
within 25 years 0.08 0.0064
Alcohol use during pregnancy and subsequent premature
birth 0.09 0.0081
Effect of nonsteriodal anti-flammatory drugs (e.g., ibuprofen)
on pain reduction 0.14 0.0196
Gender and weight for U.S. adults 0.26 0.0676
ECT for depression and subsequent improvement 0.29 0.0841
Sleeping pills and short-term improvement in chronic
insomnia 0.3 0.09
Age and episodic memory 0.33 0.1089
Elevation above sea level and lower daily temperatures in
the U.S. 0.34 0.1156
Viagra and improved male sexual functioning 0.38 0.1444
Age and reasoning 0.4 0.16
Weight and height for U.S. adults 0.44 0.1936
Age and speed 0.52 0.2704
Gender and arm length 0.55 0.3025
Nearness to equator and daily temperature in the U.S. 0.6 0.36
Gender and height for U.S. adults 0.67 0.4489
13. Paul Slovic and Ellen Peters, The Importance of Worldviews in Risk Perception, 3
Risk Decision and Policy 165–170 (1998), citing Robert Rosenthal, How Are We Doing in
Soft Psychology, 45 American Psychologist 775–7 (1990).
178 problem solving, decision making, and professional judgment
y = a + bx
linear relationship between a state’s per-pupil spending and its fourth graders’
reading achievement.
14. Often, when the outcome is categorical, the more advanced method of logistic
regression is used instead of ordinary regression.
15. See Adam Liptak, Death Penalty Found More Likely When Victim Is White, New
York Times, Jan. 8, 2003.
16. John J. Donohue III and Steven D. Levitt, The Impact of Legalized Abortion on Crime,
116 Quarterly Journal of Economics 136–44 (2001).
17. See Ian Ayres and John J. Donohue III, Shooting Down the “More Guns Less Crime”
Hypothesis, 55 Stanford Law Review 1193 (2001); Steven D. Levitt and Stephen J. Dubner,
Freakonomics 130–134 (New York: HarperCollins Publishers, 2005).
182 problem solving, decision making, and professional judgment
an analysis that showed that, although White, Black, and Hispanic drivers
are stopped by police at about the same rate, Hispanic drivers were searched
or had their vehicles searched by the police 11.4 percent of the time and
Blacks 10.2 percent of the time, compared with 3.5 percent for White driv-
ers; and Blacks and Hispanics were also subjected to force or the threat of
force more often than Whites.18
• Many Supreme Court Justices have urged Congress to increase the salaries
of federal judges, arguing (in Justice Alito’s words) that “eroding judicial
salaries will lead, sooner or later, to less capable judges and ultimately to
inferior adjudication.” In A Skeptical Response to the Judicial Salary Debate,
Stephen Choi, G. Mitu Gulati, and Eric Posner used multiple regression
analyses to find the absence of any correlation between salaries and perfor-
mance in the highest courts of the fifty states. For proxies for performance,
the scholars measured effort in terms of the number of pages of opinions
published, skill by citations by out-of-state judges and law journals, and
independence by disagreement with fellow judges with the same political
party affiliation.19
While multiple regression analysis has great power, it is subject to many of
the same limitations as simple regression analysis: multiple regression can mea-
sure the strength only of relationships that are linear or that can be “linearized”;
it is sensitive to outliers, and (unless the explanatory variables are carefully
manipulated in an experiment, instead of just observed) it cannot prove the exis-
tence of causal relationships. As we will discuss in Section 7.2.1, there could be
a so-called confounding variable that you didn’t measure that influences the
ones you did measure. To eliminate confounding variables, one must conduct
randomized controlled experiments (discussed in Sections 7.3.1 and 7.3.5).
6.3 glossary
Central tendency Mean (average), Median (middle), Mode (most common)
measures
Coefficient of The amount of total variance explained by the predictor
determination (r2)
Confidence A range of numbers in a sample that are, say, 95 percent likely
Interval to encompass the population mean
Correlation A linear relationship between two variables
Correlation The strength of the linear relationship between two variables
coefficient (r)
18. Eric Lichtblau, Profiling Report Leads to a Demotion, New York Times, Aug. 24, 2005.
19. Http://papers.ssrn.com/sol3/papers.cfm?abstract_id=1077295.
scores, dollars, and other quantitative variables 183
Dispersion How spread out the data are. (See Variance and Standard
Deviation)
Empirical The distribution of the sample
distribution
Independence Where the value of the so-called predictor or dependent
variable provides no information about the outcome,
response, or independent variable
Line of best fit A straight line that best describes the relationship between the
predictor and response variables
Multiple regression Linear regression analysis that takes into account more than
analysis one predictor variable
Normal A specific bell-shaped curve with the mean, median, and
distribution mode all at the center point
Regression A technique for determining correlation between two or more
analysis continuous variables
Regression to the The tendency of extreme predictor variables to produce
mean responses that are less extreme (i.e., closer to the mean)
Sampling The distribution of the sample mean
distribution
Scatterplot A collection of datapoints showing the predictor and outcome
values for each item
Standard deviation A measure of the dispersion or variance of a sample. Square
(SD) root of variance
Standard error (SE) The standard deviation of the sampling distribution
Variance Squared deviation of data points from the mean
184 problem solving, decision making, and professional judgment
6.4 appendix
big-mart employee salaries by months of experience and sex
The primary use of statistics in science, policy, and law is to throw light upon the
causal relationships that might exist between two or more variables. In this chap-
ter we consider some of the ways in which statistics interacts with scientific
reasoning, as well as some psychological findings that suggest caution in moving
from statistical to causal conclusions. Much of the chapter focuses on evaluating
the outcomes of policy interventions, and we also consider efforts to understand
underlying social phenomena of concern to policy makers.
Recall that the introduction to Part 2 introduced the “lens” model, under
which the real world can only be perceived through multiple fallible indicators.
This chapter is concerned largely with identifying, separating, and giving appro-
priate weight to those indicators in order to establish the existence (or nonexis-
tence) of causal relationships. One way in which this is done is through evaluation
activities aimed at assessing the conceptualization, design, implementation and
utility of social intervention programs.1
There are a number of different ways in which we use evaluation in policy
making. Oftentimes, policy makers will oversee the design of a particular social
program or select from among different types of interventions that use different
assessment designs. Thus, it is important for policy makers to understand poten-
tial design flaws and the advantages and disadvantages of different types of
program evaluation. More importantly, perhaps, is the ability for a policy maker
to understand how to analyze impact assessments, compare different evalua-
tions, think critically about the utility of such evaluations and make judgments
about the inferences that can be drawn from them.
Policy makers often design, support, and evaluate health and social programs
designed to improve people’s lives. How can they learn whether the programs
actually achieve their intended effects?
The critical issue in impact evaluation is whether or not a program produces
more of an effect or outcome than would have occurred either without the inter-
vention or with an alternative intervention. Common among all approaches to
evaluating impact are the goals of establishing (1) that a particular intervention
is correlated with the outcome, and (2) that the intervention causes the outcome—
for example, establishing that the rate of malaria infections in a village goes
down (1) when residents use bed nets and (2) because they use bed nets.
Imagine that we conducted a study of drowning deaths in Florida and discov-
ered that the number of drowning deaths in a given month was highly correlated
with ice cream sales for that month.2 We could develop a number of possible
A man rode the Fifth Avenue bus to work every day. And every day, when the bus
stopped at Forty-Second Street, he opened the window and threw the sports sec-
tion of the newspaper into a litter basket. After many years, the bus driver asked
why he did this, and the man replied: “I do it to keep the elephants off Fifth
Avenue.” “But there aren’t any elephants on Fifth Avenue,” said the bus driver, to
which the man responded: “See, it works!”
explanations for this correlation. (a) It could be that eating ice cream causes
drowning, with more people developing cramps in the water and thus not being
able to swim. (b) It could also be that drowning causes people to buy ice cream,
as friends and family of the victims seek to console themselves by eating. (c) Or
it could be that there is some other variable that causes both drowning and ice
cream consumption—for example, the average temperature in a given month.
2. This example comes from Ewart Thomas, to whom the authors are also grateful for
less direct influences.
interpreting statistical results and evaluating policy interventions 187
Further testing could tease these explanations apart, but the observation of a cor-
relation is consistent with all three causal scenarios.
A causal diagram represents the causal process that results in a correlation of
interest. The three scenarios listed above as possible explanations of the correla-
tion between drowning deaths and ice cream sales are diagramed in Figure 7.2.
A solid arrow indicates a hypothesized causal relationship in the arrow’s direc-
tion. A dashed line, on the other hand, emphasizes that, while a correlation exists
between two variables, we are not hypothesizing any causation. When one vari-
able, such as temperature, is related to two other variables (drownings and ice
cream sales), a correlation appears between those two variables. But correlation
does not imply causation.
Drownings
Scenarios (a) and (b) of Figure 7.2 show alternative causal hypotheses.
Scenario (c) implies that temperature may be a confounding or lurking variable.
Confounding is the situation in which a variable you are not studying (here tem-
perature) is related to both the variables you are studying (here ice cream sales and
drownings). Data show a correlation between the variables you are interested in,
but a third lurking variable may actually be causing both.
Consider, for example, a health survey of 21,101 Finns that indicated that pet
owners were more likely to suffer from illnesses that included high blood pres-
sure, high cholesterol, ulcers, depression, kidney disease, and being overweight.
Did owning a pet cause these diseases? Did having these diseases cause people
to own pets? Alternatively:
Pets seem to be part of the lives of older people who have settled down and
experienced an increase in the number of illnesses, whereas young healthy
single people have no time, need, or possibility for a pet. Associations of pet
ownership with disease indicators were largely explained with socio-
demographic factors. Characteristics were those that bring forward the back-
ground of poor health in epidemiological investigations: male gender, low level
of education, life without a couple relationship, and poor social standing.3
3. Leena K. Koivusilta and Ansa Ojanlatva, To Have or Not To Have a Pet for Better
Health?, 1 PLoS ONE e109 (2006), doi:10.1371/journal.pone.0000109 (2006).
188 problem solving, decision making, and professional judgment
Establishing what the outcome actually was and its correlation with the interven-
tion is often hard enough. For example, in attempting to evaluate the outcome of
the bed net program, it may be difficult to gather the data necessary to establish
a baseline of malaria infections and then to track the number of infections over
time. Establishing causation is usually more difficult, because even if there is a
change, it may be due to factors other than the intervention being studied (intro-
duction of bed nets): perhaps the weather was drier than usual, thus producing
fewer mosquitoes, or perhaps other antimalaria interventions were introduced at
the same time.
A gross outcome is all the changes that occur from time point A, before pro-
gram implementation, and time point B, after implementation.
Gross Outcome = Effects of intervention (net outcome) + Effects of extrane-
ous and confounding factors + Design effects.4
Implementation of bed net program
A---------------------------------------------------------------------> B
(no bed nets) (bed nets fully distributed)
In the above intervention, the gross outcome would consist of all changes that
occurred during the implementation period, including the changes in malaria
infection rates attributable to the program, as well as changes resulting from
other factors, such as a change in the weather or other variables that could have
also reduced malarial infection rates.
What we ultimately care about is the net outcome, or the effects that are attrib-
utable to the intervention itself. As we see above,
Net Outcome = Gross Outcome – Effects of extraneous and confounding
factors – Design effects.
Therefore, it is important to identify and, if possible, measure the effects that
are not related to the intervention that may influence the gross outcome.
5. Id. 194.
190 problem solving, decision making, and professional judgment
the same object. Some measurements, such as the measurement of height and
weight through standard instruments, have a high degree of reliability. Others,
such as surveys that rely on subjects’ responses to written or oral questions, are
more prone to unreliability because of the inherent ambiguity in language and
interviewer differences in their administration of the survey. In most cases, it is
not possible to eliminate unreliability completely, although it is possible to make
adjustments in results that take unreliability into account.
Even if a measuring procedure produces consistent results and is therefore
deemed reliable, it will not be valid if it does not measure what it is intended to
measure.
Measurement validity depends on whether there is a consensus that the mea-
sure is valid among the appropriate stakeholders, such as members of the scien-
tific community. For example, the effectiveness of the bed net program could be
measured by a number of alternative measures, including (1) self-reported
malarial infection rates, (2) hospital visits related to malaria, or (3) professional
surveys of malarial infection rates. A good outcome measure is generally one
that may be feasibly employed, given the constraints of time and budget, and
that is directly related to the goals of the program.
The process of conducting an evaluation in and of itself can also influence the
outcome of the evaluation. This can happen through what is known as the
Hawthorne or placebo effect. The Hawthorne Effect stems from a famous experi-
ment, in which efforts to discover whether changes in factory lighting intensity
would improve worker productivity disclosed that any change had that effect.6
Researchers concluded that the effect was the result of the experiment itself—
workers interpreted their participation in the experiment as a sign that the firm
was interested in their personal welfare, and thus were motivated to work harder.
In medical experiments, the Hawthorne Effect is known as the placebo effect, and
refers to the fact that subjects may be as much affected by the knowledge that they
are receiving treatment as by the treatment itself. For example, in some experi-
ments testing the effectiveness of pain relievers, the control groups receiving
sugar pills (unbeknownst to them) reported a decrease in pain levels.
The Pygmalion Effect refers to the fact that expectations can also influence
outcomes. An experiment by Robert Rosenthal and Lenore Jacobson found that
if teachers were led to expect enhanced performance from some children rather
than others, then those children did indeed perform better.7 The experiment
confirmed the researchers’ hypothesis that biased expectancies could essentially
affect reality and create self-fulfilling prophecies as a result. Policy makers
and evaluators should be especially aware of such effects where treatment is
Confidence in the results of evaluations and other empirical studies not only
requires identifying various confounding factors and potential design flaws, but
using controls. Here we discuss four important kinds of controls:
• Randomized controls—the “gold standard” of evaluation, in which partici-
pants are randomly assigned either to the treatment (or experimental)
group—those who receive the intervention—or to a control group.
• Constructed controls—a major form of quasi-experimental evaluation, in
which the treatment and control groups are constructed to be similar to
each other.
• Reflexive controls—in which one looks at the same population before and
after the intervention.
• Statistical controls—in which multiple regression analysis is used to con-
trol for suspected confounding factors.
The approaches other than randomized controls are sometimes called obser-
vational studies, because the experimenters merely observe the populations they
are studying.
had an impact because the random assignment controlled for other factors. The
main drawbacks to randomized controlled studies are that they are expensive to
perform and difficult to replicate.
The U.S. Food and Drug Administration requires that most new pharmaceu-
tical products be subject to randomized controlled studies before being intro-
duced into the market. Participants in clinical trials receive either the new drug
(the treatment) or a control (a placebo, or conventional treatment if there is one),
and researchers analyze whether the health of those receiving the new drug
improves more—or less—than the health of those in the control group.
Randomized controlled studies are not mandated for social interventions,
and they are seldom done due to time and expense constraints. But when feasi-
ble, they can provide essential insights into the actual effects of social programs.8
In 2005, the U.S. Department of Health and Human Services published the
National Head Start Impact Study, a randomized controlled study analyzing the
impact of the Head Start program on a variety of measures of children’s social
and cognitive development.9 The study employed random controls and found
that Head Start had limited effects on the social and cognitive development of
participating children. This conclusion contradicted many previous Head
Start studies, which had relied on constructed or statistical controls.
8. The Coalition for Evidence-Based Policy maintains an excellent Web site on social
programs that work: http://www.evidencebasedprograms.org.
9. U.S. Department of Health and Human Services, Administration for Children and
Families, “Head Start Impact Study: First Year Findings: (Washington, DC: U.S.
Department of Health and Human Services, 2005), available at http://www.acf.hhs.gov/
programs/opre/hs/impact_study/.
interpreting statistical results and evaluating policy interventions 193
10. Victor G. Cicerelli et al., The Impact of Head Start : An Evaluation of the
Effects of Head Start on Children’s Cognitive and Affective Development.
(Athens: Westinghouse Learning Corporation and Ohio University, 1969).
11. Debbie L. Hahs-Vaughn and Anthony J. Onwuegbuzie, Estimating and Using
Propensity Score Analysis with Complex Samples, 75 Journal of Experimental Education
31 (2006).
12. Multiple studies conclude that matching only performs well in data that satisfy
three criteria: 1) treatment and comparison groups come from identical sources, 2) indi-
viduals in both groups live in the same geographical area, and 3) “the data contain a rich
set of variables that affect both” the outcome and the probability of being treated. Jeffrey
Smith, and Petra Todd, Does Matching Overcome LaLonde’s Critique of Nonexperimental
Estimators?, 125 Journal of Econometrics 309 (2005).
194 problem solving, decision making, and professional judgment
13. The literature on this theory is nicely summarized in Wikipedia, s.v. “Fixing Broken
Windows,” http://en.wikipedia.org/wiki/Fixing_Broken_Windows (last visited June 30,
2008).
14. Jane Brody, New Therapy for Menopause Reduces Risks, New York Times, Nov. 18,
1994.
15. Gina Kolata with Melody Petersen, Hormone Replacement Study: A Shock to the
Medical System, New York Times, July 10, 2002.
16. Susan M. Love, Preventive Medicine, Properly Practiced, New York Times, July 16,
2002.
interpreting statistical results and evaluating policy interventions 195
frequency of doctor visits, and other less tangible aspects of attention to one’s
health. Variables like these, even if thought of, may be too vaguely defined to
control for completely. The suspected confounding in these studies is repre-
sented in Figure 7.3.
Take HRT
Healthy lifestyle,
regular doctor
visits
Low risk of
heart
disease
17. Kevin Arceneaux, Alan S. Gerber, and Donald P. Green, 2004-04-15 “Comparing
Experimental and Matching Methods using Two Large-Scale Field Experiments on
VoterMobilization” Paper presented at the annual meeting of the The Midwest Political Science
Association, Palmer House Hilton, Chicago, Illinois Online 2009-05-26 from http://www.
allacademic.com/meta/p83115_index.html.
196 problem solving, decision making, and professional judgment
18. Gordon C. S. Smith and Jill P. Pell, Parachute Use to Prevent Death and Major
Trauma Related to Gravitational Challenge: Systematic Review of Randomised Controlled
Trials, 327 British Medical Journal 1459–61 (2003).
interpreting statistical results and evaluating policy interventions 197
employment discrimination and toxic torts of the sort illustrated in the Big-Mart
and Terra Nueva examples are typical. Thus, the lawyer’s or policy maker’s
task—or the statistician’s task under their guidance—is to determine correlation
and often to attribute causation based on the relationship of plausible indepen-
dent variables (the insulation in a tenant’s apartment) with a dependent variable
of interest (the presence of illness).
19. Douglas Martin, Joan McCord, Who Evaluated Anticrime Efforts, Dies at 73, New York
Times, March 1, 2004.
198 problem solving, decision making, and professional judgment
services, and summer camp to high-risk boys with a control group of similar
youths. She found that boys in the program were more likely to become crimi-
nals, have employment and marital problems, and become alcoholics than those
in the control group. The evidence contradicted not only the expected outcome
but also the participants’ own belief that they had benefited from the program.
(Dr. McCord hypothesized that the boys in the treatment group may have felt
that they were given the attention because something was wrong with them,
making it a self-fulfilling prophecy—the Hawthorne Effect at work.) She found
similar paradoxical effects from a “just say no” drug education program and
from Scared Straight, a program designed to deter young people from following
a criminal path.
Thus, even (perhaps especially) when intuitions strongly support a particular
intervention, it is important to evaluate its actual impact. In this section, we look
at some cases where evaluation has demonstrated pretty conclusively that an
intervention works or that it doesn’t work and where it has been inconclusive.
The Good. The Abdul Latif Jameel Poverty Action Lab (J-PAL), a research
center at the Massachusetts Institute of Technology, specializes in conducting
randomized controlled studies in developing countries. J-PAL randomly assigned
girls in Kenya either to a group that was promised merit-based scholarships and
a cash grant for school supplies if they scored well on academic exams, or to a
control group. It turned out that just being eligible for scholarships led the girls
to think of themselves as “good students” and led to higher academic grades. In
fact, both student and teacher attendance improved in eligible schools, and even
boys showed improved test scores.20
In 2004–2005, SRI International employed a cross-sectional approach in eval-
uating the outcomes of the Bay Area KIPP (Knowledge is Power Program) char-
ter schools, whose “pillars” include high expectations, parental choice, long
school days, autonomous school leaders, and a relentless focus on results.
(Although a randomized controlled study is now under way, KIPP and its
sponsors also wanted this early, formative assessment of its impact to detect any
20. Michael Kremer, Edward Miguel, and Rebecca Thornton, “Incentives to Learn”
(NBER Working Paper No. 10971, National Bureau of Economic Research, 2004, http://
www.nber.org/papers/w10971). Another experiment by J-PAL demonstrates the failure of
an intervention to achieve its intended goals, though it certainly did some good along the
way. In a randomized controlled study to determine whether reducing the incidence of
parasitic worms among children could lead to better attendance and, consequently, higher
test scores, J-PAL administered low-cost deworming drugs to children at randomly chosen
schools in a poor, densely settled farming region near Lake Victoria. The drugs worked
well: after one year only 27 percent of children in treatment schools had moderate to heavy
worm infections, compared with 52 percent of children in control schools. The effects on
schooling, though, were minimal at best: though attendance rates in treatment schools
were 7 to 9 percent higher than in control schools, the test scores of children in interven-
tion and control schools were not significantly different.
interpreting statistical results and evaluating policy interventions 199
obvious need for midcourse corrections.) Using two standard state tests, the SAT
and CST, experimenters tracked KIPP students’ achievement over two years and
compared their achievement with that of students in comparable California
schools. The report explains the limits of the methodology:
In an ideal world we would make two comparisons as a basis for determining
if KIPP students are performing better than if they were not attending KIPP
schools. The first would be to compare achievement growth of KIPP students
to their growth trajectory prior to KIPP. However, this requires individual
student data for several years prior to enrollment in KIPP that is not publicly
available. . . . The second would be to compare KIPP students’ achievement
with that of a comparison group of students, defined based on KIPP school
waitlists (which do not yet exist), or matching students in the district based on
demographic characteristics. Again, this comparison requires access to
student-level data for both KIPP and non-KIPP students.21
With these limitations, the report noted:
The spring 2005 CST data indicate that the overall percentage of students
performing at a proficient level or above is consistently higher for KIPP
schools than for comparable schools in the district—in some cases dramati-
cally so. Similarly, when comparing students in KIPP schools to all students
in the state, more fifth graders in two of the five KIPP schools scored at or
above proficient in ELA (English language arts) than students statewide; in
three of five KIPP schools, the percentage of fifth-grade students who scored
at or above proficient in math was higher than the state average. Likewise, in
three of the four KIPP schools with sixth-grade scores, a higher percentage of
sixth-grade students reached proficiency in math and ELA compared to the
state as whole. In the one KIPP school with seventh-grade scores, the percent
proficient in both ELA and math exceeded the state average.
Although not conclusive, the size of the differences was great enough to
strongly support the hypothesis that the KIPP approach made a real difference
in children’s academic outcomes—sufficient to justify continuing and even
expanding the program while conducting more precise studies.
The Bad. The next best thing to learning that a social intervention succeeds, is
determining conclusively that it does not succeed—so that funders will seek
better options rather than pouring money down the drain. A famous example of
a demonstrably ineffective intervention is the Drug Abuse Resistance Education
(DARE) program, which sought to prevent youth substance abuse through
classroom instruction. Randomized controlled studies of the DARE program
21. SRI International, Bay Area KIPP Schools, “A Study of Early Implementation:
First Year Report 2004–05” (Menlo Park, Calif.: SRI International, 2006), available at
http:// policyweb.sri.com/cep/publications/KIPPYear_1_Report.pdf.
200 problem solving, decision making, and professional judgment
consistently showed that students in treatment and control groups had the same
rates of both short- and long-term drug use.22
In another case, the evaluation firm Mathematica Policy Research Inc. was
commissioned to conduct a randomized controlled evaluation of federal absti-
nence education initiatives designed to prevent teen pregnancy. Mathematica
worked with four different states to randomly assign schools either to receive or
not to receive abstinence education, and then analyzed students’ self-reported
sexual activity rates. The results, released in 2007, show that students whose
schools received abstinence programs were just as likely to be sexually active as
students whose schools did not.23 James Wagoner, the president of Advocates for
Youth, said of the results: “After 10 years and $1.5 billion in public funds these
failed abstinence-only-until-marriage programs will go down as an ideological
boondoggle of historic proportions.”24
The Inconclusive. In 1997, the New York City Voucher Experiment randomly
assigned two thousand low-income families with K–4 students to either a treat-
ment group or a control group. Families in the treatment group received school
vouchers for private-school tuition, worth $1400 per child per year, for four
years,25 while families in the control group did not receive vouchers. After three
years, researchers administered math and reading tests to students in each group
and analyzed the results. It turned out that vouchers did not significantly affect
children’s test scores in the aggregate, though they had a small positive impact
on the test scores of African Americans.26
The voucher experiment did not show that vouchers are ineffective. Rather,
methodological problems, especially the small sample size, made it hard to draw
a conclusion one way or the other.27 Given the difficulty of designing and imple-
menting studies of social interventions, you should be prepared for inconclusive
22. Cheryl L. Perry et al., A Randomized Controlled Trial of the Middle and Junior High
School D.A.R.E. and D.A.R.E. Plus Programs, 157 Archives of Pediatrics and Adolescent
Medicine 178–184 (2003).
23. Christopher Trenholm et al., “Impacts of Four Title V, Section 510 Abstinence
Education Programs: Final Report,” report prepared by Mathematica Policy Research Inc.
for the U.S. Department of Health and Human Services, April 2007.
24. Advocates for Youth, “10-Year Government Evaluation of Abstinence-Only
Programs Comes Up Empty,” news release, April 13, 2007, http://www.advocatesforyouth.
org/news/press/041307.htm.
25. Social Programs That Work, “New York City Voucher Experiment,” http:// www.
evidencebasedprograms.org/Default.aspx?tabid=143 (last visited June 30, 2008).
26. See Alan Krueger and Pei Zhu, Another Look at the New York City SchoolVoucher
Experiment, 47 American Behavioral Scientist 658–99 (2003); and Institute of
Education Sciences, National Center for Education Evaluation and Regional Assistance,
“Evaluation of the DC Opportunity Scholarship Program: Impacts After One Year,”
http://ies.ed.gov/ncee/pubs/20074009 (last visited June 30, 2008).
27. William G. Howell, book review, Data Vacuum, 2 Education Next (2002), available
at http://www.hoover.org/publications/ednext/3366891.html.
interpreting statistical results and evaluating policy interventions 201
results much of the time. This can happen, for example, because samples that
were originally of reasonable size diminish as a result of attrition—for example,
families in the control group move away and the experimenters lose track of them.
Or the control group coincidentally could receive some interventions from another
source—whether similar to or different from the “treatment” group.28
Unfortunately, many negative and inconclusive results never see daylight—-
the former, for reasons of motivation; the latter, for lack of interest. There is also
a strong bias in many scholarly fields against publishing statistically insignifi-
cant results. But even knowing about inconclusive results is useful, since a sub-
sequent study may be able to learn from the first and remedy some of the defects
that prevented a clearer outcome.
28. See, for example, Paul T. Decker, Daniel P. Mayer, and Steven Glazerman, “The
Effects of Teach for America on Students: Findings from a National Evaluation” (Princeton,
N.J.: Mathematica Policy Research Inc., 2004), available at http://www.mathematica-mpr.
com/publications/pdfs/teach.pdf.
29. Donald Campbell and H. Lawrence Ross, The Connecticut Crackdown on Speeding:
Time-Series Data in Quasi-Experimental Analysis, 33 Law & Society Review 33–54 (1968).
202 problem solving, decision making, and professional judgment
320
310
300
290
280
BEFORE AFTER
CRACKDOWN CRACKDOWN
(1955) (1956)
30. See Gina Kolata, Melanoma Is Epidemic. Or Is It, New York Times, Aug. 9, 2005.
“They found that since 1986, skin biopsies have risen by 250 percent, a figure nearly the
same as the rise in the incidence of early stage melanoma. But there was no change in
the melanoma death rate. And the incidence of advanced disease also did not change, the
researchers found.”
31. Tina Kelley, An Autism Anomaly, Partly Explained, New York Times, Feb. 18, 2007;
Graham Lawton, The Autism Myth, New Scientist, Aug. 13, 2005.
interpreting statistical results and evaluating policy interventions 203
325
300
275
250
225 TREATMENT
200
32. Matthew Wald, Errors in the Air and a Battle on the Ground, New York Times,
Aug. 13, 2005.
204 problem solving, decision making, and professional judgment
17 Connecticut
16 Control States
15
14
FATALITY RATE
13
12
11
10
(Why do you suppose the combined data for those four states is much smoother
than for Connecticut?).
Although the comparison lends some support to the Governor’s causal
hypothesis, Campbell and Ross mention the possibility that there may have been
some diffusion—a tendency for the intervention in the experimental group to
affect the control group. Also, there is some doubt whether the differences are
statistically significant.33
This underscores one of the central points of this chapter: purported causal
relationships must be viewed with caution. One must be wary about making
decisions based on evidence derived from observational or quasi-experimental
studies. While randomized controlled experiments are subject to less potential
errors and do a better job of proving causality, such experiments are rarely
available. Thus, the policy maker must use common sense and basic statistical
knowledge to identify potential confounding factors and design effects that could
influence the observational study’s outcome and make decisions about causality
accordingly.
33. The analysis of significance in this context is complex. Using a one-tailed test, G.V.
Glass, found significance at somewhere between p < 0.05 and p < 0.07. Gene V. Glass,
Analysis of Data on the Connecticut Speeding Crackdown as a Time-Series Quasi-Experiment,
55 Law & Society Review 55–76 (1968).
interpreting statistical results and evaluating policy interventions 205
7.5.2 Meta-Analysis
The Terra Nueva, Big-Mart, and education examples in the preceding chapters
showed statistically significant effects of sufficient magnitude to be of legitimate
interest to various stakeholders. But it is possible for a real effect to be masked
because the sample is too small. Apart from applying a less demanding signifi-
cance level—which trades off Type II for Type I errors—is there anything to be
done? In situations where a number of experiments have been done to study
essentially the same phenomenon, medical and social scientists sometimes use
the statistical technique of meta-analysis to aggregate and analyze information
from a variety of existing studies. Meta-analysis can increase the effective sample
size and thus reveal an effect that was not statistically significant in a single
study.
34. See generally Jeremy A. Blumenthal, Meta-Analysis: A Primer for Legal Scholars, 80
Temple L. Rev. 201, 209–10 (2007); John E. Hunter and Frank L. Schmidt, Cumulative
Research Knowledge and Social Policy Formulation: The Critical Role of Meta-Analysis, 2
Psychol. Pub. Pol’y & L. 324, 342–43 (1996).
206 problem solving, decision making, and professional judgment
35. http://www.pubmedcentral.nih.gov/articlerender.fcgi?artid=1694752.
36. For an excellent review of the pros and cons of meta-analysis, as well as description
of its techniques, see Blumenthal, supra.
interpreting statistical results and evaluating policy interventions 207
experiments on judgment and decision making. For this reason alone, it is worth
spending some time considering the value and limitations of such experiments.
This will also provide an opportunity to review some of the concepts developed
in the preceding chapters.
37. The example is taken from David W. Martin, Doing Psychology Experiments
(5th ed. Belmont: Thomson Brooks/Cole, 2000).
208 problem solving, decision making, and professional judgment
38. For example, on March 2, 2005, the Food and Drug Administration required that
the cholesterol drug Crestor be relabeled to add a caution that starter doses should be
reduced in Asian Americans and some other patients. A clinical trial found that levels of
Crestor in Asian patients were double those of Caucasians taking the same dose, increas-
ing the chance of muscle damage. Asians, it seemed, did not metabolize the drug at as
rapid a rate, causing it to stay in their systems longer.
39. See also Charles R. Plott and Kathryn Zeiler, The Willingness to Pay/Willingness to
Accept Gap, the “Endowment Effect,” Subject Misconceptions and Experimental Procedures for
Eliciting Valuations, 95 American Economic Review 530–45 (2004), who argue that a
major series of laboratory experiments were infected by confounding variables.
interpreting statistical results and evaluating policy interventions 209
the real world. 40 Law professor Chris Guthrie’s defense of the theory addresses
issues that apply to many other psychological experimental findings:41
Critics typically worry about three external validity factors: the subjects who
participate in the experiments, the lack of incentives they face, and the sim-
plicity of experimental designs versus the complexity of decision making in
the real world.
Subjects. Critics often contend that experimental work lacks external valid-
ity because the participants in these studies are typically undergraduates
(often enrolled in introductory Psychology classes). Although undergraduates
differ from other members of the population in terms of age, education level,
life experience, and so on, psychologists have found that they are in fact a
fairly good proxy for “real people.” Generally speaking, expert decision makers
exhibit the same decision-making patterns as undergraduates. Moreover,
many of the participants in the prospect theory studies identified in this arti-
cle were not undergraduates but rather adults with expertise relevant to the
domain being studied. For example, law students and federal magistrate
judges participated in some of the litigation studies; licensed physicians par-
ticipated in the medical treatment study; executive MBA students participated
in some of the tax studies; and corporate managers participated in some of
the managerial decision-making studies.
Incentives. When making decisions in the real world, people generally
have incentives to “get it right.” Some critics suggest that this could mean that
real-world decision makers are more “rational” than participants in experi-
mental studies. Although real-world incentives might induce decision makers
to take more care with their decisions, they do not appear to induce rational
decision-making patterns. In one rather impressive illustration of this, Steven
Kachelmeier and Mohamed Shehata conducted decision-making studies in
China where the researchers could afford to offer substantial monetary incen-
tives relative to local salaries. In one experimental condition, for instance, the
researchers were able to offer subjects an amount roughly comparable to
three months’ worth of income. Despite these substantial monetary incentives,
Kachelmeier and Shehata found that the participants, just like their uncom-
pensated counterparts in American universities, behaved consistent with
prospect theory. In short, incentives do not magically induce people to behave
in accord with rational choice theory.42
40. Chris Guthrie, Prospect Theory, Risk, Preference, and the Law, 97 Nw. U.L. Rev. 1115
(2003).
41. Id. at 1156–59. (The excerpt is based on the text of Guthrie’s article but does not
adhere strickly to it.) Reprinted by special permission of Northwestern University School
of Law, Northwestern University Law Review.
42. See also Colin Camerer and Robin Hogarth, The Effects of Financial Incentives in
Experiments: A Review and Capital-Labor-Production Framework, 7 Journal of Risk and
210 problem solving, decision making, and professional judgment
Up to this point, we have been using statistics to look for regularities in situa-
tions in which there are many similar events—an employer’s setting the wages
of employees, many residents’ being exposed (or not) to foam insulation in their
apartments. Statistical analysis of this sort plays an important role not just in
law and public policy but also in many other aspects of people’s personal and
professional lives.
In this chapter, we are interested in the probability that an event will occur
one time. We begin by reviewing our treatment of conditional probability begun
in Chapter 5, and then turn to Bayesian (or subjectivist) statistics, which is a
quite different approach from the frequentist approach explored thus far.
Note to Students: This chapter derives, uses, and transforms a number of formulas,
which readers who haven’t done math for a while may find daunting. It’s not essential
that you remember any of them—only that you understand the basic concepts of prob-
ability they express.
Recall that a conditional probability is the probability that event A will occur,
given that event B has occurred. It is written
P(A |B)
and is read: “the probability of A given B.” For example, in the Terra Nueva case,
P(rash present|foam) is “the probability of having a rash given that the tenant lives
in an apartment with foam insulation.” We repeat the contingency table for
Terra Nueva in Table 8.1. In view of the focus of this chapter, we’ll use its
synonym of probability table.
In Chapter 5, we found that:
P(rash present| foam) = P(foam & rash present)/P(foam)
which turns out to be 0.25/0.60 = 0.42.
Here is a generalized probability table, as shown in Table 8.2. (The symbol
“∼” (tilde) means “not.”)
Remember that the probabilities in the margins of the table, such as P(A), are
called marginal probabilities. The probability P(A & B) is called a joint probability.
212 problem solving, decision making, and professional judgment
Rash
Insulation Not Foam P(not foam & rash) P(not foam & no rash) P(not foam)
Rash
Probabilities:
Present Absent Marginal
P(A & B)
P (B | A) =
P(A)
B
B ∼B Total
A P(A & B) P(A & ∼B) P(A)
A
∼A P(∼A & B) P(∼A & ∼B) P(∼A)
Total P(B) P(∼B) 1.00
explaining and predicting one-time events 213
In People v. Collins, an interracial couple was prosecuted for robbery.1 Here are
the facts, as summarized by the Supreme Court of California:
On June 18, 1964, about 11:30 a.m. Mrs. Juanita Brooks, who had been
shopping, was walking home along an alley in the San Pedro area of the City
of Los Angeles. She was pulling behind her a wicker basket carryall contain-
ing groceries and had her purse on top of the packages. She was using a cane.
As she stooped down to pick up an empty carton, she was suddenly pushed to
the ground by a person whom she neither saw nor heard approach. She was
stunned by the fall and felt some pain. She managed to look up and saw a
young woman running from the scene. According to Mrs. Brooks the latter
appeared to weigh about 145 pounds, was wearing “something dark,” and had
hair “between a dark blond and a light blond,” but lighter than the color of
defendant Janet Collins’ hair as it appeared at trial. Immediately after the
incident, Mrs. Brooks discovered that her purse, containing between $35 and
$40, was missing. About the same time as the robbery, John Bass, who lived
on the street at the end of the alley, was in front of his house watering his
lawn. His attention was attracted by “a lot of crying and screaming” coming
from the alley. As he looked in that direction, he saw a woman run out of the
alley and enter a yellow automobile parked across the street from him. He
was unable to give the make of the car. The car started off immediately and
pulled wide around another parked vehicle so that in the narrow street it
passed within six feet of Bass. The latter then saw that it was being driven by
a male Negro, wearing a mustache and beard. At the trial Bass identified the
defendant as the driver of the yellow automobile. However, an attempt was
made to impeach his identification by his admission that, at the preliminary
hearing, he testified to an uncertain identification at the police lineup shortly
after the attack on Mrs. Brooks, when defendant was beardless.
1. People v. Collins, 68 Cal.2d 319, 66 Cal Rptr. 497, 438 P.2d 33 (1968).
214 problem solving, decision making, and professional judgment
In his testimony Bass described the woman who ran from the alley as a
Caucasian, slightly over five feet tall, of ordinary build, with her hair in a dark
blond ponytail and wearing dark clothing. He further testified that her pony-
tail was “just like” one which Janet had in a police photograph taken on June
22, 1964.
At the trial an expert testified as to the probability of the occurrence of the
physical characteristics of the defendants and the car:
The expert assumed that these six characteristics were independent, and cal-
culated that there was only one chance in 12 million that any couple possessed
all of these distinctive characteristics.
P(A& B)
P (B |A) =
P(A)
2. By the same token, multiplying the formula for P(A|B) by P(B) on both sides, would
give P(A & B) = P(B)P(A|B).
explaining and predicting one-time events 215
So, we can calculate the joint probability of the events A and B, P(A & B),
using the marginal probability of A and the conditional probability of B given A,
or the marginal probability of B and the conditional probability of A given B.
In our example,
P(mustache & beard) = P(mustache) P(beard | mustache)
If for every 10 men with mustaches, 9 have beards, then
P(beard | mustache) = 0.9
Thus: P(mustache & beard)
= P(mustache) P(beard | mustache)
= 0.25 * 0.9
= 0.225
3. The court also noted that the expert’s conclusions about the probabilities of the
particular events were not supported by evidence. The court did not address the accuracy
of the eyewitness identification—e.g., P(beard|witness says beard)—which is presented in
the taxicab problem in Section 8.5.
216 problem solving, decision making, and professional judgment
The simplest, easiest, cheapest and most powerful way to transform the quality
of intelligence would be to insist that analysts attach two little numbers to every
report they file.
The first number would state their confidence in the quality of the evidence
they’ve used for their analysis: 0.1 would be the lowest level of personal/
professional confidence; 1.0 would be—former CIA director George Tenet should
pardon the expression—a “slam dunk,” an absolute certainty.
The second number would represent the analyst’s own confidence in his or
her conclusions. Is the analyst 0.5— the “courage of a coin toss” confident—or
a bolder 0.75 confident in his or her analysis? Or is the evidence and environment
so befogged with uncertainty that the best analysts can offer the National Security
Council is a 0.3 level of confidence?
explaining and predicting one-time events 217
These two little numbers would provoke intelligence analysts and intelligence
consumers alike to think extra hard about analytical quality, creativity and account-
ability. Policy makers could swiftly determine where their analysts had both the
greatest—and the least—confidence in their data and conclusions. Decision-
makers could quickly assess where “high confidence” interpretations were based
on “low-confidence” evidence and vice versa.
Michael Schrage, What Percent Is ‘Slam Dunk’?” WASHINGTON POST, Feb. 20,
2005, editorial.
One consequence of the law of total probability is that the joint probability of two
events must always be less than or equal to the probability of one of those events.
P(A & B) can never be greater than just P(A) or just P(B).
One of the most famous results in the judgment and decisionmaking (JDM)
research suggests that people do not obey the law of total probability: they some-
times give more probability to the conjunction of two events than they give to the
two events on their own. This is called the conjunction fallacy.4
Perhaps the most common context in which people succumb to the conjunc-
tion fallacy is where the conjunction of two events seems intuitively more
plausible than one of the events alone. For example, read the paragraph below:
Linda is 31 years old, single, outspoken and very bright. She majored in phi-
losophy. As a student, she was deeply concerned with issues of discrimina-
tion and social justice, and also participated in anti-nuclear demonstrations.
Now, rate the following options in order of likelihood:
1. Linda is a teacher in an elementary school.
2. Linda is active in the feminist movement.
4. The example below, and any other examples in this section given without refer-
ences, are from studies described in detail in Amos Tversky and Daniel Kahneman,
Judgments Of and By Representativeness, quoted in Judgment Under Uncertainty:
Heuristics and Biases (Daniel Kahneman, Paul Slovic and Amos Tversky eds., New
York: Cambridge University Press, 1982) [hereafter, Heuristics and Biases] The con-
junction fallacy is discussed further in Amos Tversky and Daniel Kahneman, Extensional
Versus Intuitive Reasoning: The Conjunction Fallacy in Probabilistic Judgment, 90
Psychological Review 293–315 (1983).
218 problem solving, decision making, and professional judgment
a feminist bank teller) with the heuristic attribute (Linda’s resemblance to a femi-
nist bank teller).
explain the world. In this respect, consider Table 8.3 (which provides its own
narrative).
Poll conducted by Robin Dawes and Matthew Mulford. In KENDRICK FRAZIER, BARRY KARR, AND
JOE NICKELL, THE UFO INVASION: THE ROSWELL INCIDENT, ALIEN ABDUCTIONS, AND GOVERNMENT
COVERUPS (1997)
In a study of jury decision making, Nancy Pennington and Reid Hastie suggest
that, given a number of possible competing versions of the facts, jurors will select
the “best story,” with their confidence being determined by several factors:11
• coverage—how well the story accounts for all the evidence;
• coherence—which consists of:
• consistency—the absence of internal contradictions,
• plausibility—correspondence with the decision maker’s knowledge of
what typically happens in the world, and
• completeness—the story’s inclusion of relevant causes, psychological
states, actions, goals, consequences;
• uniqueness—the absence of good alternative competing stories.
Notice that a number of these criteria of confidence depend on detailed rather
than parsimonious data. We share Pennington and Hastie’s belief that the same
criteria are germane to explanations in many domains beyond the courtroom.
Persuasion often consists less of convincing others of the logic of one’s argu-
ment than having them replace their narrative construction of reality with your
own.12 And the richer and more detailed the story, the more persuasive. In these
respects, Scenario B in the North Korean conflict is far superior to A.
11. Nancy Pennington and Reid Hastie, Inside the Juror: The Psychology of
Juror Decision Making (New York: Cambridge University Press, 1993). See The Story
Model for Juror Decision Making and A Theory of Explanation-Based Decision Making.
12. See Howard Gardner, Changing Minds ch. 4 (Boston, MA: Harvard Business
School, 2004); Jerome Bruner, Making Stories (Boston, MA: Harvard University Press,
explaining and predicting one-time events 221
2002); Anthony G. Amsterdam and Jerome Bruner, Minding the Law (Boston, MA:
Harvard University Press, 2002).
13. Mark G. Kelman, The Heuristics Debate: Its Nature and Its Implications
for Law and Policy (Oxford University Press, forthcoming 2010).
14. Craig R. Fox and Richard Birke, Forecasting Trial Outcomes: Lawyers Assign Higher
Probabilities to Possibilities that Are Described in Greater Detail, 26 Law and Human
Behavior 159 (2002).
222 problem solving, decision making, and professional judgment
15. Gerd Gigerenzer, Adaptive Thinking: Rationality in the Real World 250
(Oxford: Oxford University Press, 2000). A complete discussion of the consequences of
using frequencies for probabilistic reasoning is given in Gerd Gigerenzer and Ulrich
Hoffrage, How to Improve Bayesian Reasoning Without Instruction: Frequency Formats, 102
Psychological Review 684–704 (1995).
explaining and predicting one-time events 223
A B
or 31 percent.
If two events are mutually exclusive—that is, it is impossible for both to hap-
pen—the circles in the Venn diagram will not intersect. In this case, P(A & B) = 0,
so the formula simplifies to
P(A or B) = P(A) + P(B)
(This is Axiom 3 in Section 5.1.2) For example, what is the probability that the
toss of a single die will result in an ace or a deuce? Since these events are mutu-
ally exclusive, we just add their probabilities: 1/6 + 1/6 = 1/3 = 0.33 or 33
percent.
Whenever you are interested in the probability that at least one of a number of
events will happen, summing all the probabilities and subtracting all the over-
laps can be tedious. There’s a simpler way to calculate the probability of at least
one of the events occurring. Consider this example:
It is late in the afternoon in California and you must send a signed document
to someone in New York, who must have it by 9 a.m. the next morning.16 You
can send the document simultaneously through three overnight services,
each of which costs about $20, with these probabilities of getting the docu-
ment to the destination on time:
• AirFast: 0.9
• BeThere!: 0.8
• Chronos: 0.8
Alternatively, you can give it to a courier who will take it with him on the red-
eye, at a cost of $2000. Before deciding whether to opt for this high-cost
16. The example comes from Amir D. Aczel, Chance: A Guide to Gambling, Love,
the Stock Market, and Just about Everything Else (New York: Thunder’s Mouth
Press, 2004).
224 problem solving, decision making, and professional judgment
option, you would like to know the probability of the document getting to
New York on time if you send three signed copies, one with each service.
We are interested in the probability that at least one of these services will
deliver on time. Let A represent the event that AirFast is on time, B the event that
BeThere! is on time, and C the event that Chronos is on time. We want to find
P(A or B or C) Instead of adding the probabilities and adjusting for the various
overlaps, we notice that the event that at least one happens is the complement of the
event that none happen. Recall that the probability of an event A’s complement is
P(∼A) = 1 – P(A) In this case,
P(A or B or C) = 1 – P(~A & ~B & ~C)
If we know that the three events are independent, we can use the multiplica-
tion rule to find this last probability.
P(A or B or C) = 1 – P(~A & ~B & ~C)
= 1 – P(~A)P(~B)P(~C)
Using the complement rule again on each of the three probabilities, we get:
In Bomber, Len Deighton observed that a World War II pilot had a 2 percent
chance of being shot down on each mission, and that he was “mathematically
certain” to be shot down in fifty missions. But this is incorrect.17
The pilot will not survive if he is shot down at least once in the fifty missions.
This phrase is a hint to use the complement rule. The event of being shot down
at least once is the complement of the event of being shot down none of the
times.
Let D represent the event of being shot down at least once.
17. David Freedman, Robert Pisani, and Roger Purves, Statistics (3rd ed. New
York: Norton, 1997).
explaining and predicting one-time events 225
The statistical techniques discussed in the preceding three chapters are based
upon the frequentist interpretation of probability—the idea that probabilities
refer to the frequencies with which events occur in the world, over many trials.
If an event has a probability of 0.5, it will occur half the time. This means that the
probability assigned to an event expresses its frequency of occurrence, which can
be estimated directly from experience.
As we mentioned at the start, however, there is another way of interpreting
the axioms underlying probability theory. Many modern statisticians also sub-
scribe to a subjectivist interpretation of probability, under which probabilities
indicate the strength of belief in a proposition. If something has a probability of
0.5, it has a 50 percent chance of being true. This means that the probability you
assign to an event expresses something about your beliefs. Both the frequentist
and subjectivist interpretations are consistent with the axioms of probability,
although they involve different perspectives about what probability means.
Frequentist statistics is concerned with P(D|H) the probability of a set of data
D under a hypothesis H, typically the null hypothesis H0. A frequentist interpre-
tation assumes that a hypothesis is either true or false, not as a random variable
but as a fixed state of the world, albeit unknown. However, with a subjectivist
226 problem solving, decision making, and professional judgment
interpretation, it is also possible to talk about P(H|D) the strength of one’s belief
in a hypothesis H after observing data D. Often lawyers and courts are more
interested in this latter quantity—for example, the probability that a defendant is
guilty based on the evidence presented—so it is worth learning to evaluate it.
Consider the following problem:
A taxicab was involved in a hit-and-run accident at night. Two cab companies,
the Green and the Blue, operate in the city. You are given the following infor-
mation:
1. 85 percent of the cabs in the city are Green; 15 percent are Blue.
2. A witness identified the cab as a Blue cab. The court tested her ability to
identify cabs under appropriate visibility conditions. When presented with
a sample of cabs (half of which were Blue and half of which were Green),
the witness made correct identifications in 80 percent of the cases and
erred in 20 percent of the cases.
What is the probability that the cab involved in the accident was Blue rather
than Green?
Here our hypothesis H is that the cab was Blue; we will call it B. Our data D
is the eyewitness testimony that the cab was Blue; we will call the event that she
says it was Blue SB. Our goal is to find P(H|D) in this case P(B|SB) the probabil-
ity that the cab was actually Blue, given that the witness said it was Blue.
First, let’s go through the given information.
• The proportion of taxis in the city that are Blue is called the base rate; it is
background information not specific to the particular event. If you had no
information about the accident, the probability that the cab in the accident
was Blue would be the base rate. For this reason it is also called the prior
probability that the cab was Blue. It is your best guess for the probability
that the cab was Blue prior to hearing the testimony. In general, the prior
probability P(H) is the strength of your belief in H prior to obtaining event-
specific data. The prior probability can be determined in different ways,
depending upon the situation. Sometimes it is estimated from a belief
based on the investigator’s experience, but typically it is taken to be the
base rate. In the hit-and-run case, it is the base rate of Blue taxis, P(B) =
0.15 Since this prior probability is substantially less than 50 percent, your
best guess before hearing the testimony would be that the cab was Green.
• The fact that the witness says the taxi was Blue (SB), is the data, D.
• You have one other important additional piece of information: the proba-
bility that the witness’s perception was accurate. This is the probability that
the witness says the cab is Blue, given that it was actually Blue, P(SB|B)
In general, this is P(D|H) and is called the likelihood of the data under
explaining and predicting one-time events 227
the hypothesis. It represents how likely the observed data would be if the
hypothesis was true. In the example, the likelihood is P(SB|B) = 0.8.
says blue
0.17
Green 0.20
0.85 says green
0.80
says blue
Blue 0.80 0.12
0.15 says green
0.20
The leftmost node represents the base rate, with 85 percent of the taxis being
Green and 15 percent being Blue. The next nodes represent the likelihood, that
is, the probability of the data if the given hypothesis is true. In this example, the
likelihood is the probability that the witness says the taxi was a particular color,
given its true color. She is accurate 80 percent of the time, so when the taxi was
actually Green (upper node), she says it was Green 80 percent of the time and
Blue 20 percent of the time. When the taxi was actually Blue (lower node), she
says it was Blue 80 percent of the time and Green 20 percent of the time.
Now let’s solve, or “fold back,” the tree by using the methods of Section 8.2—
that is, by multiplying the probabilities along each branch. For example, the
probability that the cab was Blue and the witness says it was Blue is
P(B & SB) = P(B)P(SB|B) = 0.15 * 0.80 = 0.12
or 12 percent. Similarly, if we let G represent the hypothesis that the cab was
Green, the probability that it was Green and she says it was Blue is
P(G & SB) = P(G)P(SB|G) = 0.85 * 0.20 = 0.17
18. The concept of a decision tree is described in greater detail in Section 15.2.
228 problem solving, decision making, and professional judgment
or 17 percent. (Since she said it was Blue, we can ignore the second and fourth
branches.) The witness says the cab was Blue 17 + 12 = 29 out of 100 times and
is correct 12 of those times. Dividing 12 by 29, we obtain 0.41, or 41 percent.
Given her testimony that the cab was Blue, there is still only a 41 percent chance
that the cab actually was Blue, so our best guess is that the cab was Green. If this
answer is surprising, keep in mind the base rate: only 15 percent of cabs are
Blue. As we discuss below, people tend to ignore the base rate when estimating
probabilities.
P(D |H ) P(H)
P(H |D ) =
P(D)
Note to Students: If you understand how to represent and analyze the taxi problem
using a decision tree, you’ve got the gist of Bayes’ Theorem. But the chapter would be
incomplete without walking through how the equation actually works.
Let’s replace the general formulation of Bayes’ Theorem with the particulars
of the taxi problem. We were given the likelihood, P(SB|B): it is 0.8. We were also
given P(B), the base proportion of Blue cabs: it is 0.15. Substituting in Bayes’
equation, we get:
Now we need the value of the denominator, P(SB), which is the overall, or mar-
ginal, probability that the witness will say the cab was Blue. Now letting SG
represent the event that the witness says the cab was Green, the probabilities of
this problem are displayed in Table 8.4.
Probabilities Testimony
SB SG Total
B P(B & SB) P(B & SG) P(B)
Truth
G P(G & SB) P(G & SG) P(G)
Total P(SB) P(SG) 1.00
explaining and predicting one-time events 229
There are only two ways the event SB—the witness saying it was a blue cab—
can come about: either the cab is blue and she says it’s blue, or the cab is green
and she says it’s blue. These are mutually exclusive and collectively exhaustive.
Therefore,19 P(SB)—the marginal probability of the testimony that the cab was
blue—is the sum of (a) the probability that the witness says the car was Blue and
it was in fact Blue, and (b) the probability that she says the car was Blue and it
was in fact Green:
P(SB) = P(B & SB) + P(G & SB)
As with the example of finishing the brief on time (in Section 8.2.1 and
Section 8.2.3), we calculate the two terms:
P(B & SB) = P(SB|B) * P(B)
P(G & SB) = P(SB|G) * P(G)
Since the witness is correct 80 percent of the time, the probability that she will
say the taxi was Blue if it was Blue is 0.8, and the probability that she will say it
was Blue if it was Green is 0.2. Plugging in these numbers, we get:
P(B & SB) = P(SB|B) * P(B) = 0.8 * 0.15 = 0.12
P(G & SB) = P(SB|G) * P(G) = 0.2 * 0.85 = 0.17
P(SB) = P(B & SB) + P(G & SB)
= 0.12 + 0.17
= 0.29
Above, we had the partially solved equation
0.8 * .015
P (B | SB) =
P(SB)
Now that we know that P(SB) is 0.29, we can plug that into the denominator,
getting:
0 .8 * 0 .15
P (B | SB) = = 0 .41
0 .29
8.5.3 Diagnosticity
Recall that we can discuss uncertainty in terms of odds instead of probabilities.
Here, just as we have prior and posterior probabilities of an event, we have prior
and posterior odds, the odds of the event before and after we observe the data,
respectively.
In the previous section, with the Bayes tree method, we used the likelihood to
transform the prior probability—P(B) = 0.15—into the posterior probability—
P(B|SB) = 0.41. To transform the prior odds into the posterior odds, we use the
likelihood ratio.
Recall from the previous section that the likelihood is the probability of
observing data given a certain hypothesis, P(D|H) In the example, the data (D)
consisted of the testimony SB that the cab was Blue. The likelihood of this testi-
mony under the hypothesis (H) that the cab was Blue is P(SB|B) = 0.8 and the
likelihood under the hypothesis that the cab was Green is P(SB|G) = 0.2. The
likelihood ratio, therefore is
P (SB | B ) 0 .8
= =4
P (SB | G ) 0 .2
This means that the testimony that the cab was Blue is four times as likely if
it actually was Blue than if it was Green. In general, the likelihood ratio is
P (D | H)
P (D | ~H)
The numerator is the likelihood of the data given that H is true, and the
denominator is the likelihood of the data given that H is false.
To transform prior odds into posterior odds, we multiply by the likelihood
ratio:
Posterior odds = Likelihood ratio * Prior odds
If the data are more likely when H is true, the likelihood ratio is greater than
one, and observing the data increases the odds. If the data are more likely when
H is false, the likelihood ratio is less than one, so observing the data decreases
the odds.
The likelihood ratio is also called the diagnosticity because of its meaning in
medical diagnostic tests.
Many diagnostic tests have significant false positive rates. If we let D repre-
sent a positive test result and H the hypothesis that a patient has the disease in
question, then the true positive rate (also called the sensitivity) of the test is
P(D|H), the probability of a positive result when the patient has the disease. The
false positive rate is P(D|∼H), the probability of a positive result given that
the patient does not have the disease. The diagnosticity is therefore the ratio of
the true positive rate to the false positive rate.
Suppose you wake up one morning with a sore throat. You have heard that,
of all sore throats, 10 percent are bacterial infections and the other 90 percent
explaining and predicting one-time events 231
are viral. In other words, for every one bacterial infection, there are nine viral
infections, so you figure that your odds of having a bacterial infection—let’s say
strep throat—are 1:9.
You go to the doctor, and she gives you a rapid strep test. She explains that,
among people with strep throat, 81 percent will get a positive test result, and
among those with a virus, 1 percent will get a positive test result. In other words,
the sensitivity or true positive rate is 0.81, and the false positive rate is 0.01. You
calculate that the diagnosticity of the rapid strep test is 81.
The doctor calls you with the news that you tested positive. Now what are your
odds of having strep?
Posterior odds = Diagnosticity * Prior odds = 81 * 1/9 = 9/1
Your odds have increased dramatically from 1:9 to 9:1, because of the high
diagnosticity of the test, and she writes a prescription for antibiotics.
A high diagnosticity is an indication of a good test. Whenever the diagnostic-
ity is greater than 1, we know two things:
• Because diagnosticity is the likelihood ratio P(D|H)/P(D|∼H) we know the
true positive rate P(D|H) is greater than the false positive rate P(D|~H).
• Because posterior odds = Diagnosticity * Prior odds, we know that the pos-
terior odds (after a positive test result) are greater than the prior odds. That
is, a positive test result increases the odds that the patient has the disease.
figure 8.3 way too general practitioner © The New Yorker Collection ,
feb. 14–21 2005, 120. all rights reserved. reprinted with permission.
232 problem solving, decision making, and professional judgment
On the other hand, if the diagnosticity of a test is 1, it does not change the
odds and is not worth performing.
The rarer the disease in question, the higher a test’s diagnosticity must be to
give one confidence in the result. For example, in order to have better-than-even
odds of having SARS, you would have to test positive on a test with very high
diagnosticity, in order to make up for SARS’s very low base odds. Even if your
symptoms are more consistent with SARS, you might still be more likely to have
the flu, just because the flu is a much more common ailment.
For an analysis of the taxi problem using the odds version of Bayes’ Theorem,
see the Appendix, Section 8.7.20
CT scans have become a significant industry even for customers who have no
particular risk of disease. Are they useful? Steven Woloshin and colleagues cite a
report that people whose lung cancer is found early by such scans have a five-year
survival rate of 80 percent, as opposed to 15 percent for the typical lung-cancer
patient whose condition is detected later.20 They are nonetheless skeptical about
whether the scans do more good than harm.
One problem is that CT scans detect tiny tumors, but do not have a very high
diagnosticity, and thus must often be followed up by biopsies, which can be pain-
ful and even dangerous. So for starters, the test can give rise to harm to people
who don’t have, and are not in danger of developing, lung cancer. And even when
cancerous abnormalities are discovered, they may be indolent and never develop
into anything dangerous. But because there’s no way to know, most people who
are diagnosed are treated with surgery, radiation therapy, or chemotherapy—also
painful and often dangerous.
But what of the survival statistics? Here’s how they are calculated: Suppose
that 1000 people were diagnosed with lung cancer five years ago. If 150 are alive
today, the five-year survival is 150/1000, or 15 percent; if 800 are alive today the
five-year survival is 80 percent.
But Woloshin and his colleagues argue that even if CT screening raised the
five-year survival rate from 15 percent to 80 percent, no one might get an extra day
of life. They suggest this thought experiment. Imagine a group of people with
lung cancer who will all die at age 70. If they are diagnosed when they are 67, their
five-year survival rate would be zero percent. But if they were diagnosed when
they were 64, their five-year survival rate would be 100 percent. Early diagnosis
increases the five-year survival statistic but doesn’t postpone death.
20. Steven Woloshin, Lisa Schwartz, and Gilbert Welch, Warned, but Worse Off,
New York Times, Aug. 22, 2005.
explaining and predicting one-time events 233
The fact that CT scans identify small indolent tumors also inflates the survival
rate. They suggest another thought experiment. Imagine a population where,
after having symptoms such as a persistent cough, 1000 people are diagnosed
with progressive lung cancer. Five years later,150 are alive. Suppose the entire
population gets CT scans, and that 5000 are diagnosed with cancer, but the addi-
tional 4000 actually have indolent forms and are alive five years later. This would
raise the five-year survival rate to 83 percent—because these healthy people
would appear in both parts of the fraction: survivors/people diagnosed with
cancer = (150 + 4000)/(1000 + 4000) = 4150/5000 = 83%. But what has really
changed? Some people have been unnecessarily told they have cancer (and may
have experienced the harms of therapy), and the same number of people (850)
still died.
The phenomenon was borne out by a randomized trial of chest X-ray screen-
ing at the Mayo Clinic, where the five-year survival was higher for those who were
screened, but death rates were slightly higher for those who received the
screening.21
21. The New England Journal of Medicine published the study on which the Woloshin
article is based without disclosing that its principal investigator and her institution
held patents related to CT scanning or that the study was funded by a cigarette manufac-
turer. In January 2009 the journal announced a new disclosure policy. New York
Times, Jan. 9, 2009, available at http://www.nytimes.com/2009/01/09/us/09journal.
html?_r=1&ref=health.
22. Amos Tversky and Daniel Kahneman, Evidential Impact of Base Rates, (1982), in
Heuristics and Biases, supra at 153.
234 problem solving, decision making, and professional judgment
estimate that appears to take some account of the base rate. Kahneman and
Tversky’s explanation, which is supported by other experiments, is that the
base rate in the variant problem provides some causal information: it suggests
that the drivers of the Green cabs are less competent or careful drivers than the
drivers of Blue cabs, and this induces people to attend to the base rates of care-
ful and careless drivers.
Kahneman and Tversky23 gave two groups of undergraduate students some
short descriptions of individuals and asked them to judge whether the individu-
als were likely to be lawyers or engineers. Here is a typical description:
Jack is a 45-year-old man. He is married and has four children. He is gener-
ally conservative, careful, and ambitious. He shows no interest in political
and social issues and spends most of his free time on his many hobbies,
which include home carpentry, sailing, and mathematical puzzles.
One group of students was told that the descriptions came from a sample of
people of whom 70 percent were lawyers and 30 percent engineers. The other
group of students was told that the descriptions came from a sample of people of
whom 30 percent were lawyers and 70 percent engineers. Thus the two groups
were given quite different base rates.
Most of the students judged Jack to be very likely to be an engineer. Most
interestingly, their judgments were unaffected by the different base rates. In
Bayesian terms, the students entirely ignored P(H). Here, again, the represen-
tativeness heuristic is at work: People seem to be making decisions based upon
how representative the description is of a particular profession, rather than by
combining prior and likelihood information in a fashion consistent with
Bayesian inference.
These and other findings suggest that humans do not combine probabilistic
information in a Bayesian fashion. When told:
Suppose . . . that you are given no information whatsoever about an individual
chosen at random from the sample.
respondents correctly referred to the base rates in predicting the person’s
profession. But when given a diagnostically worthless description, such as:
Dick is a 30-year-old man. He is married with no children. A man of high
ability and high motivation, he promises to be quite successful in his field. He
is well liked by his colleagues.
23. Amos Tversky and Daniel Kahneman, Causal Schemas in Judgments Under
Uncertainty, (1973), in Heuristics and Biases, supra at 117.
explaining and predicting one-time events 235
Legal Problem
We have focused on people’s tendency to neglect base rates. Can you think
of an instruction that judges give jurors as a matter of course in criminal
cases that is intended to counteract their tendency to take account of base
rates?
they neglected the base rates, judging Dick to be equally likely to be a lawyer or an
engineer regardless of the relative frequencies of these two professions in the
samples.
Let us define match as the presence of the accused’s blood type, AB+, at the
scene of the crime. And let us treat the presence or absence of the accused’s
actual blood at the scene of the crime as equivalent to his being guilty or not
guilty. We know that the probability of a match for a randomly selected member
of the population, P(match) = 0.03. Applying the prosecutor’s argument in
Collins to the present case, P(match|not guilty) = 0.03, and therefore the prosecu-
tor would want the jury instructed that there is a 97 percent chance that the
suspect committed the murder.
But we’re interested in P(not guilty|match). Calculating this number requires
knowing the population of people who might have committed the crime. Let’s
assume that the relevant population contained 1000 people, and that there is no
other evidence connecting the suspect to the crime—so that if one other person
in the relevant population has AB+, the suspect cannot be convicted “beyond a
reasonable doubt.” We can use Bayes’ Theorem,
P(D | H ) P(H)
P(H | D ) =
P(D)
Forensic Evidence
are not the same25. By contrast to fingerprint matches, which courts typically
treated as certain, the relative novelty of DNA evidence opened it to a probabilis-
tic understanding, which has begun to affect more traditional forms of forensic
evidence.26
25. Jonathan Koehler, quoted in Gigerenzer at 167, says 1 in 100 or 200. Hans Zeisel
and David H. Kaye, Prove it With Figures: Empirical Methods in Law and
Litigation (New York: Springer-Verlag, 1997).
26. See Michael J. Saks and Jonathan J. Koehler, The Coming Paradigm Shift in Forensic
Identification Science, 309 Science 892 (2005).
27. Gerd Gigerenzer, Calculated Risks: How to Know When Numbers Deceive
You (New York: Simon & Schuster, 2002). The particular questions quoted in the text
were put forth by one expert, leading to the development of a broader survey with results
along the same lines.
238 problem solving, decision making, and professional judgment
correctly to reach the correct result of 7.5 percent. Another two were close, but
their reasoning was not correct.28
The second group of physicians was given information in terms of natural
frequencies.
Ten out of every 1,000 women have breast cancer. Of these 10 women with
breast cancer, 8 will have a positive mammogram. Of the remaining 990
women who don’t have breast cancer, some 99 will still have a positive mam-
mogram. Imagine a sample of 100 women who have positive mammogram
screening. How many of these women actually have breast cancer?
Eleven of this group of twenty-four physicians gave the correct answer, and
quite a few of the others were close. Gigerenzer argues that the computation
based on natural frequencies is much simpler than using Bayes’ Theorem—
something that most readers who worked through the problems above would
agree with.
We think that a tree, of the sort used in the taxi problem, provides an easy way
of representing this problem, as shown in Figure 8.4.
positive test
has cancer 0.8
0.01 negative test
0.2
Positive test
does not have cancer 0.1
0.99 Negative test
0.9
Since we are only interested in cases where someone tested positive, we focus
on the two branches with positive tests:
has cancer * positive test = 0.01 * 0.8 = 0.008
28. Id. 43. Gigerenzer states that the correct result was 8 percent, but the actual
number, as derived below, is 7.476.
explaining and predicting one-time events 239
8.7 appendix
Let us consider those components in turn. Every moment of our lives, we are
bombarded with vast amounts of information, of which we attend to only a small
fraction. We encode the information, structuring, evaluating, and interpreting it
and transforming into some sort of mental representation. You might think of
perception (not on the chart) as overlapping attention and encoding. We store
information in memory and, on occasion, retrieve it from memory (i.e., become
aware of it) and process it with respect to particular objectives. Our response to
processing may be a factual judgment or inference, an evaluative judgment or
opinion, a choice among alternatives or decision, or a solution to a problem.1
This and the following chapter examine biases and distortions that can affect
these stages of information processing. Although some of the psychological
1. See Verlin B. Hinsz, R. Scott Tindale, and David A Vollrath, The Emerging
Conceptualization of Groups as Information Processors, 43 Psychological Bulletin 121
(1997).
242 problem solving, decision making, and professional judgment
A large amount of any lawyer’s time is focused on past events. It is the under-
standing and analysis of past events that give rise to legal liability—whether civil,
administrative, or criminal—and that provide the basis for the predictions
necessary for policy making and planning. By definition, understanding the past
calls upon memory. In many contexts, however, memories tend to be highly
fallible.
Our memory does not store complete representations of what was perceived,
but only fragments of our interpretations of the relevant facts or events.
Recollection requires reconstructing one’s interpretation, often using (more
or less) logical inferences to fill in missing details.2 In Eyewitness Testimony,
Elizabeth Loftus divides memory processes into three stages: acquisition, in
which the information is entered; retention, the period between acquisition and
the demand to “remember”; and retrieval, where one recollects the information.
Mr. Hill was so engrossed in the call that he ran a red light and didn’t notice Linda
Doyle’s small sport utility vehicle until the last second. He hit her going 45 miles
per hour. She was pronounced dead shortly after. Later, a policeman asked
Mr. Hill what color the light had been. “I never saw it,” he answered.6
Section 1.5.2 introduced the idea that all of our perceptions are mediated by sche-
mas or expectations: when we perceive a stimulus from our environment, our
first task is to fit that information into some existing knowledge structure repre-
sented in memory. This is essential and inevitable—and a hazard for the careful
empiricist. In a classic experiment, Jerome Bruner and Leo Postman briefly
5. Dolly Chugh and Max H. Bazerman, Bounded Awareness: What You Fail to See Can
Hurt You, 6 Mind & Society 1–18 (2007).
6. Matt Richtel, Drivers and Legislators Dismiss Cellphone Risks, New York Times,
July 19, 2009, http://www.nytimes.com/2009/07/19/technology/19distracted.html?hp.
biases in perception and memory 245
showed participants five playing cards, one of which was a three of hearts colored
black.7 A vast majority of the participants engaged in “perceptual denial,” confi-
dently recalling that the card was a normal three of hearts. The phenomenon of
illusory correlation described in Section 5.7 exemplifies the same tendency.
Elizabeth Loftus reports on an accident where hunters killed a companion: “One
of the men . . . saw something moving and said to his friend, ‘That’s a deer, isn’t it?’
The friend replied that he thought so too, and the first man shot at the deer. The
deer pitched forward and cried out—a sound which seemed like the cry of a
wounded deer. The hunters fired more shots to bring it down.” Loftus explains:8
The hunters who eagerly scanned the landscape for a deer perceived the
moving object as a deer. They expected to hear the cry of a deer and they heard
their friend’s cry that way . . . Yet a policeman testified that when he later saw
a man under the same conditions he perceived the object as a man . . .[T]he
policeman knew he was supposed to be looking at a man; thus, he perceived
the object he saw as a man.
Consider these more complex examples:
fair” and over a third introduced their own category of “rough and fair” to
describe the action. Although a third of the Dartmouth students felt that
Dartmouth was to blame for starting the rough play, the majority of Dartmouth
students thought both sides were to blame . . .
When Dartmouth students looked at the movie of the game, they saw both
teams make about the same number of infractions. And they saw their own
team make only half the number of infractions the Princeton students saw
them make. The ratio of “flagrant” to “mild” infractions was about one to one
when Dartmouth students judged the Dartmouth team, and about one “fla-
grant” to two “mild” when Dartmouth students judged infractions made by
the Princeton team.
Granted that the students’ perceptions of fault were biased in favor of their
own schools’ teams, what explains the difference in perceptions? Hastorf and
Cantril answer:
It seems clear that the “game” actually was many different games and that
each version of the events that transpired was just as “real” to a particular
person as other versions were to other people . . .
. . . An “occurrence” on the football field or in any other social situation
does not become an experiential “event” unless and until some significance is
given to it. And a happening generally has significance only if it reactivates
learned significances already registered in what we have called a person’s
assumptive form-world.
At the very least, people interpret the same ambiguous information in differ-
ent ways: I see the tackle as gratuitously violent while you see it as hard-hitting
but fair. At the worst, from all the events going on in the environment, we some-
times notice those that are significant from our egocentric positions, and fail to
notice those that don’t.
10. Robert P. Vallone, Lee Ross, and Mark R. Lepper, The Hostile Media Phenomenon:
Biased Perception and Perceptions of Media Bias in Coverage of the Beirut Massacre, 49
Journal of Personal and Social Psychology 577 (1985).
biases in perception and memory 247
The pro-Arab participants saw the media as biased toward Israel; the pro-
Israeli participants saw the news programs as biased against Israel. The pro-
Arab participants thought that the programs applied lower standards to Israel
than to other countries, that they did not focus enough on Israel’s role in the
massacre, and that the editors made a positive case for Israel and were biased in
its favor. Pro-Israeli participants thought that the programs applied a higher
standard to Israel, focused too much on Israel’s role, made a negative case
against Israel, and were biased against Israel. Furthermore, participants who
deemed themselves more knowledgeable or more emotionally involved had
stronger perceptions of bias.11 Neutrals were about midway between the parti-
sans. (Viewers who thought themselves knowledgeable thought the main source
of bias was what was omitted, particularly that the media did not provide the full
context.)
Just as Hastorf and Cantril believed that Dartmouth and Princeton students
had seen different football games, Vallone, Ross, and Lepper concluded that
“the pro-Arab and pro-Israeli participants ‘saw’ different news programs—that
is, they disagreed about the very nature of the stimulus they had viewed.”
For example, the groups reported that a higher percentage of the references to
Israel were, respectively, favorable or unfavorable. And each believed that the
programs would lead an undecided or ambivalent viewer to favor the opposing
position. The authors offer this analysis of the phenomenon:
Our results highlight two mechanisms—one apparently evaluative or cogni-
tive, the other apparently more perceptual in nature—that combine to pro-
duce the partisans’ conviction that they have been treated unfairly. According
to the first mechanism, in which opposing partisans believe, respectively, that
the truth is largely “black” or largely “white,” each complain about the fairness
and objectivity of mediated [i.e., media] accounts that suggest that the truth
might be at some particular hue of gray. According to the second mechanism,
opposing partisans further disagree about the color of the account itself: One
side reports it to be largely white (instead of the blackish hue that the other
side thinks it should be), the other side reports it to be largely black (instead
of the whitish hue that the first side thinks it should be), and both sides believe
the discrepancy between the mediated account and the unmediated truth to
be the intended result of hostile bias on the part of those responsible.
We note that our results do not permit us to speak authoritatively about
either the source or the depth of the perceptual bias we have claimed to
document; nor, obviously, do they shield us from the age-old difficulties of
11. While knowledgeable neutral participants did not perceive a bias one way or the
other, unknowledgeable neutrals more closely resembled the pro-Israeli than pro-Arab
groups. The authors note that American public opinion tends to be pro-Israeli and that the
neutrals may be more pro-Israeli than they realize.
248 problem solving, decision making, and professional judgment
12. Linda Babcock, George Loewenstein, Samuel Issacharoff, and Colin Camerer,
Biased Judgments of Fairness in Bargaining, 85 American Economic Review 1337 (1995);
Linda Babcock and George Loewenstein, Explaining Bargaining Impasse: The Role of Self-
Serving Biases, 11 Journal of Economic Perspectives 109 (1997). Abstract available at
SSRN: http://ssrn.com/abstract=11367.
biases in perception and memory 249
answering the questions, and negotiating, while those in the second group first
read the testimony and answered the questions and were then assigned roles
before they negotiated.
The assessment of what would be a fair judgment by plaintiffs and defen-
dants who had been assigned roles before reading the testimony diverged by
$19,756 on average; their predictions of the judicial outcome diverged by $18,555.
Moreover, 28 percent of the parties in this group failed to settle. By contrast,
plaintiffs and defendants in the second group did not diverge significantly in
their assessments of either of fairness or the judicial outcome, and only 6 per-
cent failed to settle. Apparently, knowing one’s role in advance causes one to
encode the facts in a way that is self-serving and that leads to undue optimism
about the assessments likely to be made by third parties.
Not only students, but experienced litigators and negotiators are prone to this
self-serving bias. Whereas purely economic analyses attribute the failure to settle
to strategic behavior, the authors suggest: “Perhaps disputants are not trying to
maximize their expected outcome, but only trying to achieve a fair outcome.
However, what each side views as fair tends to be biased by self-interest, which
reduces the prospects for settlement.”
Merely informing disputants about the prevalence of self-serving bias had
no effect on their bias. Is it possible to counter the bias? Babcock et al. did a fol-
low-up experiment in which the participants, after being assigned roles and
reading the testimony, were instructed:
Disputants don’t always think carefully about the weaknesses in their own
case: they are therefore surprised when the judge’s ruling is worse than their
expectations. For plaintiffs, this means that the judge’s award is often less
than their expectations. For defendants, this means that the judge’s award is
often greater than their expectations. Therefore, please think carefully about
the weaknesses in your case. In the space below, please list the weaknesses in
your own case.
Being asked to assess the weaknesses of one’s own case had a significant
effect. When given the debiasing instruction, plaintiffs’ and defendants’ predic-
tions of the judge’s award differed by only $5000, and the litigants reached settle-
ment much more quickly. The authors suggest that a debiasing procedure could
be built into court-mandated pretrial settlement conferences.
It isn’t so astonishing, the number of things that I can remember, as the number of
things I can remember that aren’t so.
—Mark Twain, A Biography13
13. Http://www.twainquotes.com/Memory.html.
250 problem solving, decision making, and professional judgment
For any reader who doesn’t recall the phenomenon of forgetting, consider
this a reminder. Of particular interest in this section are events that distort the
content of what is remembered. Distorted recollections are often exacerbated by
people’s overconfidence in the correctness of their memories.
It is useful to treat retention and retrieval together, because intermediate
retrievals are among the major sources of distortion of ultimate retrievals, includ-
ing those that have legal consequences. A witness’s ultimate retrieval may take
place when she testifies at trial, but her recollections may be distorted by ques-
tions posed earlier by investigators or in a deposition, or by chatting with friends
about the event.
One group of biases results from the phenomenon of memory binding, in
which different components of an event, or different events, are bound together.14
For example, a Ryder employee who observed Timothy McVeigh rent a van two
days before the 1995 bombing of the Federal Building in Oklahoma City recalled
there being two men. One was tall and fair, and fit McVeigh’s description; the
other was short, stocky, dark-haired, and had a tattoo. It turned out that a day
after McVeigh rented the van (alone), two men, who matched the descriptions of
McVeigh and his supposed companion, came in to rent a van. As Daniel Schacter
explains: “People recall correctly a fact they learned earlier, or recognize accu-
rately a person or object they have seen before, but misattribute the source of
their knowledge.”15
More broadly, experiences after an event can distort one’s recollection of the
event. For example, people who read an inaccurate newspaper description of an
event they personally witnessed are prone to incorporate the inaccuracies into
their recollections.16 Leading or misleading questions can also distort recollec-
tion. Elizabeth Loftus and colleagues showed participants photos of a car-
pedestrian accident involving a red Datsun. One group saw the car at a “stop”
sign; the other saw the car at a “yield” sign. Within each group, half the partici-
pants were asked “Did another car pass the red Datsun while it was stopped at
the ‘stop’ sign?” and half were asked “Did another car pass the red Datsun while
it was stopped at the ‘yield’ sign?” Finally, after performing some unrelated
intervening tasks designed to put temporal distance between the intermediate
and ultimate retrieval, the participants were shown slides depicting the red
Datsun either at a “stop” or “yield” sign and asked which comported with the
original photo. Of participants who had originally seen a “stop” sign and had
been asked whether the car was stopped at a “stop” sign, 75 percent responded
correctly. Of participants who had been asked the misleading intermediate
question—“Did another car pass the red Datsun while it was stopped at the
14. Daniel L. Schacter, The Seven Sins of Memory: How the Mind Forgets and
Remembers 94 (New York: Houghton Mifflin, 2001) [hereafter Schacter].
15. Schacter 93.
16. Loftus 55.
biases in perception and memory 251
The heuristics and biases agenda, begun in the 1970s by the psychologists Amos
Tversky and Daniel Kahneman, continues to inform much psychological research
today, especially in social psychology and judgment and decision making (JDM).
The heuristics identified by Tversky and Kahneman are based on processes that
operate unconsciously and automatically in System 1, and that tend to be insuf-
ficiently corrected by the conscious analytic processes of System 2.23 We discussed
the representativeness heuristic in detail in Section 8.3. Here we consider the
availability heuristic. While representativeness involves categorizing, availability
involves noticing.
Imagine that you have been considering buying a new car for some time. After
poring through Consumer Reports and some auto magazines, you decide that the
Volvo sedan is just what you want—comfortable, safe, a good maintenance
record; all in all, a great buy. The evening before you plan to purchase the car, you
are at a dinner party where a guest describes his terrible experience with the
Volvo he bought last year—a real lemon, which has spent more time in the shop
than on the road.24 How much will this one anecdote influence your decision
compared to all the research you have done?
What Kahneman and Tversky call the availability heuristic leads people to
make judgments of probability and causation based on how readily events come
to mind because they are vivid, imaginable, or otherwise perceptually salient.
Thus, in attributing the cause of the problems at Terra Nueva, residents will
more likely be influenced by vivid stories of tenants’ illnesses and media cover-
age of “foam insulation syndrome” elsewhere in the country than in epidemio-
logical evidence.
In one experiment, two groups of participants were given samples of the
heights of a number of individuals and asked to estimate the proportion of people
over 6’ tall. Twenty percent of the individuals in both samples were over 6’ tall,
but one group had several very tall outliers. The participants estimated that this
sample had more individuals over 6’ tall. By the same token, participants esti-
mated that a sample of people that included some criminals contained more
criminals when some of them had committed heinous offenses than a similar
sample with the same number of criminals who had only committed minor
offenses.25 What impressions do you imagine various media audiences have of
the number of people with extreme versus moderate views on issues such as
abortion, gun control, and gay marriage?
Consider the following problem:
In four pages of a novel (about 2000 words), how many words would you
expect to find that have the form
_ _ _ _ _ n _ ? (five letters, then n, then another letter)
_ _ _ _ i n g ? (four letters, then ing)
What were your estimates? In between-subjects experiments, people usually
say that they expect more words of the second kind than of the first. But, of
course, this cannot be true, since ing is a subset of _n_. By the same token,
people estimate that there are more “dairy products” than “white foods” in a
supermarket.
These are logical errors—akin to the error made in the example of Linda the
(feminist) bank teller in Section 8.3. Most people understand that there cannot
be more instances of a subset than of the set of which it is a part. The errors arise
from people’s tendency to estimate the frequency or probability of an event by
how readily examples come to mind. (Consider how people would respond to the
24. Richard E. Nisbett and Lee Ross, Human Inference: Strategies and
Shortcomings in Social Judgement. (Englewood Cliffs, NJ: Prentice Hall, 1980).
25. Id. 74–75.
254 problem solving, decision making, and professional judgment
question: “How many white foods are there in a supermarket?—and by the way
don’t forget dairy products.”) It is easy to see why using availability as a proxy for
frequency can be a useful strategy—if events happen more often, we tend to
recall instances of such events more easily. But events can come to our attention
and/or be retrieved more readily for many reasons other than the frequency of
their occurrence.
26. Norbert Schwarz and Leigh Ann Vaughn, The Availability Heuristic Revisited: Ease
of Recall and Content of Recall as Distinct Sources of Information, in Thomas Gilovich,
Dale Griffin, and Daniel Kahneman, Heuristics and Biases, The Psychology of
Intuitive Judgment 103 (New York: Cambridge University Press, 2002).
27. Steven J. Sherman, Robert B. Cialdini, Donna F. Schwartzman, and Kim D.
Reynolds, Imagining Can Heighten or Lower the Perceived Likelihood of Contracting a Disease:
The Mediating Effect of Ease of Imagery, in Thomas Gilovich, Dale Griffin, and Daniel
Kahneman, Heuristics and Biases, The Psychology of Intuitive Judgment 98
(2002).
28. The authors of the study suggest that this has implications for preventive health,
such as persuading people to have immunization shots or to take measures against hyper-
tension: “The health profession might increase . . . compliance by making vivid presenta-
tions of the medical problems and by describing symptoms and consequences . . . in
easy-to-imagine terms.” Note, however, that unless this actually leads to changes in health
behavior, it may just increase people’s anxiety.
biases in perception and memory 255
on how people evaluate the content of the information they are exposed to, theories
about metacognition hinge on the fact that people also make judgments by evalu-
ating features of the experience of thinking. In particular, experiencing the process-
ing of information as easy or difficult can significantly influence our judgments
of the content of that information. Of course people’s prior assumptions or expec-
tations about the ease or difficulty of a task are also relevant. For example, when
the people who listed twelve examples of assertive behavior were made aware of
the objective fact that listing twelve examples was not an easy task, they were less
likely to attribute their experienced difficulty to lack of assertiveness.29
The mere exposure effect can have serious public policy consequences. Since
we are constantly bombarded with facts, figures, and information, it is crucial for
consumers and citizens to discern between true and false assertions. While the
mere exposure effect robustly suggests that information that is frequently
repeated tends to stick, experiments conducted by Norbert Schwarz and his col-
leagues indicate that the ability to distinguish between the truth and falsehood of
that repeated information fades with time.33 Their study also indicates that
memory of context—which we use to retroactively determine whether a piece of
information is true or false—declines when one is engaged in cognitively
demanding tasks, while familiarity is hardly affected by such tasks.
Not only repeated false claims about an individual, product, or policy, but
efforts to debunk them, can contribute to their apparent truthfulness by height-
ening their availability. For a vivid example of this, consider Richard Nixon’s
statement, “I am not a crook,” in rebutting claims that he was involved in the
Watergate break-in.
9.6.2.b Using Metacognitive Monitoring to Repair Judgment Biases People
sometimes use metacognition to try to correct for what they perceive as biases.
For example, how common is the name Kennedy? When people are asked to
make such estimates, they might rely on the availability heuristic based on how
easily the name comes to mind. However, people’s awareness that the name of
a prominent political family appears so often in the media may lead to self-
correction of the heuristic. In a recent experiment, when asked to estimate the
frequency of names which happened to be the names of prominent celebrities
(e.g., politicians and musicians), participants estimated that the names were
much less frequent than they actually were. In other words, they overcorrected
for the bias caused by the availability heuristic.34
Metacognitive correction of perceived biases can have especially important
implications in the domain of justice. Research done by Duane Wegener and
Richard Petty35 suggests that people can use metacognition to identify potentially
biasing factors and attempt to remove bias from their perceptions and judg-
ments. Wegener and Petty posited that when trying to prevent bias or debias
initial perceptions (e.g., when on jury duty), people consult their own metacogni-
tive beliefs about how various biasing factors might influence their perceptions.
33. E.g., Ian Skurnik, Carolyn Yoon, Denise C. Park, and Norbert Schwarz. How
Warnings About False Claims Become Recommendations. 31, Journal of Consumer
Research, 713–724 (2005).
34. Daniel M. Oppenheimer, Spontaneous Discounting of Availability in Frequency
Judgment Tasks, 15 Psychological Science 100–05 (2004).
35. Reported in Duane T. Wegener, and Richard E. Petty, The Flexible Correction Model:
The Role of Naïve Theories of Bias in Bias Correction, in Advances in Experimental Social
Psychology 29, (M. Zanna ed., 1997).
biases in perception and memory 257
In one experiment, Petty and his colleagues asked college students to judge
whether a person accused of the crime of rape was guilty or not. Participants in
the control group were presented with ambiguous information that led them to
be uncertain as to whether or not the defendant had committed the crime.36
A second group read the same ambiguous information but also learned that the
defendant had pled guilty to a prior crime of rape. A third group read the ambig-
uous information but also learned that the defendant had pled guilty to a prior
crime of burglary.
Participants were then asked to give two judgments: their own personal opin-
ion of the defendant’s guilt, and the verdict that they would render if they were a
juror. When giving their own opinions, participants were more likely to consider
the defendants with prior records (rape and burglary) guilty than the defendant
with no prior record. Perhaps they regarded some people as “criminal types.”
When acting as jurors and rendering verdicts, however, there was no difference
between the defendant with the burglary and the defendant with no record.
However, more participants judged the defendant with the prior rape conviction
to be guilty. The researchers concluded that when trying to be good jurors, people
may correct for biases due to factors they perceive as irrelevant to making an
impartial judgment (prior burglary conviction for a rape suspect), but not for
biases due to information they perceive as relevant (prior rape conviction for a
rape suspect) even when instructed to disregard it.
People may also correct for biases that they believe resulted from stereotypes
or from their mood at the time they made a judgment. But there is evidence that
attempts to suppress stereotypical thinking may actually increase stereotyping,
and that people sometimes overcorrect, for example, leading to a less favorable
evaluation of a candidate when they are in a good mood. 37
36. Richard E. Petty, Duane T. Wegener, and Monique A. Fleming, “Flexible Correction
Processes: Perceived Legitimacy Of Bias And Bias Correction,” paper presented at the
annual meeting of the Society for Experimental Social Psychology (Sturbridge, MA:
1996).
37. Leonard Berkowitz et al., On the Correction of Feeling-Induced Judgmental Biases
131, in Feeling and Thinking: The Role of Affect in Social Judgments (New York:
Cambridge University Press, 2000); Linda M. Isbell and Robert S. Wyer, Jr., Correcting for
Mood-Induced Bias in the Evaluation of Political Candidates: The Roles of Intrinsic and
Extrinsic Motivation, 25 Personality & Soc. Psychol. Bull. 237 (1999). See also Ehud
Guttel, Overcorrection, 93 Geo. L.J. 241 (2004) (noting that the legal system sometimes
corrects for overcorrection); C. Neil Macrae et al., Out of Mind But Back in Sight: Stereotypes
on the Rebound, 67 J. Personality & Soc. Psychol. 808 (1994).
258 problem solving, decision making, and professional judgment
retrieved at some later date and to affect later inference.”38 Factors that contrib-
ute to the vividness of information include:
• Emotional interest. Events in which we or people we know are involved have
more emotional interest than those involving strangers.
• Concreteness. Even events involving strangers can be emotionally gripping
if they are described in sufficient detail to prompt imagery. Compare “Jack
was killed by a semi-trailer that rolled over his car and crushed his skull” to
“Jack sustained fatal injuries in a car accident.” Concrete descriptions have
more information, which, even if redundant or irrelevant to the attribution
of causation, is more likely to be remembered. Statistics may be the least
concrete form of information.
• Direct experience. First-hand information has more salience than the reports
of others. By the same token, a face-to-face recommendation is likely to be
more effective than the same recommendation in writing.
In addition to its availability, vivid information is more likely to recruit addi-
tional information from one’s memory and people are more likely to rehearse, or
mull over, vivid information, making it even more memorable.
The normative problem of using vividness, and availability more generally, as
criteria for inference is that they are not always strongly correlated with their
evidential value. This is illustrated by a number of experiments in which vivid-
ness caused participants to generalize from obviously biased samples. For exam-
ple, Ruth Hamill,Timothy Wilson and Richard Nisbett showed participants a
videotape interview of a prison guard, which depicted him as either humane or
brutal. Some were told that the guard was typical, some that he was highly atypi-
cal. When later asked to generalize about prison guards, both groups of partici-
pants concluded that prison guards generally tended to be like the one they had
seen in the interview.39 Vivid anecdotal evidence, can have a powerful effect on
people’s perceptions.
With a summer season framed by a shark attack on a boy in Florida two days after the
Fourth of July and the death of a man on Labor Day on the Outer Banks of North
Carolina, the danger from sharks could easily be seen as rising dramatically.
This is being seen as the “Summer of the Shark” as Time magazine put it in a July 30
cover story bristling with images of razor-sharp teeth . . .
But notwithstanding the bloody attacks . . . scientists said the new fears were overblown.
There is no rampage. If anything, they say, the recent global trend in shark attacks is
down, even as news media attention soars.40
. . . [Y]ou cannot talk sensibly about shark attacks this summer unless you talk first of
the sheer number of humans who flock to the beaches in the summertime. Shark attacks
will drop off precipitously now that Labor Day has come, because there will be less
human flesh in the water to be bitten . . .
Twenty-eight children in the United States were killed by falling television sets between
1990 and 1997 . . . [T]hat is four times as many people as were killed by great white
sharks in the 20th century. Loosely speaking, this means that “watching ‘Jaws’ on TV is
more dangerous than swimming in the Pacific.”41
In the aftermath of the terrorist attacks of September 11, 2001, on the World
Trade Center and on the Pentagon, and the SARS epidemic in Toronto in 2003,
Neal Feigenson and his colleagues studied Americans’ and Canadians’ percep-
tions of the two risks.42
Canadians estimated the percentage chance that they would become seriously
ill or die from SARS within the next year as 7.43%, and the percentage chance
that they would suffer a similar outcome from terrorism as 6.04%, a signifi-
cant difference. By contrast, Americans estimated the percentage chance that
they would become seriously ill or die from SARS within the next year as
2.18%, and the percentage chance that they would suffer a similar outcome
from terrorism as 8.27%, also a significant difference.
The authors estimate that the probability of a Canadian’s contracting SARS
was in fact less than .0008 percent and (albeit more speculatively) that the prob-
ability of an American dying from terrorism within the year was about .001 per-
cent. They explain the different perceptions in terms of “systematically different
media coverage of those risks, making those risks differently available to them”:
[A] sampling of national and local print coverage of SARS and terrorism indi-
cates that Canadian media sources devoted about 40% more articles to SARS
than American media sources did, while more than 14 times as many articles
about terrorism appeared in American as opposed to Canadian print media.
These threat by country interactions roughly parallel the national patterns of
respondents’ perceptions of SARS and terrorism risks.
40. William J. Broad, Scientists Say Frenzy Over Shark Attacks Is Unwarranted, New York
Times, Sept. 5, 2001.
41. The Statistical Shark, New York Times, Sept. 6, 2001.
42. Neal Feigenson, Daniel Bailis, and William Klein, Perceptions Of Terrorism And
Disease Risks: A Cross-National Comparison, 69 Mo. L. Rev. 991 (2004).
260 problem solving, decision making, and professional judgment
Paul Slovic and his colleagues have conducted numerous surveys of people’s
estimates of the frequency of causes of death. The social scientists found a strong
correlation between media coverage—especially the coverage of spectacular
events—and errors in estimation. “In general, rare causes of death were overes-
timated and common causes of death were underestimated . . . [A]ccidents were
judged to cause as many deaths as diseases, whereas diseases actually take about
15 times as many lives. Homicides were incorrectly judged to be more frequent
than diabetes and stomach cancer. Homicides were also judged to be about as
frequent as stroke, although the latter annually claims about 11 times as many
lives. Frequencies of death from botulism, tornadoes and pregnancy (including
childbirth and abortion) were also greatly overestimated.”43
Chip Heath and his colleagues studied media coverage of mad cow disease
outbreaks in France to see how language frames affect the public’s perception of
risk.44 Monthly beef consumption was affected more by the number of articles
that referred to the disease as “mad cow” than by articles that referred to the
disease by one of its scientific names, such as bovine spongiform encephalopa-
thy (BSE.). The “mad cow” articles tended to deal less with the scientific research
conducted on the disease and more with its sensational personal and social ram-
ifications, whereas the BSE articles were more likely to include technical content
and cite current studies. The authors conclude that “mad cow,” which was used
far more often than the scientific names, made the public more risk averse in
consuming beef. (Policy makers’ responses seemed more closely aligned with
scientific reports in the media, suggesting a degree of deliberation independent
of their constituents’ fears.)
Returning to the earlier discussion of fluency as a source of judgment,
research by Hyunjin Song and Norbert Schwarz suggests that people perceive
objects, diseases, foods, etc. with difficult to pronounce names to be more risky
or threatening than those with familiar-sounding names.45 In particular, when
people were asked to rate the dangerousness of a food additives named magnal-
roxate versus hnegripitrom, or roller-coaster rides named Chunta versus
Vaiveahtoishi the latter, more difficult-to-pronounce additives and rides were
judged to be more dangerous. The researchers conclude that easy-to-pronounce
or seemingly familiar names can be processed fluently, which gives rise to
the belief that one has “heard it before,” and therefore to the belief that it is
relatively safe.
43. Paul Slovic, Baruch Fischhoff, and Sarah Lichtenstein, Rating the Risks, in Paul
Slovic, The Perception of Risk 105 (London: Earthscan Publications, 2000).
44. Marwan Sinaceur, Chip Heath, and Steve Cole, Emotional and Deliberative Reactions
to a Public Crisis: Mad Cow Disease in France, American Psychological Society 247–54
(2005), http://faculty-gsb.stanford.edu/heath/documents/PsychSci-Mad%20Cow.pdf.
45. Hyunjin Song, and Norbert Schwarz, If It’s Difficult to Pronounce, It Must Be Risky,
20 Association for Psychological Science 135–38 (2009).
biases in perception and memory 261
The duet from Gigi quoted in Section 9.2 exemplifies how participants or observ-
ers may recall the same event differently even when there is no ostensible reason
for the difference. Beyond this, people’s recollections and judgments tend to be
clouded by an egocentric bias. When married couples estimate their proportion-
ate contribution to household tasks, the sum exceeds 100 percent.46 So too of
collaborators’ estimates of their contributions to joint projects, such as this book.
While the phenomenon may be partially explained by self-serving motives or
having inflatedly positive views of oneself, it extends to negative behaviors as
well. Spouses report causing, as well as resolving, most of the problems in their
relationship,47 and people are prone to overestimate their own errors and unat-
tractive behaviors.48
One plausible explanation for the egocentric bias is that our own actions are
more readily accessible to our memories than the actions of others. Furthermore,
our recollection of behaviors may be guided by a theory of one’s own behavior.
Rather than go through the laborious process of recalling particular activities, we
ask ourselves, “how often do I do this kind of thing?” and take the further short-
cut of looking to our disposition by asking, “am I the sort of person who does
this?” Of course, we might ask the same questions about the other person. But
either because we find it “difficult or time consuming to retrieve two separate
pieces of information and combine them to procedure a single judgment,” or
because we anchor on our own contributions and then (insufficiently) adjust for
the other’s, it is our own behavior that dominates.49
Even when an individual holds himself accountable for an outcome vis-à-vis
others, the question remains open whether the outcome was due to internal fac-
tors such as skill or its absence, or to external factors, such as bad luck. In an
analysis of newspaper accounts of sporting events, Richard Lau and Dan Russell
found that players attributed good outcomes to their skill 75 percent of the time,
but took responsibility for bad outcomes only 55 percent of the time.
Sportswriters—who one might expect to have more distance—were less likely
than players or coaches to attribute wins to internal factors, but even sportswriters
46. Michael Ross and Fiore Sicoly, Egocentric Biases in Availability and Attribution, 37
Journal of Personality and Social Psychology 322 (1979).
47. See Linda Babcock et al., Biased Judgments and Fairness in Bargaining, 85 American
Economic Review 1337, 1337–38 (1995).
48. Nisbett and Ross, supra at 76; Justin Kruger and Thomas Gilovich, Naive Cynicism,
in Everyday Theories of Responsibility Assessment: On Biased Perceptions of Bias, 76 Journal
of Personality and Social Psychology 743–53 (1999).
49. Suzanne C. Thompson and Harold H. Kelley, Judgments of Responsibility for Activities
in Close Relationships, 41 Journal of Personality and Social Psychology 469 (1981).
262 problem solving, decision making, and professional judgment
were more likely to provide internal attributions for their home teams’ wins than
for losses.50
Egocentrism is often correlated with self-servingness. Among people working
different amounts of time in a joint enterprise, those who work more believe
they should be paid more, while those who work less believe that both parties
should be paid equally. People may come to self-serving results indirectly, by
disproportionately valuing the particular task they contribute to the joint enter-
prise (e.g., washing the laundry compared to sweeping the floors). But people
often believe they are entitled to a greater share than others even in the absence
of joint activities. For example, in a simulation of a commons problem involving
fishing associations that needed to reduce their harvests to preserve the stock of
fish for long-run profits, participants believed they were entitled to a greater por-
tion of the catch than the others.51
50. Richard R. Lau and Dan Russell, Attributions in the Sports Pages, 39 Journal of
Personality and Social Psychology 29 (1980).
51. Kimberly A. Wade-Benzoni, Ann E. Tenbrunsel, and Max H. Bazerman, Egocentric
Interpretations Of Fairness In Asymmetric, Environmental Social Dilemmas: Explaining
Harvesting Behavior and the Role of Communication, 67 Organizational Behavior and
Human Decision Processes 111–26 (1996).
52. Eugene M. Caruso, Nicholas Epley, and Max H. Bazerman, The Good, The Bad,
And The Ugly Of Perspective Taking In Groups, in E. A. Mannix, M. A. Neale, &
A. E. Tenbrunsel, (series eds), Research On Managing Groups And Teams, 8 Ethics in
Groups 201–224 (London: Elsevier, 2006).
53. Eugene M. Caruso, Nicholas Epley, and Max H. Bazerman. When Perspective
Taking Increases Taking: Reactive Egoism in Social Interaction, 91 Journal of Personality
and Social Psychology 972 (2006).
biases in perception and memory 263
participants actually took in the simulation. In a sense, cynicism about the other
participants resulted in reactive egoism.54
9.7.2 The Recollection of One’s Own Past Beliefs, Feelings, and Events
Herein lies a difficulty in any autobiographical sketch which purports to deal with one’s
mental development. It is a story of oneself in the past, read in the light of one’s present
self. There is much supplementary inference—often erroneous inference—wherein “must
have been” masquerades as “was so.”
—C. Lloyd Morgan, History of Psychology in Autobiography (1930)55
Research suggests that emotions may not be stored directly in memory but
rather “are reconstructed based on memories of the emotion-eliciting
circumstances.”56 Reconstructions may go beyond emotions to involve the mem-
ories of attitudes and even actions. Michael Ross57 argues that such memories
are based on people’s implicit theories of human stability and change. All things
being equal, we think that attitudes are consistent over time. For this reason,
when we infer past attitudes from present ones, we tend to exaggerate their con-
sistency. For example, participants whose views changed about school busing to
achieve desegregation or about the health effects of vigorous physical exercise
misremembered their earlier views; university students whose attitudes toward
a dating partner had changed over several months recalled their earlier impres-
sion as being more consistent with their current attitudes than was the case.
Furthermore, because we tend to seek consistency in attitudes and behavior, a
change in attitudes can produce inaccurate recollection of behaviors. Thus, when
experimenters caused participants to believe that regularly brushing their teeth
was not as important as they had thought, they recalled less brushing than they
actually had done.
While a belief in personal consistency is the norm, we also have theories of
change. So, for example, people who have engaged in education and self-
improvement programs tend to exaggerate the difference between their past and
the present knowledge or behavior. In an experiment involving a program to
improve study skills, students recalled their original assessment of their study
skills to be worse than they had originally reported and reported greater improve-
ment than the facts warranted. They not only overpredicted the resulting
54. Kruger and Gilovich; Epley, Caruso, and Bazerman. However, selfish behavior was
reduced when the group project was structured or characterized as cooperative rather
than competitive.
55. Quoted in Michael Ross, Relation of Implicit Theories to the Construction of Personal
Histories, 96 Psychological Review 341 (1989).
56. Id.
57. Id.
264 problem solving, decision making, and professional judgment
improvement in their grades, but subsequently recalled the new grades as being
higher than they actually were.
58. Lee Ross and Donna Shestowsky, Contemporary Psychology’s Challenge to Legal
Theory of Practice, 97 Nw. U. L. Rev. 1081 (2003).
biases in perception and memory 265
would agree with them if apprised of the “real” facts; partisans therefore
assume that disinterested third parties would agree with them.
• Partisans tend to see viewpoints that differ from their own as highly reveal-
ing both of personal dispositions (for example, gullibility, aggressiveness,
pessimism, or charitableness) and of various cognitive and motivational
biases. In fact, differences in judgment often reflect differences in the way
a given issue or object of judgment is perceived and construed rather than
a difference in the perceivers’ values or personality traits.
• When partisans acknowledge the influence of particular features of their
experience or identity, they see it not as a bias but as a source of enlighten-
ment, while they see the other parties unique experiences or identities as a
source of bias.
• Partisans on both sides of an issue will typically perceive evenhanded
media to be biased against them and to favor their adversaries. They are apt
to see the same hostile bias in the efforts and decisions of evenhanded
third-party mediators or arbitrators.
• Partisans will be polarized and extreme in their view of others. Once they
believe a person to be biased, they tend to discount his views entirely, and
will underestimate areas of agreement and underestimate the prospects of
finding “common ground” through discussion or negotiation.
We have seen that being asked to articulate the weaknesses of one’s own posi-
tion may counter perceptual biases in some situations. A broader hope of this
book is that students and readers who understanding those biases and the phe-
nomena of naïve realism will be less likely to fall into their traps. But we should
be among the first to admit that the hope does not rest on much of an empirical
foundation.
This page intentionally left blank
10. biases in processing and judging
information
1. Amos Tversky and Daniel Kahneman, Judgment Under Uncertainty: Heuristics and
Biases. 185 Science, 1124–1130 (1974).
2. Edward J. Joyce and Gary C. Biddle, Anchoring and Adjustment in Probabilistic
Inference in Auditing, 19 Journal of Accounting Research 120 (1981).
268 problem solving, decision making, and professional judgment
3. Gregory B. Northcroft and Margaret A. Neale, Experts, Amateurs, and Real Estate:
An Anchoring-And-Adjustment Perspective on Property Pricing Decisions, Organisational
Behavior and Human Decision Processes 39 (1987).
biases in processing and judging information 269
made the final offer seem comparatively generous.4 And Jeff Rachlinski, Chris
Guthrie, and Andrew Wistrich presented federal judges with the facts of a per-
sonal injury case, in which the plaintiff had been hospitalized for several months
and was left paraplegic and confined to a wheelchair. One group was anchored
by being informed that the defendant had moved for dismissal on the ground
that the case did not meet the $75,000 jurisdictional minimum for a diversity
case;5 the other group was given no such anchoring information. Though almost
none of the judges granted the motion to dismiss, the average award in the
anchored group was $882,000 compared to $1,249,000 in the unanchored
group.6
Anchoring in Jury Awards. In an experiment by John Malouff and Nicola
Shutte, when plaintiff’s lawyer requested $100,000 damages, mock jurors
awarded $90,000, and when plaintiff’s lawyer requested $500,000 damages, they
awarded nearly $300,000 on average.7 While jurors assessing punitive damages
often anchor on the amount of actual damages, they also anchor on the amount
including punitive damages requested in the closing argument of the plaintiff’s
attorney. In an experiment involving a train derailment that dumped herbicide
and caused serious environmental damage, plaintiff’s attorney told one group of
mock jurors: “So the range you may want to consider is between $15 million and
about a half a year’s profit, $50 million”; another group was given the range
between $50 million and $150 million. The median awards by the two juries
4. Russell Korobkin and Chris Guthrie, Opening Offers and Out-of-Court Settlements: A
Little Moderation May Not Go a Long Way, 10 Ohio State Journal on Dispute Resolution
1 (1994).
5. Federal courts can hear cases arising under state law if the amount in controversy
is $75,000 or more and the plaintiff and defendant reside in different states.
6. Chris Guthrie, Jeffrey J. Rachlinski, and Andrew J. Wistrich, Inside the Judicial
Mind, 86 Cornell L. Rev. 777 (2001).
7. John Malouff and Nicola S. Schutte, Shaping Juror Attitudes: Effects of Requesting
Different Damage Amounts in Personal Injury Trials, 129 Journal of Social Psychology
491 (1989).
270 problem solving, decision making, and professional judgment
were $15 million and $50 million, respectively.8 Experimental data also suggests
that statutory caps on damages may affect settlement rates.9
8. Reid Hastie, David A. Schkade, and John W. Payne, Do Plaintiffs’ Requests and
Plaintiffs’ Identities Matter?, in Cass R. Sunstein et al., Punitive Damages: How Juries
Decide (Chicago: University of Chicago Press, 2002).
9. A low cap increased settlement rates while a high cap decreased them. The authors
term the latter phenomenon, “motivated anchoring,” in which a relatively high damage
cap disproportionately anchors the plaintiff’s estimate of the likely damage award. The
result is a widened disparity in opposing litigants’ judgments, and less settlement. Greg
Pogarsky and Linda C. Babcock, “Damage Caps, Motivated Anchoring, and Bargaining
Impasse” (Aug. 1, 2000), available at SSRN: http://ssrn.com/abstract=235296
10. Fritz Strack and Thomas Mussweiler, Explaining the Enigmatic Anchoring Effect:
Mechanisms of Selective Accessibility, 73 Journal of Personality and Social Psychology
437 (1997).
11. Daniel M. Oppenheimer et al., Anchors Aweigh: A Demonstration of Cross-Modality
Anchoring and Magnitude Priming, Cognition (2007). When the anchors were placed in
appropriately scaled contexts—e.g., a shorter line tracing the length of the Mississippi
on a small scaled map, and a longer line tracing its length on a larger scaled map—the
influence of the anchoring effect diminished.
biases in processing and judging information 271
away from the initial anchor are insufficient. Under this view, we “adjust the
anchor until shortly after it enters a range of plausible values for the target item.
Thus, when adjusting from a high anchor, decision makers stop at the high end
of plausible values, but stop at the low end when adjusting from low anchors.”
Adjustment is cognitively effortful, and a “lack of effort or lack of cognitive
resources cause adjustment to be terminated too soon, resulting in a final
response that is too close to the anchor.”12
Alternatively, the anchor may increase the availability of features that the
anchor and the target—e.g., the number to be determined—hold in common,
selectively activating information about the target that is consistent with the
anchor (while not activating other information).13 Gretchen Chapman and Eric
Johnson hypothesize:
Assume a decision maker, when requested to make a judgment, can access a
variety of features about the target item, either from memory or from the
external environment. Some subset of this information is retrieved and used
as the basis for . . . judgment. If a representative sample is retrieved, the judg-
ment will be unbiased. However, . . . the presence of an anchor increases the
activation of features that the anchor and target hold in common while reduc-
ing the availability of features of the target that differ from the anchor.
For example, suppose a decision maker asked to evaluate a bus trip was
presented with a low anchor [for the cost of the bus ticket]. The decision maker
might contemplate a bumpy, noisy, crowded ride aboard a city-owned vehicle
that spouts diesel fumes and has hard plastic seats and no air conditioning.
Conversely, if a high anchor were presented, the decision maker might
imagine a luxurious tour bus with reclining plush velour seats, sound system,
and bar . . ..
Under either of these explanations, a person might sometimes believe that
the anchor provides information relevant to the actual amount. This might be
true in the fraud, real estate, and jury award examples described above (though
certainly not in anchoring on the percentage of African countries in the United
Nations based on the spin of a wheel.) Even so, the phenomenon is likely to
result in a biased judgment.
12. Gretchen B. Chapman and Eric J. Johnson, Incorporating the Irrelevant: Anchors
in Judgments of Belief And Value, in Thomas Gilovich, Dale Griffin, and Daniel
Kahneman, Heuristics and Biases, The Psychology of Intuitive Judgment 127
(2002) [hereafter, Heuristics and Biases].
13. Gretchen B. Chapman and Eric J. Johnson, Anchoring, Activation, and the
Construction of Values, 79 Organizational Behavior and Human Decision Processes
115 (1999).
272 problem solving, decision making, and professional judgment
10.1.5 Debiasing
The insufficient adjustment and activation theories may account for anchoring
phenomena in different circumstances. In any event, the phenomenon is extraor-
dinarily robust. Simply warning people about it does not decrease anchoring;
nor does offering incentives for accurate estimates.14
Baruch Fischhoff, who pioneered research in the subject, describes hindsight bias
as follows:
In hindsight, people consistently exaggerate what could have been anticipated
in foresight. They not only tend to view what has happened as having been
inevitable but also to view it as having appeared “relatively inevitable” before
it happened. People believe that others should have been able to anticipate
events much better than was actually the case.15
In a classic experiment, Fischhoff provided a description of the 1814 war
between the British and the Nepalese Gurkhas and informed participants that it
had four possible outcomes: British victory, Gurkha victory, military stalemate
with no peace settlement, or stalemate with a peace settlement. In each of four
conditions participants were told that one of the four outcomes had actually
occurred and were then asked to estimate its prior probability as well as those of
the three alternatives. As Fischhoff had predicted, participants’ knowledge of
what they believed to be the actual outcome significantly affected their estimates
of the prior probability.
In an experiment examining hindsight bias in a legal context, Kim Kamin and
Jeffrey Rachlinski gave two groups of participants the same facts, except that one
group was asked whether to take precautions to avert a potential accident and the
second was asked whether to assess liability after an accident had already
occurred:
Foresight condition. Participants were told that a city had constructed a draw-
bridge and needed to determine whether the risk of a flood warranted main-
taining a bridge operator during the winter when the bridge was not in use.
Hiring the operator would serve as a precaution. The operator would monitor
weather conditions and raise the bridge to let debris through if the river
16. See Mark Kelman, David Fallas and Hillary Folger, Decomposing Hindsight Bias, 16
Journal of Risk and Uncertainty 251 (1998).
274 problem solving, decision making, and professional judgment
case illustrates, this is an essential part of the negligence standard in tort liability
and plays a significant role in many other legal domains, ranging from trust and
estate law (did the trustee invest the beneficiary’s funds prudently?) to substan-
tive criminal law (did the defendant claiming self-defense reasonably believe that
he was in danger of being killed?) and criminal procedure (did the police officer
have probable cause for the arrest and search?). In all of these situations, the trier
of fact is vulnerable to hindsight bias.
Summarizing the rich psychological literature since this 1975 study, Jeffrey
Rachlinski17 writes that hindsight bias
results primarily from the natural (and useful) tendency for the brain to auto-
matically incorporate known outcomes into existing knowledge, and make
further inferences from that knowledge. For example, participants in the
British-Gurkha study who were told that the British had won, probably made
inferences about colonial warfare, generally. They may have concluded, for
example, that the advantages that the British had over the Gurkhas (better
weapons and training) were more important in such conflicts than the Gurkha
advantages (better knowledge of the terrain and higher motivation).
Participants informed of a Gurkha victory may have inferred the opposite.
When asked to estimate the [prior] probability of a British victory, the partici-
pants relied on their new belief that better weapons and training were more
important than knowledge of the terrain and higher motivation. These infer-
ences then induced them to make high estimates for the probability of
the known outcome, relative to the estimates of people who were unaware of
the outcome and hence had not developed such beliefs.
More generally, ignoring a known outcome is unnatural. Normally, people
should integrate new information into their existing store of knowledge and
use it to make future predictions. Assessing the predictability of past events
requires that the outcome and the subsequent inferences that depend upon
learning the outcome be ignored. Perhaps because such judgments are so
uncommon and unnatural, people cannot make them accurately.
Hindsight bias is obviously related to the availability heuristic. The concrete-
ness and vividness of the actual outcome of an event prompts us to explain it,
and hindsight explanations are usually easy to provide. Moreover, we do not usu-
ally develop casual scenarios for other possible, and even more probable, out-
comes that did not occur—although we could have readily explained those
outcomes had they occurred. The ease of explanation, and the relative availability
of explanations for occurring versus non-occurring events, becomes, in effect, a
heuristic for assessing their likelihood.
17. Jeffrey Rachlinski, Heuristics and Biases in the Courts: Ignorance or Adaptation?, 69
Oregon L. Rev. 61 (2000).
biases in processing and judging information 275
Consider, for example, the many ways that another terrorist attack could be
launched against the United States—explosives in air freight shipments, suicide
bombers in shopping malls, dirty bombs, blowing up a chemical plant, bioterror-
ism, etc. Especially because we have heard warnings about all of these and many
other possibilities targeted at various locations, if and when an attack does occur
the causal chain will be readily available in hindsight. And, of course, people will
be held accountable for having failed to take adequate precautions against the
attack.18
18. For an argument that takes account of hindsight bias and nonetheless concludes
that airlines and government officials should have anticipated the 9/11 attack, see Max H.
Bazerman and Michael D. Watkins, Predictable Surprises: The Disasters You
Should Have Seen Coming and How to Prevent Them (Boston: Harvard Business
School, 2004).
19. W. Kip Viscusi, Corporate Risk Analysis: A Reckless Act?, 52 Stan. L. Rev. 547
(2000).
276 problem solving, decision making, and professional judgment
20. Norbert Schwarz and Leigh Anne Vaughn, The Availability Heuristic Revisited:
Ease of Recall and Content of Recall as Distinct Sources of Information, in Heuristics and
Biases 103.
biases in processing and judging information 277
setting the standard of proof . . . operate[s] against the party most likely to benefit
from the hindsight bias—the plaintiff.” And he goes on to describe other legal
mechanisms:
First, courts suppress evidence that would exacerbate the bias, such as subse-
quent remedial measures in negligence and products liability cases. Second,
they rely on standards of conduct developed ex ante, when reliable standards
are available. Examples include the use of custom as a defense in medical
malpractice cases and the business judgment rule in corporate law. Although
accurate assessments of the ex ante probabilities would be superior to these
strategies, courts seem to recognize such assessments will be biased in hind-
sight and are therefore unattainable.
The human understanding when it has once adopted an opinion draws all things else to
support and agree with it. And though there be a greater number and weight of instances
to be found on the other side, yet these it either neglects and despises, or else by some
distinction sets aside and rejects, in order that by this great and pernicious
predetermination the authority of its former conclusion may remain inviolate.
—Francis Bacon, First Book of Aphorisms (1620)
Faced with the choice between changing one’s mind and proving that there is no need to
do so, almost everyone gets busy on the proof.
—John Kenneth Galbraith
[Herbert] Spencer began, like a scientist, with observation; he proceeded, like a scientist,
to make hypotheses; but then, unlike a scientist, he resorted not to experiment, nor to
impartial observation, but to the selective accumulation of favorable data. He had no
nose at all for “negative instances.” Contrast the procedure of Darwin, who, when he
came upon data unfavorable to his theory, hastily made note of them, knowing that they
had a way of slipping out of the memory a little more readily than the welcome facts.
—Will Durant, The Story of Philosophy 296 (1926)
Recall than in the examples of Bayesian statistics in Section 8.5, P(H), the
prior probability—or the strength of our belief in the hypothesis prior to obtain-
ing event-specific data—was determined by the base rate of the event in ques-
tion. (In our example, the base rate was the proportion of Blue taxis in a town.)
Bayes’ Theorem tells us how to revise an estimate after gaining new informa-
tion—say, about the eyewitness’s identification of the taxi involved in the
accident.
278 problem solving, decision making, and professional judgment
Consider, though, that at any given time we hold many beliefs that cannot
readily be described in terms of base rates—for example, the belief that, as of
2003, Iraq did or did not possess weapons of mass destruction. Those beliefs are
based on a variety of sources, including our own experience and the assertions of
others—teachers, scientists, politicians, journalists—which we may deem more
or less accurate or trustworthy. In these myriad instances, P(H) does not reflect
a calculable probability, but rather the subjective strength of our prior belief in
H. When we are called upon to revise a prior belief in terms of new evidence, the
task is not formulaic as much as intuitively combining the prior belief with the
current data.
In many if not most circumstances, it is entirely rational to give considerable
weight to one’s prior beliefs. However, people tend to interpret conflicting evi-
dence with a variety of confirmation biases, manifesting a pervasive tendency to
seek evidence supporting their prior beliefs or hypotheses, and to ignore or den-
igrate evidence opposing them.21 The belief by the CIA and other members of
the Western Intelligence Community that Iraq had weapons of mass destruction
provides a recent and dramatic example of the phenomenon. This was, in effect,
the conclusion of the Senate Select Committee on Intelligence:22
The Intelligence Community (IC) suffered from a collective presumption that
Iraq had an active and growing weapons of mass destruction (WMD) pro-
gram. This . . . led Intelligence Community analysts, collectors, and managers
to both interpret ambiguous evidence as conclusively indicative of a WMD
program as well as ignore or minimize evidence that Iraq did not have active
and expanding weapons of mass destruction programs . . ..
. . . [I]ntelligence analysts, in many cases based their analysis more on their
expectations than on an objective evaluation of the information. Analysts
expected to see evidence that Iraq had retained prohibited weapons and that
Iraq would resume prohibited WMD activities once United Nations (UN)
inspections ended . . ..
The . . . IC had a tendency to accept information which supported the pre-
sumption that Iraq had active and expanded WMD programs more readily
than information which contradicted it. . . . Information that contradicted the
IC’s presumption that Iraq had WMD programs, such as indications in the
intelligence reporting the dual-use materials were intended for conventional
or civilian programs, was often ignored . . ..
The presumption that Iraq had active WMD programs affected intelligence
collectors as well. None of the guidance given to human intelligence collectors
21. See generally Joshua Klayman, Varieties of Confirmation Bias, 32 The Psychology
of Learning and Motivation 385 (New York: Academic Press, 1995).
22. The committee report argued that the confirmation bias was caused or reinforced
by “group think.”
biases in processing and judging information 279
• You might have trouble generating viable new hypotheses, even when you
do feel like abandoning an old one.
The opportunity to seek confirming or disconfirming evidence can take place
at different phases of developing and testing hypotheses, and what is essential in
one phase may be counterproductive in another. Ryan Tweney et al. write:23
Success in a very difficult inference task requires that hypotheses be “good”
ones in the sense that they must possess at least some correspondence to real-
ity to be worthwhile candidates for the test. Thus, seeking confirmatory data
and ignoring disconfirmatory data may be the most appropriate strategy early
in the inference process, since it could then produce good hypotheses. Once
such hypotheses have been found, however, deliberate attempts ought to
be made to seek disconfirmatory data, as the strongest possible test of an
explanation.
Most of the instances of confirmation bias examined below do not involve the
development of hypotheses, but their testing.
26. The underlying assumption of the self-affirmation theory is that people are ulti-
mately motivated to see themselves as good. Claude M. Steele, The Psychology of Self-
Affirmation: Sustaining the Integrity of the Self, in L. Berkowitz (ed.,) 21 Advances in
Experimental Social Psychology 261 (New York: Academic Press, 1988). Most motivational
biases serve that purpose—a good (or intelligent) person would not drink a lot of coffee if
doing so were unhealthy, so drinking coffee must not be that unhealthy. Self-affirmation
doesn’t make people more skeptical about welcome news.
282 problem solving, decision making, and professional judgment
27. Charles G. Lord, Lee Ross, and Mark R. Lepper, Biased Assimilation and Attitude
Polarization: The Effects of Prior Theories on Subsequently Considered Evidence, 37 Journal
of Personality and Social Psychology 2098–2109 (1979).
28. Consider, in this respect, scientists’ belief in the theory of evolution. From time to
time one comes across complex functions—grist for the mill of so-called “intelligent
biases in processing and judging information 283
design”—that are not readily explicable by natural selection. But the theory has proved
robust in the 150 years since Darwin, and such phenomena have always ended up being
accounted for.
29. Jonathan J. Koehler, The Influence of Prior Beliefs on Scientific Judgments of Evidence
Quality, 56 Organizational Behavior and Human Decision Processes 28 (1993).
284 problem solving, decision making, and professional judgment
and continuing to hear the answer “yes,” he will never be able to prove his
hypothesis, because he will never have tested alternative hypothesis. It would be
more beneficial to try sequences like “3 5 7” or “3 6 9,” for which a “yes” would
disprove the hypothesis.30 Superstitious behaviors are often premised on positive
hypothesis testing: if you were wearing your favorite green striped socks when
you aced the exam, then you might as well wear them next time. (And if you
don’t do so well the next time, we’ve seen that there are lots of ways to explain
that away.)
The second study involves a logic problem, essentially asking people to apply
the rule, if p then q:
Imagine four cards with the symbols E, K, 4, and 7 written on them. Each card
has a number on one side and a letter on the other. I tell you: “Every card with
a vowel on one side has an even number on the other,” and ask you to test
whether this statement is true. If you were only allowed to turn over two
cards, which two cards would they be?31
When asked this question, most people choose the cards marked with E
and 4.32 In fact, this is not the best way to test whether the rule is true.
• Turning over the E card is a good idea, because finding an odd number on
the other side would disprove the rule.
• But turning over the 4 card is not informative if the rule is false: it can only
provide confirming evidence, giving us another instance in which the rule
turns out to be true. Consider the two possibilities: If there is a vowel on
the other side, it is consistent with the rule but cannot prove it. If there is a
consonant on the other side, the card is irrelevant. since the statement tells
us what to expect only for cards with a vowel on one side. Thus the 4 card
cannot give us evidence against the statement.
The correct choice is E and 7. If there is a vowel on the other side of the 7
card, or an odd number on the other side of the E card, the statement is false. If
there is a consonant with the 7 and an even number with the E, the statement
is true.
It is difficult to see in these examples the motives for confirmation bias
that were so apparent in the death penalty study. Rather, they seem to be instances
of a more cognitive, dispassionate form of bias involving the tendency to engage
in positive hypothesis testing, that is, testing hypotheses by looking for evidence
that would support rather than disconfirm them. Participants in an experiment
who were asked to determine whether a person was an extrovert asked questions
such as, “What would you do if you wanted to liven things up at a party?”
When asked to determine whether he was an introvert, participants asked questions
such as, “What makes you the most uneasy or anxious in social situations?”33
It turns out that people are better at hypothesis testing when solving prob-
lems that are familiar or situated in readily imaginable domains. For example,
Hubert L. Dreyfus and Stuart E. Dreyfus note that that people have less difficulty
with this problem than with the logically identical card-turning problem described
above:34
Suppose that you are a supervisor at a supermarket and have the checks
received that day stacked before you—some face up and some face down. The
supermarket has a rule: cashiers may only accept checks for $50 or more if
approved on the back by the manager. Which checks do you need to turn over
to see if the cashiers have followed the rule?35
As we discussed in the preceding chapters (see particularly the discussion of
diagnosticity in Section 8.5.3), when testing whether a particular hypothesis is
more likely than an alternative, one must consider not only the probability that
the data support the hypothesis, but also the probability that the data support the
alternative. This is negative hypothesis testing, which requires looking for data that
33. Mark Snyder and William B. Swann, Jr. Behavioral Confirmation in Social
Interaction: From Social Perception to Social Reality, 14 Journal of Experimental Social
Psychology 148 (1978). In a related experiment, participants read a story about a woman
with an equal number of behaviors that could be characterized as introverted (e.g., spend-
ing her office coffee break by herself) or extroverted (e.g., chatting with another patient in
the doctor’s office). Several days later, some were asked to assess her suitability for a ste-
reotypically introvert job (research librarian) and others for an extrovert job (real estate
salesperson). Those who evaluated her for the two positions respectively recalled her
introverted and extroverted behaviors and evaluated her accordingly. Mark Snyder and
Nancy Cantor, Testing Hypotheses about Other People: The Use of Historical Knowledge, 15
Journal of Experimental Social Psychology 330 (1979).
34. Hubert L. Dreyfus, Stuart E. Dreyfus, and Tom Anthanasiou, Mind Over
Machine: The Power of Human Intuition and Expertise in the Era of the
Computer 18 (New York: Free Press, 1986). See also Klayman, Varieties of Confirmation
Bias, supra at 403.
35. The evolutionary psychologists Leda Cosmides and John Tooby assert that people’s
ability to solve this problem, or an analogous one involving a bar’s checking IDs to assure that
drinks are not served to minors, involves a particular “cheater-detection” capacity. Jerome H.
Barkow, Leda Cosmides, and John Tooby, The Adapted Mind: Evolutionary Psychology
and the Generation of Culture (New York: Oxford University Press, 1995); Leda Cosmides
and John Tooby, Neurocognitive Adaptations Designed for Social Exchange. In Evolutionary
Psychology Handbook (David M. Buss, ed., New York: Wiley, 2005). For a critical review
of this work, see Mark G. Kelman, The Heuristics Debate: Its Nature and Its Implications
for Law and Policy (New York: Oxford University Press, forthcoming 2010).
286 problem solving, decision making, and professional judgment
would disconfirm the hypothesis. If I observe Alex eating nothing but junk food,
I could hypothesize that Alex likes to eat only junk food. Positive hypothesis test-
ing would involve giving Alex junk food to see if he likes it. Negative hypothesis
testing requires that I give Alex some healthy food to see if he might like that as
well. Note that if I try giving Alex broccoli and he doesn’t like it, that doesn’t
mean I have confirmed my hypothesis. Although broccoli is not junk food, many
people who like healthy food do not like it. The strongest test of my hypothesis
would be to give Alex a lot of different healthy foods and see if he likes any
of them.
In many situations, though, “people rely most heavily on information about
what happens when the presumed cause is present, and much less on informa-
tion about what happens in its absence.”36 They have “cognitive difficulty in
thinking about any information which is essentially negative in its conception.”37
This tendency is implicitly noted by the great psychologist Arthur Conan Doyle,
in this dialogue from The Adventure of Silver Blaze:
Inspector Gregory: “Is there any other point to which you would wish to draw
my attention?”
Holmes: “To the curious incident of the dog in the night-time.”
“The dog did nothing in the night-time.”
“That was the curious incident,” remarked Sherlock Holmes.
Positive hypothesis testing is not necessarily an irrational strategy. It tends to
uncover false positives—i.e., proffered hypotheses not supported by the data—
which often are more costly than false negatives.38 It also makes sense when you
believe that the circumstances in which rules apply arise only rarely.39
the idea that a hypothesis acts as a suggestion, instilling at least a transient belief
in its validity.41 This is consistent with Spinoza’s view (contrary to Descartes’)
that understanding and believing are elided—that people at least tentatively accept
or believe propositions in order to comprehend them, and only then “unbelieve”
propositions that do not comport with their other beliefs.42
41. See Karen E. Jacowitz and Daniel Kahneman, Measures of Anchoring in Estimation
Tasks, 21 Personality and Social Sciences Bulletin 1161 (1995).
42. See Daniel T. Gilbert, Inferential Correction, in Heuristics and Biases.
43. Richard E. Nisbett and Lee Ross, Human Inference: Strategies and
Shortcomings in Social Judgement 175 ff (Englewood Cliffs, NJ: Prentice Hall, 1980).
44. Id. 169.
45. Craig A. Anderson, Mark R. Lepper, and Lee Ross, Perseverance of Social Theories:
The Role of Explanation in the Persistence of Discredited Information, Journal of Personality
and Social Psychology 39 (1980).
46. Lee Ross, Mark R. Lepper, and Michael Hubbard, Perseverance in Self-Perception
and Social Perception: Biased Attributional Processes in the Debriefing Paradigm, 32 Journal
of Personality and Social Psychology 880 (1975).
288 problem solving, decision making, and professional judgment
that they had done well or poorly. Even after being informed that the evaluations
actually had no relationship to their performance, the participants continued to
believe that they were relatively good or bad, not only on the specific suicide note
task but on other tasks involving social sensitivity.
Nisbett and Ross suggest several reasons why people persevere in their beliefs
even after being explicitly told that they were duped: We are likely to recruit evi-
dence from our own memory consistent with our initial beliefs (“I have always
been good at guessing people’s occupations”) and continue to give it weight
because of its availability. Moreover, we are likely to have generated a causal
explanation of why the initial belief is true (“Of course I’m good at telling if sui-
cide notes are real, I’m very empathic”), and we tend to treat the ease of explana-
tion as evidence of its truth (ignoring how easy it is to develop explanations for
almost any outcome). Not only are the recruited evidence and causal theory not
discredited but, like the initial data that stimulated them, they remain available
even after the data turn out to be spurious.47
10.4 overconfidence
Edward Russo and Paul Schoemaker asked 1000 American and European busi-
ness managers to answer ten trivia questions calling for numerical estimates
48. Dan Simon, Lien B. Pham, Quang A. Le and Keith J. Holyoak, Bidirectional
Reasoning in Decision Making by Constraint Satisfaction, Journal of Experimental
Psychology—General 3 128 (1999); Dan Simon, Lien B. Pham, Quang A. Le, and Keith
J. Holyoak, The Emergence of Coherence over the Course of Decision Making, 27 Journal of
Experimental Psychology—Learning, Memory, and Cognition 1250 (2001).
49. See the discussion of connectionist networks in Section 11.3.2.
50. Charles G. Lord, Mark R. Lepper, and Elizabeth Preston, Considering the Opposite:
A Corrective Strategy For Social Judgment. 47 Journal of Personality and Social
Psychology 1231 (1984).
290 problem solving, decision making, and professional judgment
and to provide a “low and a high guess such that you are 90% sure the correct
answer falls between the two.” The questions included.51
• Martin Luther King’s age at death.
• Length of the Nile River in miles.
• Number of books in the Old Testament.
• Diameter of the moon in miles.
• Air distance from London to Tokyo.
Most respondents got considerably more than 10 percent of their answers
wrong. What is interesting about Russo and Schoemaker’s experiment is not the
state of the participants’ substantive knowledge, but their belief that they knew
more than they did. Indeed, not only amateurs answering trivia questions but
professional making judgments in their own domains of expertise tend to be
overconfident. Also noteworthy: although people are overconfident about par-
ticular answers, they are much more realistic in estimating their overall hit rate.52
Unfortunately, most judgments are made one at a time.
People tend to be overconfident about nontechnical items of moderate or
extreme difficulty.53 For example, participants given the nearly impossible task of
discriminating between Asian and European children’s drawings or predicting
fluctuations in stock prices did little if any better than chance (i.e., 50 percent
correct). But their mean confidence levels were between 70 percent to 80 per-
cent.54 Moreover, people become more confident as they are given more infor-
mation, even if the information is not diagnostic.55 Overconfidence diminishes
as the tasks become easier. Indeed, knowledgeable participants responding to
easy questions tend to be underconfident.
Granted that people often tend to be overconfident, is there a correlation
between confidence and accuracy? Alas, not. For example, in a study by Scott
Plous and Philip Zimbardo, citizens were given a description of actual Soviet
and American military actions, with the names of the countries removed. They
were asked to identify which of the superpowers was involved and to say how
confident they were in their answers. The respondents were correct slightly less
than half the time, and their confidence ratings were almost orthogonal to the
accuracy of their responses.56
The relationship between individuals’ assessment of the probability of their
beliefs being true and the actual truth is termed calibration. Meteorology is
among the few professions—along with bookmaking on horse races—in which
judgments prove to be well calibrated. By contrast, the calibration of physicians’
diagnosis of pneumonia in patients examined because of a cough, and their
diagnoses of skull fractures, has proven to be poor. Sarah Lichtenstein, Baruch
Fischhoff, and Lawrence Phillips conjecture:
Several factors favor the weather forecasters. First, they have been making
probabilistic forecasts for years. Second, the task is repetitive; the question to
be answered (Will it rain?) is always the same. In contrast, a practicing physi-
cian is hour by hour considering a wide array of possibilities (Is it a skull
fracture? Does she have strep? Does he need further hospitalization?). Finally,
and perhaps most important, the outcome feedback for weather forecasters is
well defined and promptly received. This is not always true for physicians;
patients fail to return or are referred elsewhere, or diagnoses remain
uncertain.57
In many situations, overconfidence seems to result from the availability and
anchoring and adjustment phenomena. Dale Griffin and Amos Tversky distin-
guish between the strength (extremity on the relevant predictive dimension) of
evidence and its weight (predictive value).58 In statistical terms, strength is analo-
gous to the size of the effect, and weight is analogous to its significance or
reliability (taking into account, for example, the size of the sample). They argue
that people focus on the perceived strength of the evidence based on the degree
to which available evidence is consistent with the hypothesis in question—and
then adjust (insufficiently) for its weight. For example, when judging whether
a coin is biased, people first focus on the proportion of heads and tails in the
sample and only then (insufficiently) adjust for the number of tosses. By the
same token, the fact that a suspect fails a lie detector test is treated as strong
evidence of guilt notwithstanding the poor reliability of such tests.59
56. Scott L. Plous and Philip G. Zimbardo, How Social Science Can Reduce Terrorism
(2004), Chronicle of Higher Education, Sept. 10, 2004, B9-10.
57. Sarah Lichtenstein, Baruch Fischhoff, and Lawrence D. Phillips, Calibration of
Probabilities: The State of the Art to 1980, in Daniel Kahneman, Paul Slovic, and Amos
Tversky, Judgment Under Uncertainty: Heuristics and Biases 306 (1982).
58. Dale Griffin and Amos Tversky, The Weighing of Evidence and Determinants of
Confidence, in Heuristics and Biases, 230.
59. Lyle A. Brenner, Derek J. Koehler, and Yuval Rottenstreich, Remarks on Support
Theory: Recent Advances and Future Directions, in Heuristics and Biases.
292 problem solving, decision making, and professional judgment
The tendency (discussed in Section 8.5.4) to ignore base rates and to over-
value the more immediate and available data also leads to overconfidence. For
example, after interviewing individuals, participants were asked to predict
whether, if given the choice, the interviewees would take a free subscription to
Playboy or The New York Review of Books. Although informed that 68 percent of
all the people interviewed preferred Playboy, participants tended to ignore this
base rate and premise their predictions on their own impressions from the inter-
views. Participants whose predictions went contrary to the base rate were over-
confident, with a mean accuracy of only about 50 percent in the face of a mean
confidence of 72 percent. When they just relied on the base rate they were quite
well calibrated.
Although people tend to be overconfident about their knowledge or predic-
tions of particular items or events, they tend to be better calibrated in terms of
their overall judgment about similar events. For example, entrepreneurs were
overconfident about the success of their own new ventures even when they were
realistic about the general failure rate for such ventures. Griffin and Tversky sug-
gest that the entrepreneurs ignore base rates and focus on factors pertaining to
the individual decision.60 Similarly, “people often make confident predictions
about individual cases on the basis of fallible data (e.g., personal interviews or
projective tests) even when they know that these data have low predictive
validity.”61
Are there individual characteristics that correlate with calibration? While men
and women were equally represented in the Plous and Zimbardo study described
above, two-thirds of the highly confident respondents were male; and about twice
as many of the highly confident respondents distrusted the Soviet Union and
were advocates of strong defense spending than their low-confidence peers.
However, in a different study of relationships between overconfidence and sev-
eral personality measures—authoritarianism, conservatism, dogmatism, and
intolerance of ambiguity—only authoritarianism showed a (modest) correlation
with overconfidence.62 The nature of the task seems to matter more than the
individual who is performing it.63
As for debiasing, merely alerting participants to the difficulty of the task does
not improve calibration. Nor does increasing the motivation to calibrate
accurately. Paralleling a strategy for correcting self-serving biases, however,
calibration did improve when participants were asked to write down all the rea-
sons that contradicted their answers.64 But the most powerful strategy for improv-
ing calibration is clear and immediate feedback.
Overcoming Intuition. People’s overconfidence increases to the extent their
judgments rely on intuition. Could overconfidence be overcome by “forcing” the
brain to think more deliberately? A 2007 finding by Adam Alter, Daniel
Oppenheimer, and their colleagues indicates that intuition (and the associated
overconfidence) can be overcome when faced with “metacognitive difficulty”—-
that is, when presented with information that makes them think harder about
their intuitive assumptions and judgments. The authors postulate that cognitive
disfluency (which occurs when information pertaining to a decision is difficult
to process) can compel individuals to switch from System 1 cognitive processes
(intuition) to System 2 cognitive processes (deliberation).65 In other words, cog-
nitive disfluency makes individuals “less confident” in their intuitive judgments
and forces them to switch into a more deliberative thinking mode.
64. Asher Koriat, Sarah Lichtenstein, and Baruch Fischhoff, Reasons for Confidence,
6 Journal of Experimental Psychology: Human Learning and Memory 107 (1980).
65. Adam L. Alter, Daniel M. Oppenheimer, Nicholas Epley, and Rebecca N. Eyre,
Overcoming Intuition: Metacognitive Difficulty Activates Analytic Reasoning, Journal of
Experimental Psychology, General 136.4 569–76 (2007).
294 problem solving, decision making, and professional judgment
parole violations, loan defaults and bankruptcies, and the academic success of
graduate students.66 For example, while statistical prediction of parole failures
based on three variables (type of offense, number of past convictions, and
number of violations of prison rules) produced a (very modest) 0.22 correlation
with actual parole failures, this far surpassed professional parole interviewers,
whose predictions had a correlation of only 0.06.67 In the Cook County Hospital
emergency room, the use of four objective factors did a far better job than clinical
diagnosis in predicting whether a patient with chest pain was in danger of suf-
fering a heart attack.68
Baseball offers an interesting example of the difference between clinical and
statistical decision making.69 On the clinical side, the scout, uses his intuition to
recruit players based on speed, quickness, arm strength, hitting ability, and
mental toughness, and qualities such as “a live, active lower body, quick feet,
agility, instinct, . . . [and] alertness.” To the extent statistics play a role in evaluat-
ing a hitter’s potential contribution to a team’s success, they traditionally have
centered around batting average, runs batted in, and home runs.
Bill James, who published the Baseball Abstract, found that these few statistics
did not account for a player’s contribution to the team’s winning. He began to
create new statistics that were the product of a sophisticated analysis of hun-
dreds of games, with the ultimate measure of a hitter’s success based on how
many runs he creates.
Billy Beane, general manager of the Oakland Athletics, began to experiment
with James’s approach. He discovered that on-base percentage (the percentage
of the time a batter reached base by any means) and slugging percentage (total
bases divided by at-bats) were the two best predictors of success, and on-base
percentage was three times as important as slugging percentage.
Beane thus selected players based upon an entirely different model than his
clinically-minded competitors. Just as clinicians generally resist statistical pre-
diction, the scouts were reluctant to yield their intuitions. In Moneyball, Michael
Lewis recounts a paradigmatic conversation between Beane and a recruiter:
“The guy’s an athlete, Billy,” the old scout says. “There’s a lot of upside
there.”
66. Robyn M. Dawes, David Faust, and Paul E. Meehl, Clinical Versus Actuarial
Judgment, in Heuristics and Biases; Robyn M. Dawes, House of Cards: Psychology
and Psychotherapy Built on Myth (New York: The Free Press, 2002).
67. See Robyn Dawes, House of Cards, supra.
68. The studies are summarized in Malcolm Gladwell, Blink 125–36 (2005).
69. See Michael Lewis, Moneyball: The Art of Winning an Unfair Game (New York:
Norton & Company, 2003); Cass Sunstein and Richard Thaler, Who’s on First, New
Republic (August 2003); Ehren Wassermann, Daniel R. Czech, Matthew, J. Wilson, and
A. Barry Joyner, An Examination of the Moneyball Theory: A Baseball Statistical Analysis
Abstract, 8 The Sports Journal (Winter 2005), http://www.thesportjournal.org/
2005Journal/Vol8-No1/daniel_czech.asp.
296 problem solving, decision making, and professional judgment
70. Another proponent of Beane’s method of evaluating talent, Boston Red Sox
General Manager Theo Epstein, saw his team win the World Series for the first time in
eighty-eight years, albeit with a significantly higher payroll than Beane’s As. It should be
noted that the claim that statistics are more valuable than informed intuition in recruiting
and deploying a baseball team has met with some vehement resistance. David Leonhardt,
To Play Is the Thing, New York Times, Aug. 28, 2005, reviewing Buzz Bissinger, Three
Nights in August (2005), and Bill Shanks, Scout’s Honor (2005).
71. See Robyn M. Dawes and Bernhard Corrigan, Linear Models in Decision Making, 81
Psychology Bulletin 95 (1974). It is questionable whether one can improve on a simple
linear model through the process of bootstrapping—of developing and implementing a
regression equation based on observing an expert decision maker. Id.
72. William M. Grove and Paul E. Meehl, Comparative Efficiency of Informal (Subjective,
Impressionistic) and Formal (Mechanical, Algorithmic) Prediction Procedures—The Clinical
Statistical Controversy, 2 Psychology, Public Policy, and Law 293 (1996).
biases in processing and judging information 297
73. Paul J. Hoffman, Paul Slovic, and Leonard G. Rorer, An Analysis-of-Variance Model
for Assessment of Configural Cue Utilization in Clinical Judgment, 69 Psychological
Bulletin 338 (1968).
74. Lewis R. Goldberg, Man Versus the Model of Man: A Rationale, Plus Some Evidence,
for a Method of Improving on Clinical Inferences, 73 Psychology Bulletin 422 (1970).
298 problem solving, decision making, and professional judgment
Second, the general superiority of statistical over clinical prediction raises the
question—of ego as well as fact—of what use are experts? We now turn to this
question.
75. Gary Klein, Sources of Power: How People Make Decisions 32 (Cambridge,
MA: MIT Press, 1998).
biases in processing and judging information 299
76. Benedict Carey, In Battle, Hunches Prove to be Valuable, New York Times, July 28,
2009, http://www.nytimes.com/2009/07/28/health/research/28brain.html?pagewanted=
2&_r=1&hp.
300 problem solving, decision making, and professional judgment
• Validity. Expert judgments in many fields are often just wrong. (Compared
to what, though? Expert judgments are usually more accurate than those of
laypersons.)
• Reliability. Given exactly the same fact situation an hour or a week later, an
expert will often come to a different judgment. A review of studies of reli-
ability found an overall correlation of 0.79 between the first judgment and
a subsequent one—though meteorologists scored 0.91.77
• Calibration. Experts’ certainty about their judgments tends to be poorly
calibrated with their validity, with meteorologists again being an excep-
tion.78 (On the other hand, there is some evidence that laypersons are
attracted to overconfident experts.79)
• Under-use of available relevant information. While one of the advantages of
expertise is the ability to exclude extraneous information, experts often do
not seek or use information that is relevant and diagnostic. One possible
explanation, related to schematic processing, is that an expert’s “feeling of
knowing”—akin to the “strong sense of general familiarity” that can mis-
lead eyewitness identification80—may demotivate further inquiry. When
an expert “must choose either to retrieve a previously computed solution or
to compute the solution anew, she may judge the likelihood that she pos-
sesses the correct answer in memory by assessing her feeling of knowing
it.”81 While the expert may often be correct, overconfidence in one’s initial
judgment can readily lead to short-circuiting any further learning.
• Combining indicators. As we saw in the preceding section, experts’ intu-
itions are not as reliable or consistent as simple linear models in combin-
ing or aggregating indicators to predict outcomes. Nonetheless, experts are
an important source for determining what indicators are relevant and for
discarding supposed indicators that are not. But more fundamentally,
many situations that experts deal with—situations like the kitchen fire—do
not have the similar, repetitive quality that allow for the application of a
linear model.82
77. Robert H. Ashton, A Review and Analysis of Research on the Test-Retest Reliability of
Professional Judgment, 13 Journal of Behavioral Decision Making 277 (2000).
78. Stuart Oskamp, Overconfidence in Case Study Judgments, 29 Journal of Consulting
Psychology 261–65 (1965).
79. Paul C. Price and Eric R. Stone, Intuitive Evaluation of Likelihood Judgment Producers:
Evidence of a Confidence Heuristic, 17 Journal of Behavioral Decision Making 39 (2004).
80. Daniel L. Schacter, The Seven Sins of Memory: How the Mind Forgets and
Remembers (New York: Houghton Mifflin, 2001).
81. Stacy L. Wood and John G. Lynch, Jr., Prior Knowledge and Complacency in New
Product Learning, 29 Journal of Consumer Research 416 (2002).
82. On the other hand, James Shanteau suggests that expert judgments are most reli-
able when dealing with relatively repetitive tasks with static stimuli. James Shanteau,
biases in processing and judging information 301
10.6 conclusion
Nisbett and Ross report that a colleague, after reading a draft of their manu-
script of Human Inference, asked: “If we’re so dumb, how come we made it
to the moon?”87 Their response is that advances in the physical and natures
sciences and technology are ultimately due to formal research methodology and
principles of inference, and are the result of collective rather than individual
learning. For example, “the individual is not usually required to detect covaria-
tion anew,” but can learn from experts or through cultural transmission:
Each culture has experts, people of unusual acumen or specialized knowl-
edge, who detect covariations and report them to the culture at large. Thus,
most (though not by any means all) cultures recognize the covariation between
intercourse and pregnancy. This is a covariation detection task of enormous
complexity, given the interval between the two types of events and the rarity
of pregnancy relative to the frequency of intercourse, not to mention the occa-
sional unreliability of the data (“Honest, I never did it!”). . . . [O]nce a covaria-
tion is detected and the new hypothesis seems to be confirmed by perusing
available data, the entire culture is the beneficiary and may take action in
accordance with the knowledge. Such a cultural transmission principle
applies to almost all human affairs from farming (“Plant corn when the oak
leaf is as big as a mouse’s ear”) to tourism (“The German restaurants in
Minnesota are generally quite good”) to urban survival (“The South Side is
unsafe.”)”88
Although cultures can embrace and transmit false knowledge, adaptive
pressures tend to correct misinformation that has practical consequences.
11.1 introduction
Many of the decisions that lawyers, judges, and policy professionals make involve
interpreting and predicting other people’s behavior. Indeed, making sense of
other people figures centrally in the performance of virtually every aspect of law-
yers’, judges’, and public officials’ work. Lawyers and other “forensic profession-
als” functioning as investigators, or judges acting in their roles as fact-finders,
for example, have to assess credibility and attribute motivation. As counselors,
lawyers must understand what makes their clients “tick,” win their clients’ trust,
and predict how their clients will feel and behave in the future. In evaluating
specific cases and designing advocacy strategies, lawyers have to predict how
legal decision makers, like judges, administrative officials, or jurors, will respond
to specific legal or factual arguments. What arguments will they find persuasive?
What narratives will they find compelling? What inferences will they draw from
ambiguous evidence susceptible of varying interpretations? In negotiating an
agreement or designing a process, lawyers have to structure incentives that will
encourage certain behaviors and discourage others, and this requires under-
standing a great deal about human behavior. Policy makers have analogous tasks
in dealing with constituents, legislatures, and other decision-making bodies and
in regulating or otherwise influencing the behavior of individuals and entities.
Errors of judgment in any of these, and many similar contexts, can lead to bad
decisions and ineffective policies. For this reason, understanding the sources of
bias and error that can distort social perception and judgment, and knowing how
to structure decision making so as to minimize their negative effects, will
improve the quality of a lawyer or policy maker’s performance in virtually every
aspect of his or her professional role.
Social judgment, like judgment more generally, is influenced by systematic
biases and other sources of distortion similar to those described in earlier chap-
ters. Biases in social judgment, however, are in some ways the most insidious of
all. Because social interaction constitutes so much of the grist of daily profes-
sional (and personal) life, we tend to think of ourselves as experts in understand-
ing other people and predicting their future behavior. Our self-confidence as
“common-sense” or “intuitive” psychologists often leads us to think that we
understand other people better than we do.
In this chapter, we consider how people go about making sense of other
people. In the process, we introduce readers to various sources of bias and distor-
tion in social perception and judgment, describe how “intuitive” psychological
304 problem solving, decision making, and professional judgment
theories can lead us astray in our attempts to understand and predict people’s
behavior, and make some suggestions as to what might be done about it.
The psychological processes we explore in this chapter, particularly those
having to do with racial, gender, or other forms of stereotyping and bias, raise a
number of highly controversial legal and policy issues. You have probably already
been exposed to at least parts of the large and at times contentious literature
engaging many of the controversies, whether you are studying law, public policy,
or some other discipline. It is not our purpose in this chapter to rehearse, to
weigh in on, or even to summarize, these controversies. Rather, we seek only to
describe a number of the major—and for the most part, consensus—findings in
empirical social and cognitive psychology about various psychological processes
that can conduce to intergroup misperception and perpetuate subtle—often
unintentional—forms of intergroup bias.
affects how those targets behave toward us. If, for example, you apply for a job
and are interviewed by someone who, because of your race, gender, or grade
point average, expects you to be unqualified for the job, their attitude may leak
out in subtle aspects of their behavior, such as seating distance, eye contact, and
facial expressions, all of which may give you a vague sense that they don’t like
you. You may in turn behave in ways that project social discomfort, reticence, or
self-doubt, confirming the interviewer’s prior impression of you. We examine
these kinds of expectancy feedback effects later in this chapter, in sections on the
behavioral confirmation effect and a phenomenon known as “stereotype threat,”
through which a stereotyped person’s fear of performing in a way that will con-
firm a negative stereotype about his group actually does impair his performance
on a stereotype-relevant task.
For clarity’s sake, much of the following discussion of the biasing effects of
social expectancies centers on racial and gender stereotypes and stereotyping.
Readers should bear in mind, however, that social expectancies derive from
other knowledge structures as well, including individual person prototypes and
role prototypes. Where possible, we will provide examples and illustrations
incorporating these as well.
regularly used to classify people and then to construe, encode, recall, and predict
their behavior, motivations, and character attributes.
Stereotypes often function implicitly, meaning that they can influence social
perception and judgment outside of conscious awareness, whether or not they
are believed. As social psychologist Patricia Devine has shown, even if a person
has consciously rejected a particular stereotyped belief (e.g., that females, as a
group, have less innate scientific aptitude than males), the stereotype may
linger on as an explicit expectancy, capable of influencing a perceiver’s subjec-
tive assessment of a particular male or female target’s scientific aptitude or
ability.10
Stereotypes are spontaneously activated when the stereotype-holder encoun-
ters a member of the stereotyped group. This activation is automatic, and occurs
without intention or awareness, with little cost to other, ongoing mental opera-
tions. In this way, stereotypes, and other social schemas, function in much the
same way as the availability, representativeness, and other heuristics described
in earlier chapters. Although they can serve as useful “rules of thumb,” they can
also lead to poor judgment, not to mention violations of law and important social
and personal ideals.
Law professor Jerry Kang illustrates the process by which stereotypes influ-
ence social perception, judgment, and behavior as shown in Figure 11.1.
Priming effects/
mapping rules
mapped by the
perceiver through into
Person
Perceived social
categories
meanings
10. Patricia G. Devine, Stereotypes and Prejudice: Their Automatic and Controlled
Components, supra n. 8.
308 problem solving, decision making, and professional judgment
example, a human being wearing a dress, having long, styled hair, and wearing
makeup will usually be mapped as female. However, category assignment may
be influenced not only by culturally provided mapping rules, but also by features
of the stimulus situation, which may “prime” particular categories, increasing
the likelihood that they will be used in categorizing the person perceived. So, in
a drag bar, the above-mentioned human might not be spontaneously mapped as
female. In either case, however, assignment of an individual to a particular cat-
egory activates the category. The category, which comprises a network of mental
associations, supplies the perceiver with a set of implicit and explicit meanings.
These meanings, in turn, systematically influence the perceiver’s judgments of
and interaction with the categorized individual.
Kang’s model can be thought of as a visual representation of the schematic
information processing model of “naturalistic,” or “System 1,” thinking described
in Section 1.6.1. Indeed, stereotyping is best understood not as something that
only “prejudiced” people do, but as one type of System 1 thinking, which is auto-
matic, rapid, associative, effortless, and process opaque.11 Of course, this does
not mean that the application of stereotypes in social judgment is inescapable.
As with other types of System 1 thinking, an initial stereotype-driven impression
can be revised through the application of deliberative, or System 2 thinking,
which is controlled, effortful, and self-conscious.10 We describe this correction
process in Section 11.7.
Liberal Political
Church
Ideals Student
Black Church
Exposure to Culture(s)
(Art, Music, Literature)
Choir Gospel
Young black male
Church Youth
Urban black
R&B
youth
Gangster Rap
Violence
Hip
Hop
Mugger
BlingBling
Crime “Sagging”
Mugging
Script
Oversized Clothing
Ski Cap
Gun
Poverty
Gang Member
Visual Exemplar
parents, teachers, and friends, the media, or one’s own experience and imagina-
tion. Once activated, the nodes of the network mediate subsequent social percep-
tion and judgment. Like any schema, the components of the network contain
affective components (e.g., fright, pride, warmth),17 action “scripts” (e.g., walk
faster, say hello, smile), and causal theories explaining why stereotyped group
members “are the way they are.”15 This is one way of understanding what it
means to say that race is a “social construction.” The social environment “con-
structs” the associative network in the brain that constitutes racial categories.
Notice in Figure 11.2 the wide variety of concepts associated with the subcat-
egory, “young Black male.” Many of these concepts, like “crime,” “violence,”
“poverty,” “gun,” “gang member,” and “mugging script,” which are intercon-
nected at the bottom of Figure 11.2, carry negative connotations. Taken together,
they construct the schema “criminal urban Black male.” But notice also that other
concepts, such as “Black church,” “college student,” “church music,” “church
youth,” “liberal political ideals,” “exposure to culture,” “self as young man,” and
“jazz,” carry positive connotations. Grouped together, these concepts would
constitute a quite different subcategory, perhaps “Black male college student.”
Given all the different constructs that an encounter with William Carter
might activate, how is it that, on any particular occasion, he is subcategorized
and then construed—or misconstrued—in one way rather than another? Because
in American culture young Black men are so frequently associated in media and
other representations with urban violence, and because the dark, deserted New
York street might well have primed the concept of crime in Jackson’s mind,
Jackson would have been predisposed, upon first seeing Carter, to categorize
him as he did. Additionally, Carter’s physical attributes, dress, and posture,
although inherently ambiguous, would have reinforced this initial categoriza-
tion, because of their association with youth gangster culture. The net effect is
the activation of those elements of the Figure 11.2 associative network bolded
and shaded in Figure 11.3.
What might have happened if Carter had been walking down the street sing-
ing the tenor melody from the Gloria section of the B Minor Mass? This new
stimulus feature might have dramatically altered how Jackson classified the
approaching Carter. The schematic network activated by the encounter might
have looked more like Figure 11.4 than like Figure 11.3.
Presented with a William Carter singing the Gloria from Bach’s B Minor
Mass, concepts like “classical music,” “sacred choral music,” “church,” “God,”
and “exposure to culture” would have been activated. These in turn would have
activated other concepts, like “church youth,” “choir,” and “student,” and from
these, perhaps specific exemplars of young Black men from Jackson’s own
17. Russell H. Fazio, Joni R. Jackson, Bridget C. Dunton, and Carol J. Williams,
Variability in Automatic Activation as an Unobtrusive Measure of Racial Attitudes: A Bona
Fide Pipeline?, 69 Journal of Personality and Social Psychology 1013–27 (1995).
312 problem solving, decision making, and professional judgment
Civil Disobedience
SNCC SCLC
Satyagraha
Civil Rights
Activist
Law Abiding
Liberal Political
Church Ideals Student
Black Church
Exposure to Culture(s)
(Art, Music, Literature)
Choir Gospel
Young black male
Church Youth
Urban young
black male R&B
Gangster Rap
Violence
Hip
Hop
Mugger
BlingBling
Crime “Sagging”
Mugging
Script
Oversized Clothing
Ski Cap
Gun
Poverty
Gang Member
family, religious community, or even of himself as a young man. The end result
would have been the instantiation of a particular subtype of “Black young man”
that incorporated an interconnected network of positive associations and emo-
tional reactions very different from those constituting the stereotype of young
Black urban male criminality depicted in Figure 11.4.
Indeed, recent research suggests that the activation of certain elements of a
schematic network can result not only in the activation of elements with which
they are connected, but also in the deactivation of elements with which they are
disassociated.18 Seen in this way, the various constructs comprising our mental
representations can stand in three types of relation to each other. They can be
associated, in which case the activation of one will increase the probability that
the other will be activated. They can be completely disassociated, in which case
the activation of one will have no impact on the probability of the others’ activa-
tion. Or, they can be negatively associated, or mutually inhibitory, such that the
activation of one actually decreases the probability that the other will be
invoked.
For example, many female executives and professionals report that, as soon
as they have their first child, people at work start responding to them differently.
They are given less responsible job assignments; their suggestions are taken less
seriously; they are accorded less deference and respect. One explanation of this
reported phenomenon is that it is very difficult for many people to associate the
social role construct “mother” with the trait constructs “aggressive” or “hard-
driving.” The activation of “mother” not only activates a set of trait constructs,
such as “nurturing,” or “soft,” it actually inhibits the activation of others, like
“intellectually sharp” or “professionally productive.”19
Just as it is hard for many people to think of another person simultaneously
as a “new mother” and a “hard-driving executive,” it is hard to think simultane-
ously about sacred choral musicianship and a propensity toward street violence.
By singing Bach’s B Minor Mass, William Carter has not only behaved in a way
that will activate certain constructs in Jesse Jackson’s mind, he has behaved in a
way that will inhibit the activation of others. The result is a profound change in
what Jackson “sees”—at least as profound as what Dartmouth and Princeton
students saw in the football game recounted in Section 9.4.1.
We have presented two alternative scenarios of the external stimuli in
Jackson’s encounter with Carter. In our hypothetical case, whistling Bach made
all the difference in terms of which schematic elements were activated. But other
factors can determine which aspects of an incoming perception most influence
18. Paul Thagard, and Zeva Kunda, Making Sense of People: Coherence Mechanisms, in
Connectionist Models of Social Reasoning and Social Behavior 3–26 (Stephen J.
Read and Lynn C. Miller eds., Mahwah, N.J.: Erlbaum,1998).
19. Amy. J. C. Cuddy, Susan T. Fiske, and Peter Glick, When Professionals Become
Mothers, Warmth Doesn’t Cut the Ice, 60 Journal of Social Issues 701 (2004).
314 problem solving, decision making, and professional judgment
Civil Disobedience
SNCC SCLC
Satyagraha
Civil Rights
Emotional Activist
Warmth Law Abiding
Liberal Political
Church Ideals Student
Black
Exposure to
Church
Culture(s) (Art,
Choir Gospel Music, Literature)
Young black male
Church Youth
Urban black
youth R&B
Gangster Rap
Violence
Hip
Hop
Mugger
BlingBling
Crime “Sagging”
Mugging
Script
Oversized Clothing
Ski Cap
Gun
Poverty
Gang Member
schema activation. Some kinds of stimuli are more salient than others. All other
things being equal: visual stimuli,20 negative stimuli,21 and stimuli that are
unusual or unexpected (like a young person walking down a Brooklyn street
singing a portion of Bach’s B Minor Mass) are particularly salient.
Also, the more frequently or recently a particular schema has been activated,
the more available it is, and thus the more likely it is to be reactivated by a stimu-
lus that relates to it.22 Thus, given the prominence of news stories about crimes
committed by young Black men, seeing a young Black man on an empty Brooklyn
street late at night is likely to activate the negative elements of a schematic net-
work together with a “mugging” script and a fear/flight response.23
Priming also exerts a powerful impact on which schemas are activated by a
particular stimulus. Priming refers to the notion that people’s judgment, atti-
tudes, and behavior can be influenced by incidental cues in their surroundings.
Exposure to subtle situational cues, such as words, people, and even scent or
physical objects, can automatically activate knowledge structures in memory
such as trait concepts, stereotypes, and goals. Such automatic activation can
occur without people’s intention or awareness and carry over for at least a while
to exert an unintended influence on judgment and behavior.
For example, John Bargh and his colleagues primed people with words related
to the elderly (e.g., Florida, bingo) in what was said to be a sentence unscram-
bling task.24 After completing the task, when participants believed the session
had ended, the researchers measured how long it took participants to walk to the
elevator on their way out of the building. Surprisingly, participants primed with
words related to the elderly walked significantly more slowly than other partici-
pants who were primed with neutral words (e.g., normal, send). In another
study, Bargh and his colleagues found that people primed with words related to
rudeness (e.g., disturb, impolitely) were subsequently quicker to interrupt an
20. Susan T. Fiske and Martha G. Cox, Person Concepts: The Effect of Target Familiarity
and Descriptive Purpose on the Process of Describing Others, 47 Journal of Personality
136–61 (1979).
21. Felicia Pratto, and John P. Oliver, Automatic vigilance: The Attention-Grabbing Power
of Negative Social Information, Journal of Personality and Social Psychology (1991).
22. Felicia Pratto, and John A. Bargh, Stereotyping Based on Apparently Individuating
Information: Trait and Global Components of Sex Stereotypes Under Attention Overload, 27
Journal of Experimental Social Psychology 26 (1991); E. Tory Higgins, and Gilliam A.
King, Accessibility of Social Constructs: Information-Processing Consequences of Individual and
Contextual Variability (1981), in Personality, Cognition, and Social Interaction 69–121
(Nancy Cantor and John F. Kihlstrom eds., Hillsdale, NJ: Erlbaum, 1986); Robert S. Wyer and
Thomas K. Srull, Human Cognition in its Social Context, Psychological Review, 93, 322–59.
23. Jerry Kang, Trojan Horses of Race, 118 Harv. L. Rev. 1489–1593 (2005).
24. John A. Bargh, Mark Chen, and Lara Burrows, Automaticity of Social Behavior:
Direct Effects of Trait Construct and Stereotype Activation on Action, 71(2) Journal of
Personality and Social Psychology 230–44 (1996).
316 problem solving, decision making, and professional judgment
ongoing conversation between the experimenter and a third person than people
primed with neutral words. These findings imply that the mental circuits used to
perceive social stimuli, such as words, are closely linked to behavioral schemata
and other mental representations that affect people’s judgments and behavior.
Priming effects not only have consequences for simple, behavioral responses
such as speed of walking, but can also affect more complex behaviors and goals.25
One group of White participants was asked to write about a day in the life of
Tyrone Walker, who most people assumed to be an African American. Other White
participants were asked to write about a day in the life of Eric Walker, who they
assumed to be White. Next, all participants were asked to complete a math section
of the Graduate Record Examination (GRE). Participants who wrote about Tyrone
Walker performed poorly compared to those who wrote about Eric Walker.
As this example illustrates, racial stereotype activation can elicit maladaptive,
stereotype-consistent behavior among individuals who hold a stereotype. In other
words, because the stereotype of African Americans includes an association with
low academic achievement, activating the African American stereotype noncon-
sciously caused people to behave in a way consistent with that stereotype.
Priming can take the form of entirely unrelated stimuli occurring before the
event—for example, seeing a film (not necessarily about race) that leaves the
viewer with feelings of brotherly love on the one hand, or dread on the other.
One can also be primed by one’s own thoughts. For example, walking down a
deserted Brooklyn street late at night might give rise to scary images or feelings.
Such thoughts, even if subliminal and fleeting can promote, or prime, a negative
construal of inherently incoming perception.26
Finally, schema activation can be influenced by the perceiver’s interaction
goals. Much of our interaction with other people is motivated toward some end.
When you first meet a new coworker upon whom you must depend for the suc-
cess of your own efforts, you will probably be motivated to like the person and
feel confident in his or her abilities. To feel otherwise would prompt distress.
Likewise, if you need surgery, you will be motivated to perceive the surgeon who
will be operating on you as competent. People who are motivated to avoid preju-
dice tend to monitor their own reactions to members of other groups, attempt-
ing to notice and correct for the effects of stereotypes or other forms of bias.
We discuss issues relating to controlling or correcting for the biasing effects of
stereotype activation later in Section 11.7.
25. Based on S. Christian Wheeler, Jarvis W. Blair, and Richard E. Petty, Think
Unto Others: The Self-Destructive Impact of Negative Racial Stereotypes, 37 Journal of
Experimental Social Psychology 173–80 (2001).
26. Jerome Bruner, On Perceptual Readiness, 64 Psychological Review 123 (1957);
Jerome Bruner, Going Beyond the Information Given, in Contemporary Approaches to
Cognition, Jerome Bruner, ed., (Cambridge, MA: Harvard University Press, 1957).
the social perceiver: processes and problems in social cognition 317
27. Andrew H. Sagar, and Janet W. Schofield, Racial and Behavioral Cues in Black and
White Children’s Perceptions of Ambiguously Aggressive Acts, 39 Journal of Personality
and Social Psychology 590–98 (1980).
318 problem solving, decision making, and professional judgment
28. Birt L. Duncan, Differential Social Perception and Attribution of Intergroup Violence:
Testing the Lower Limits of Stereotyping of Blacks, 34 J. Personality & Soc. Psychol. 590–98
(1976).
the social perceiver: processes and problems in social cognition 319
Not all participants were asked to predict Hannah’s achievement level right
away, however. Some watched a second videotape before they were asked to
predict Hannah’s academic performance. This video, which was identical across
both conditions, depicted Hannah responding verbally to twenty-five achieve-
ment test problems. Researchers told participants that the test included easy,
moderate, and difficult problems. Hannah’s performance was designed to be
ambiguous. After watching the second video, participants who had seen the high
socioeconomic status (SES) Hannah rated her academic ability significantly
higher than did participants who saw her low SES double. Open-ended com-
ments revealed differences in perception as well. For example, participants
reported the low-income Hannah as having “difficulty accepting new informa-
tion,” while they frequently described high-income Hannah as demonstrating
the “ability to apply what she knows to unfamiliar problems.”
Darley and Gross suggested that exposure to ambiguous diagnostic informa-
tion actually promotes social stereotyping.29 People who do not believe it appropri-
ate to base social predictions on stereotypes based on socioeconomic information
alone, nonetheless did so in what they perceived to be a more diagnostic con-
text—even though it was thoroughly ambiguous. From this, Darley and Gross
argued that stereotypes function not so much as conscious ex ante decision rules,
but as dormant expectancies that impose a hypothesis-confirming bias on the
interpretation of incoming social information.
The effects of stereotypes and other schematic expectancies on social percep-
tion and judgment have been well documented across a variety of different
contexts.30 Perhaps most disturbing are recent studies showing that that a
target person’s race has a significant effect on whether both Black and White
participants think he is holding a weapon, like a knife or a gun, as opposed to
a harmless object, like a wallet or cell phone.31
Although most of the research on the biasing effects of social schemas has
focused on race, age, and gender stereotypes, the basic idea that schemas shape
how we interpret other people’s behavior applies in a variety of other contexts
29. John M. Darley and Paget H. Gross, A Hypothesis-Confirming Bias in Labeling Effects,
44 J. Personality & Soc. Psychol. 20–33 (1983).
30. David Dunning and David A. Sherman, Stereotypes and Tacit Inference, 73
J. Personality & Soc. Psychol. 459–71 (1997); Zeva Kunda and Bonnie Sherman-
Williams, Stereotypes and the Construal of Individuating Information, 19 J. Personality &
Soc. Psychol. 90–99 (1993).
31. Joshua Correll, Geoffrey R. Urland, and Tiffany A. Ito, Event-Related Potentials and
the Decision to Shoot: The Role of Threat Perception and Cognitive Control, 42 Journal of
Experimental Social Psychology 120–28 (2006); Joshua Correll, Bernadette Park, Charles
M. Judd, and Bernd Wittenbrink, The Police Officer’s Dilemma: Using Ethnicity to Disambiguate
Potentially Threatening Individuals, 83 J. Personality & Soc. Psych. 1314 (2002); B. Keith.
Payne, Prejudice and Perception: The Role of Automatic and Controlled Processes in Misperceiving
a Weapon, 81 J. Personality & Soc. Psychol . 181–192, 185–186 (2001).
320 problem solving, decision making, and professional judgment
32. Robert. J. Robinson and Dacher Keltner, Defending the Status Quo: Power and Bias
in Social Conflict, 23 Personality and Social Psychology Bulletin 1066–77 (1997);
Robert J. Robinson and Raymond A. Friedman, Mistrust and Misconstrual in Union-
Management Relationships: Causal Accounts in Adversarial Contexts, 6 International
Journal of Conflict Management 312–27 (1995); Roderick M. Kramer, The Sinister
Attribution Error: Paranoid Cognition and Collective Distrust in Organization, 18 Motivation
and Emotion 199–230 (1994); Dacher Keltner and Robert J. Robinson, Imagined Ideological
Differences in Conflict Escalation and Resolution, 4 International Journal of Conflict
Management 249–62 (1993).
33. Leslie Z. McArthur, What Grabs You?: The Role of Attention in Impression
Formation and Causal Attribution in Social Cognition (1981); Shelley E. Taylor
and Susan T. Fiske, Salience, Attention, and Attribution: Top of the Head Phenomena, in 11
Advances in Experimental Social Psychology 249, 264–65, Leonard Berkowitz ed.
(New York: Academic Press, 1978).
the social perceiver: processes and problems in social cognition 321
experimenters projected a slide of that person’s picture. By playing the same tape
recording, while varying the race of the person supposedly speaking, Taylor and
her colleagues compared participants’ judgments of a Black person when he was
the only Black person in an otherwise all White group, and when he was in a fully
integrated group. In the “solo” condition, participants judged the Black person in
more extreme ways and perceived him as playing a more prominent role in the
discussion than Blacks were perceived in the “integrated” condition.
In a second experiment, Taylor and her colleagues found similar, even stron-
ger effects on perceptions of solos in gender-mixed groups. Token women and
minorities, in short, are rated more extremely and are viewed (for better or for
worse) as playing a more prominent role in group process, than are women or
minorities in integrated groups.
34. David W. Jamieson and Mark P. Zanna, Need for Structure in Attitude Formation
and Expression, in Attitude Structure and Function 383–406 (Anthony R. Pratkanis,
S. J. Breckler, and Anthony G. Greenwald eds., Hillsdale, NJ: Erlbaum, 1989).
35. Joshua Klayman and Young-Won Ha, Confirmation, Disconfirmation, and
Information in Hypothesis Testing, 94 Psychological Review 211–28 (1987).
the social perceiver: processes and problems in social cognition 323
36. Mark Snyder, and Nancy Cantor, Hypothesis Testing in Social Interaction, 36
J. Personality & Social Psychol. 1202–12 (1979).
37. Jennifer Crocker, Darlene B. Hanna, and Renee Weber, Person Memory and Causal
Attributions, 44 J. Personality & Social Psychol. 55 (1983).
324 problem solving, decision making, and professional judgment
As we have seen, people tend to use social knowledge structures to fill in gaps in
the information they have about another person and to predict how that person
38. Galen Bodenhausen and Robert Wyer, Effects of Stereotypes in Decision Making and
Information-Processing Strategies, 48 Journal of Personality and Social Psychology
267–82 (1985).
39. George S. Bridges and Sara Steen, Racial Disparities in Official Assessments of
Juvenile Offenders: Attributional Stereotypes as Mediating Mechanisms, 63 Am. Soc. Rev. 554
(1998).
the social perceiver: processes and problems in social cognition 325
will behave in the future.40 Once a perceiver has formed expectancies about a
target person, those expectancies can play a significant role in shaping the per-
ceiver’s behavior toward the target in subsequent interactions. This may, in turn,
affect the target’s behavior toward the perceiver, often in ways that confirm the
perceiver’s initial expectancy. This behavioral confirmation effect represents one
way in which stereotypes can function as self-fulfilling prophesies. Stereotype threat
represents a second process by which stereotypes become self-fulfilling. In stereo-
type threat situations, a person’s awareness that his or her group is negatively
stereotyped impairs his or her performance in a stereotype-relevant domain.
40. James M. Olson, Neal J. Roese and Mark P. Zanna, Expectancies, in Social
Psychology: Handbook of Basic Principles 211–38, E. Tory Higgens and Arie
Kruglanski eds., (New York: Guilford, 1996).
41. Robert Rosenthal, and Lenore Jacobson, Pygmalion in the Classroom
(New York: Holt, Rinehart, & Winston, 1968).
326 problem solving, decision making, and professional judgment
1. Tentative expectancy:
(I’m told he’s very bright.)
2. Ambiguous behavior:
(Could be seen as bright.)
3. Expectancy strengthened
by perceptual confirmation:
(He does seem bright!)
4. Expectancy-influenced
behavior (i.e., eye contact,
body posture, facial
expressions, patterns of
verbal interaction (the
expectancy-behavior link)
5. Lively, energized response
(behavior-behavior link)
6. Expectancy further
strengthened by behavioral
confirmation: (I was right.
He is bright!)
46. The amygdala is a small, bilateral structure located in the brain’s temporal lobe. It
plays a significant role in emotional learning and evaluation, and particularly in the evoca-
tion of emotional responses to fear or aggression-inducing stimuli. Ralph Adolphs, Daniel
Tranel, and Antonio R. Damasio, The Human Amygdala in Social Judgment, 393 Nature
470 (1998).
47. Elizabeth A. Phelps, Kevin J. O’Connor, William A. Cunningham, E. Sumie
Funayama, J. Christopher Gatenby, John C. Gore, and Mahzarin R. Banaji, Performance
on Indirect Measures of Race Evaluation Predicts Amygdala Activation, 12 J. Cognitive
Neuroscience 729-738, 730 (2002).
328 problem solving, decision making, and professional judgment
Members of negatively stereotyped groups often claim that they can “just tell”
when someone they are dealing with is biased against them. Given that people
are exquisitely attuned to subtle facial, verbal, and other physical cues, and use
these to decode others’ attitudes toward them, it is certainly reasonable to believe
that that subtle behavioral “leakages” from implicitly biased perceivers can be
sensed by members of stereotyped groups.
Just as perceivers’ behavior toward stereotyped targets is affected by the
expectancies that stereotypes activate, stereotyped targets respond in kind, thus
creating the behavior-behavior link described as step 5 of Jones’s model. The
earliest evidence for the behavior-behavior link comes from Rosenthal
and Jacobson’s 1968 study, described earlier, in which students responded to
their teacher’s arbitrarily created expectancies by academically outperforming
their peers. Substantial additional evidence for the link has accumulated since
then.
For example, in the McConnell and Leibold study described above, White
confederates were trained to mimic the “friendly” and “unfriendly” interview
styles that had been observed in Word, Zanna, and Cooper’s 1974 experiment. In
the “friendly” condition, trained interviewers sat closer to interviewees, made
fewer speech errors, and extended the length of the interview. Blind raters (i.e.,
people who knew nothing about the nature of the experiment) coded the inter-
viewees’ behavior, and found that those interviewees who had been interviewed
using the “unfriendly” style performed more poorly in the interviews than White
interviewees who had been treated in the “friendly” style. Similar effects were
observed in experiments manipulating perceivers’ beliefs that they were speak-
ing with either an attractive or unattractive partner48 or to an obese or normal
weight partner. More recently, Serena Chen and John Bargh demonstrated that
African American targets who interacted with White participants who had been
subliminally primed with pictures of African American faces were rated by con-
dition-blind coders as behaving with more verbal hostility than African American
targets who had interacted with Whites primed with Caucasian faces. The tar-
gets themselves had not been primed, and yet their behavior conformed to the
stereotypic expectancies induced in their partners.49
In short, there is ample evidence, far more than can be reviewed here,
suggesting that when subtle expectancy-consistent behaviors “leak out” of
stereotype-holding perceivers, targets respond in kind, often contributing to
stereotype confirmation. While these effects can be moderated by such factors as
48. Allen. R. McConnell and Jill M. Leibold, Relations Among the Implicit Association
Test, Discriminatory Behavior, and Explicit Measures of Racial Attitudes, supra note 45.
49. Mark Chen, and John A. Bargh, Nonconscious Behavioral Confirmation Processes:
The Self-Fulfilling Consequences of Automatic Stereotype Activation, 33 Journal of
Experimental Social Psychology 541–60 (1997).
the social perceiver: processes and problems in social cognition 329
50. Steven L. Neuberg, Nicole T. Judice, Lynn M. Virdin, and Mary A. Carrillo, Perceiver
Self-Presentation Goals as Moderators of Expectancy Influences: Ingratiation and the
Disconfirmation of Negative Expectancies, 64 Journal of Personality and Social
Psychology 409–20 (1993).
51. James L. Hilton, and John M. Darley, Constructing Other Persons: A Limit on the
Effect, 21 Journal of Experimental Social Psychology 1–18 (1985); William B. Swann,
Jr., and Robin J. Ely, A Battle of Wills: Self-Verification Versus Behavioral Confirmation, 46
Journal of Personality and Social Psychology 1287–1302 (1984).
52. Claude M. Steele, A Threat in the Air: How Stereotypes Shape Intellectual Identity and
Performance of African Americans, 52 American Psychologist 613–29 (1997).
53. Claude M. Steele and Joshua Aronson, Stereotype Threat and the Intellectual
Performance of African Americans, Journal of Personality and Social Psychology 69:
797–811 (1995).
54. Ryan P. Brown, and Robert A. Josephs, A Burden of Proof: Stereotype Relevance and
Gender Differences in Math Performance, 76 Journal of Personality And Social
Psychology 246–57 (1999).
330 problem solving, decision making, and professional judgment
studies, stereotype threat effects have been found in Latinos and low socioeco-
nomic status Whites.55
The way in which a test is framed and the conditions under which it is admin-
istered have a significant impact on whether stereotype threat occurs. For exam-
ple, Steven Spencer, Claude Steele, and Diane Quinn demonstrated that the
women-math stereotype threat effect could be eliminated when a test was
described as being “gender fair.” 56 Becca Levy found that elderly participants
performed more poorly on a memory test after being primed with negative ste-
reotypes of the elderly, but performed better when primed with positive stereo-
types.57 Stereotype threat is also affected by whether stigmatized group status is
made salient before group members take a test. In one particularly fascinating
study, priming Asian American women with female identity before taking a
math test depressed their performance, while priming them with Asian identity
enhanced it.58
The mechanisms underlying stereotype threat are not yet well understood.
Claude Steele hypothesizes that, when members of negatively stereotyped groups
know that their performance is being assessed in a stereotype-relevant domain,
fear of confirming the negative stereotype impairs their performance. The effect,
he suggests, occurs because this sense of self-threat raises anxiety levels and
consumes cognitive resources.59 Over time, Steele notes, the effects of stereotype
threat on performance may cause promising minority and/or female students to
disidentify with certain academic performance domains. Ironically, he notes, the
very members of stigmatized groups who are, at the outset, most talented at a
particular academic endeavor are also the most likely over time to suffer from
stereotype threat and withdraw from academic pursuits in response to its painful
effects.
55. Margaret Shih, Nalini Ambady, Jennifer A. Richeson, Kentaro Fujita, and Heather
M. Gray, Stereotype Performance Boosts: The Impact of Self-Relevance and the Manner of
Stereotype Activation, 83 Journal of Personality and Social Psychology 638 (2002).
56. Steven J. Spencer, Claude M. Steele, and Diane M. Quinn. “Stereotype Threat
and Women’s Math Performance.” Journal of Experimental Social Psychology 35
(1999): 4–28.
57. Becca Levy, Improving Memory in Old Age Through Implicit Self-Stereotyping,
Journal of Personality and Social Psychology 71: 1092–1107 (1996).
58. Id.; Margaret Shih, Todd L. Pittinsky, and Nalini Ambady, Stereotype Susceptibility:
Identity Salience and Shifts in Quantitative Performance, 10 Psychological Science 80–83
(1999).
59. Claude. M. Steele, S. J. Spencer, and Joshua Aronson, Contending with Group
Image: The Psychology of Stereotype and Social Identity Threat, in Advances in Experimental
Social Psychology (Vol. 34), 379–440 (Mark Zanna ed., New York: Academic Press,
2002).
the social perceiver: processes and problems in social cognition 331
63. Lee Ross, The Intuitive Psychologist and His Shortcomings: Distortions in the Attribution
Process, in Advances in Experimental Social Psychology (Vol. 10) 173–220 (Leonard
Berkowitz ed., New York: Academic Press, 1977).
64. Edward E. Jones and Robert E. Nisbett, The Actor and the Observer: Divergent
Perceptions of the Causes of the Behavior, in Attribution: Perceiving the Causes of
Behavior 79–94 (Edward E. Jones, David E. Kanouse, Harold H. Kelley, Robert E. Nisbett,
Stuart Valins, and Bernard Weiner eds., Morristown, NJ: General Learning Press, 1972).
65. Fritz Heider, The Psychology of Interpersonal Relations (Hoboken, NJ: John
Wiley & Sons, Inc., 1958).
the social perceiver: processes and problems in social cognition 333
Bernard Weiner and his associates built on Heider’s insights to construct the
Heider/Weiner matrix to show how these parameters affect one’s judgment and
prediction of individuals’ behavior. (See Table 11.1.)
Stable Unstable
Behaviors attributed to internal, stable factors (i.e., skill, laziness, criminal pro-
pensity) are viewed both as highly predictive and warranting of moral judgment.
Behaviors attributed to internal, unstable factors (e.g., moods or how one priori-
tizes tasks on a particular occasion) are also viewed as relevant to moral judgment,
but as less predictive of future behavior. Attributions to external, stable factors dif-
fuse moral responsibility, but are seen as predictive, while attributions to external,
unstable factors do not provide a basis for either judgment or prediction.
Biases in causal attribution affect the exercise of professional judgment in
lawyering, judging, policy making, and management, not to mention in our per-
sonal lives. First, our tendency to attribute our own failings or negative outcomes
to situational factors (which we often overoptimistically view as transient) and to
attribute others’ failings or negative outcomes to internal, dispositional factors
makes us vulnerable to a host of ego-enhancing biases. These biases often impair
our willingness or ability to compromise with others, to be open to understand-
ing their points of view, and to predict accurately how well or poorly we or other
people will perform on future tasks.
Second, once we have formed an initial impression of another person, assign-
ing her a set of character traits, those traits operate as prior expectancies that
shape how we attribute her actions and outcomes in the future. Specifically, we
tend to attribute outcomes or behaviors that confirm our impressions to stable,
internal factors (the person’s character), while attributing events that contradict
our prior expectancies to transient or environmental causes. In this way as in
others, our initial person impressions tend to be self-perpetuating.
334 problem solving, decision making, and professional judgment
First, the bad news. Social psychology can presently tell us far more about how
social perception falls prey to the biasing effects of stereotypes and other social
expectancies than about how those biases can reliably be reduced or eliminated.
But there is good news as well. First, we do know something about bias reduction.
And second, debiasing is presently a subject of intense investigation in cognitive
social psychology, and it is reasonable to expect that a decade from now we will
know much more than we know now about how to constrain what psychologist
John Bargh calls “the cognitive monster.”66
66. John A. Bargh, The Cognitive Monster: the Case against the Controllability of Automatic
Stereotype Effects, in Shelly Chaiken and Yaacov Trope, Dual-Process Theories in
Social Psychology 361 (New York: Guilford Press, 1999).
67. James S. Uleman, Leonard S. Newman, and Gordon B. Moskowitz, People as
Flexible Interpreters: Evidence and Issues from Spontaneous Trait Inference, in Advances in
Experimental Social Psychology (Vol. 28) 211–79 (Mark P. Zanna ed., San Diego:
Academic Press, 1996); Timothy D. Wilson and Nancy Brekke, Mental Contamination and
Mental Correction: Unwanted Influences on Judgments and Evaluations, 116 Psychological
Bulletin 117 (1994); Daniel T. Gilbert, Brett W. Pelham, and Douglas S. Krull, On
Cognitive Busyness: When Person Perceivers Meet Persons Perceived, 54 Journal of
Personality and Social Psychology 733–40 (1988).
the social perceiver: processes and problems in social cognition 335
Notice, however, that correction of this kind would require three things:
Trujillo’s motivation to reconsider his initial characterization; the cognitive
resources (time, spare processing capacity) required to do so; and information
from which a reasoned corrective judgment might be made.
There is virtual consensus among social psychologists that stereotyping
includes both automatic and controlled components. If a person has learned, but
then consciously rejects, the stereotypes and attitudes associated with a devalued
social group, those stereotypes and attitudes do not just disappear. Indeed, tech-
niques like the Implicit Association Test68 show them to be present, even in
many participants who rate low on measures of explicit bias or who are them-
selves members of the negatively stereotyped group.69 In such situations, the
older implicit attitudes and associations continue to exist alongside the newer,
consciously held beliefs and commitments. The implicit attitudes function as a
Type 1 or “low road” system, while the conscious beliefs function as a Type 2 or
“high road” system. In fact, the neurological substrates underlying the implicit/
explicit distinction can even be observed through such technologies as functional
magnetic resonance imaging (fMRI).70
There is relative consensus within cognitive social psychology that stereotype
activation, when it occurs, is spontaneous. But there is also a growing consensus
that stereotype activation does not necessarily lead to stereotype expression, and
that stereotype activation itself can be affected by environmental factors.
To understand the state of current thinking on these issues, consider Figure
11.6, an extension of Jerry Kang’s model of social categorization portrayed ear-
lier in Figure 11.1.
As Figure 11.6 suggests, even where a social schema (like a stereotype) is
activated and generates a schema-consistent impression, that initial impression
can be overridden through the application of conscious, “high road” thinking.71
68. See note 44, supra. A great deal has been written about the Implicit Association Test.
For a relatively brief description of the instrument, see Anthony G. Greenwald & Linda
Hamilton Krieger, Implicit Bias: Scientific Foundations, 94 Cal. L. Rev. 945–967 (2006).
69. Nilanjana Dasgupta, Implicit Ingroup Favoritism, Outgroup Favoritism, and Their
Behavioral Manifestations, 17 Social Justice Research 143 (2004); Brian A. Nosek,
Mahzarin R. Banaji, and Anthony G. Greenwald, Harvesting Implicit Group Attitudes and
Beliefs from a Demonstration Web Site, 6 Group Dynamics 101 (2002).
70. William. A. Cunningham, Marcia K. Johnson, Carol L. Raye, J. Chris Gatenby,
John C. Gore and Mahzarin R. Banaji, Neural Components of Social Evaluation, 85 Journal
of Personality and Social Psychology 639–49 (2003); E. A. Phelps, et al., Performance
on Indirect Measures of Race Evaluation Predicts Amygdala Activation, supra note 47.
71. Margo J. Monteith and Corrine I. Voils, Exerting Control over Prejudiced Responses,
in Cognitive Social Psychology: The Princeton Symposium on the Legacy and
Future of Social Cognition (Gordon B. Moskowitz ed., Mahwah, NJ: Erlbaum, 2001).
336 problem solving, decision making, and professional judgment
d
Which May
oa
Trigger
R
gh
Hi
Spontaneous Schematic Content
Impression
Which
generates
72. John A. Bargh, The Cognitive Monster: The Case Against the Controllability of
Automatic Stereotype Effects, in Shelly Chaiken and Yaacov Trope, Dual-Process
Theories in Social Psychology 361 (1999); Timothy D. Wilson and Nancy Brekke,
Mental Contamination and Mental Correction: Unwanted Influences on Judgments and
Evaluations, 116 Psychological Bulletin 117 (1994).
73. There is some evidence indicating that, with sufficient motivation and effort on the
perceiver’s part, even these subtle behaviors are amenable to conscious control. See Steven
L. Neuberg, Expectancy-Confirmation Processes in Stereotype-Tinged Social Encounters: The
Moderating Role of Social Goals, in The Psychology Of Prejudice: The Ontario
Symposium (Vol. 7) 103–30 (Mark P. Zanna and John M. Olson eds., Hillsdale, NJ:
Erlbaum, 1994).
the social perceiver: processes and problems in social cognition 337
Consider how the presence or absence of these factors might influence the
degree of racial bias present in hiring decisions. If those who make hiring deci-
sions are simply told to be “color-blind” in evaluating potential candidates, if
they do not believe themselves to be race-biased, and if there are no meaning-
fully endorsed and enforced system of goals relating to minority hiring, con-
scious correction of implicitly race-biased evaluations is unlikely to occur. Under
such conditions, hiring decision makers are unlikely to be aware of the possibil-
ity that implicit racial stereotypes and attitudes could be skewing their judgment,
and they will have little motivation to engage in conscious correction.
In the hiring or promotion context, at least with respect to high-level jobs,
ultimate decision making is often quite deliberate, with ample time available
for systematic evaluation of the final competing candidates. In the formation
of managers’ day-to-day impressions of employee performance, however, this
may not be the case. Managers often function under conditions of extreme time
constraint and high cognitive load. In these situations, correction of stereotype-
influenced impressions is far less likely to occur.74
As earlier mentioned, to correct biased impressions, decision makers also
need ample decision-relevant information and access to a structured decision-
making processes. Carefully spelling out the applicable evaluative criteria, pro-
viding decision makers with objective, criterion-relevant information, and
requiring them to write down how the information provided relates to the rele-
vant criteria can help reduce the biasing effect of stereotypes on evaluations.75
Additionally, providing decision makers with systemwide information revealing
broad patterns of good or bad outcomes can help them identify biases they likely
would not recognize from case-by-case data.76 Finally, as with any expectancy,
“considering the opposite” can help correct expectancy-confirmation bias.
In summary, even though the implicit expectancies created by stereotypes
and other social knowledge structures bias initial impression formation
spontaneously, bias is, at least under certain conditions, amenable to correction.
Individuals or organizations interested in maximizing the likelihood that correc-
tion will occur should implement the following strategies:
• educate decision makers that their evaluative judgments may be influenced
by subtle forms of bias, no matter how sincere their conscious commit-
ment to fairness, objectivity, and nondiscrimination;
74. John A. Bargh, The Cognitive Monster: The Case Against the Controllability of
Automatic Stereotype Effects, in Shelly Chaiken and Yaacov Trope, Dual-Process
Theories in Social Psychology 361 (New York: Guilford Press, 1999); Daniel T. Gilbert,
Brett W. Pelham, and Douglas S. Krull, On Cognitive Busyness: When Person Perceivers Meet
Persons Perceived, 54 Journal of Personality and Social Psychology 733–40 (1988).
75. Veronica F. Nieva and Barbara A. Gutek, Sex Effects on Evaluation, 5 Acad. of
Mgmt. Rev. 267, 270–71 (1980).
76. Faye J. Crosby, Susan Clayton, Olaf Alksnis, and Kathryn Hemker, Cognitive Biases
in the Perception of Discrimination: The Importance of Format, 14 Sex Roles 637–46 (1986).
338 problem solving, decision making, and professional judgment
77. John A. Bargh, supra note 74, The Cognitive Monster: The Case Against the
Controllability of Automatic Stereotype Effects, in Shelly Chaiken and Yaacov Trope,
Dual-Process Theories in Social Psychology 361 (1999); Susan T. Fiske, Examining
the Role of Intent: Toward Understanding Its Role in Stereotyping and Prejudice, in
Unintended Thought, supra note 62, pp. 253–83.
78. Patricia G. Devine, Stereotypes and Prejudice: Their Automatic and Controlled
Components, supra note 8; Susan T. Fiske, Monica Lin, and Steven L. Neuberg, The
Continuum Model: Ten Years Later, in Dual-Process Theories in Social Psychology
231–54 (Shelly Chaiken and Yaacov Trope eds., New York: Guilford Press, 1999).
the social perceiver: processes and problems in social cognition 339
79. Irene V. Blair and Mahzarin R. Banaji, Automatic and Controlled Processes in
Stereotype Priming, 70 Journal of Personality and Social Psychology 1142–63 (1996).
80. Kerry Kawakami, John F. Dovidio, J. Moll, S. Hermsen, and A. Russin, Just Say No
(to Stereotyping): Effect of Training in the Negation of Stereotypic Associations on Stereotype
Activation, 78 Journal of Personality and Social Psychology 871–88 (2000).
81. Nilanjana Dasgupta and Anthony G. Greenwald, On the Malleability of Automatic
Attitudes: Combating Automatic Prejudice with Images of Admired and Disliked Individuals,
81 Journal of Personality and Social Psychology 800–14 (2001).
82. Irene V. Blair, Jennifer E. Ma, and Alison P. Lenton, Imagining Stereotypes Away:
The Moderation of Implicit Stereotypes Through Mental Imagery, 81 Journal of Personality
and Social Psychology 828–41 (2001).
83. Lisa Sinclair and Zeva Kunda, Reactions to a Black Professional: Motivated Inhibition
and Activation of Conflicting Stereotypes, 7 Journal of Personality and Social
Psychology 885–904 (1999).
84. Jennifer A. Richeson and Nalini Ambady, When Roles Reverse: Stigma, Status, and
Self-Evaluation, 31 Journal of Applied Social Psychology 1350–78 (2001).
85. Brian S. Lowery, Curtis D. Hardin, and Stacey Sinclair, Social Influence Effects on
Automatic Racial Prejudice, 81 Journal of Personality and Social Psychology 842–55
(2001).
86. Gretchen B. Sechrist and Charles Stangor, Perceived Consensus Influences Intergroup
Behavior and Stereotype Accessibility, 80 Journal of Personality and Social Psychology
645–54 (2001).
340 problem solving, decision making, and professional judgment
messages we absorb from the media, making conscious decisions to limit expo-
sure to negatively stereotyped images of women and minorities, or making sure
that we supplement those we do take in with positive, non-stereotyped portrayals
as well. Organizations can take affirmative steps to ensure that members of ste-
reotyped groups are present in leadership positions. We can change the pictures
on the walls of our homes, workplaces, and schools, even on our screen savers.
These situational modifications can’t hurt, and in another decade, empirical
research may demonstrate that, alone or in combination with other factors, they
help a great deal.
part three
making decisions
People make myriad decisions every day. Most are of little consequence: What
clothes should I wear to work? What should I order for lunch? Other decisions
carry more weight: What job should I take? Should I enter into this deal, and on
what terms? What litigation strategy should I employ? Where should we site the
wastewater treatment plant? It is the presence of options that creates both oppor-
tunities and a decision problem.
Part 1 introduced a general framework for deliberative decision making. To
focus specifically on decision making (rather than problem solving in general),
the framework entails:
1. Identifying and prioritizing the relevant interests and objectives.
2. Generating a range of plausible decision alternatives. (Sometimes it’s a
simple “yes” or “no” decision; sometimes, as in the wastewater treatment
plant decision, there are numerous alternatives.)
3. Predicting the consequences of those alternatives in terms of the interests
or objectives.
4. Comparing the alternatives and selecting the course of action that opti-
mizes the interests or objectives.
We covered the first two of these in Part 1. The third is fundamentally an
empirical undertaking, often requiring probabilistic assessments of the kinds
explored in Part 2 of the book. A decision process is good to the extent that the
underlying empirical analysis follows the scientific method and that judgments
of uncertain consequences adhere to the rules of probability theory.
We got a head start on the fourth step in Chapter 4, with the survey of various
strategies for making decisions, and working through an example of a formal
model of decision making—the subjective linear model—in which the decision
making specifies the major interests involved and assigns weights to them. We
continue this exploration in Part 3.
Chapter 12, Choices, Consequences, and Trade-offs, asks what people seek to
achieve when they are making decisions. It introduces the concept and com-
plexities of utility and the criteria for rational decision making, focusing on a
widely accepted set of axioms for rational decision making—the axioms of subjec-
tive expected utility theory.
342 problem solving, decision making, and professional judgment
2. Barry Schwartz, The Paradox of Choice: Why More Is Less (New York: Ecco/
Harper Collins Publishers, 2004).
choices, consequences, and trade-offs 345
theories that affirm that there are a multitude of good lives and that each individ-
ual is in a unique position to ascertain which of the many reasonable conceptions
of the “good” best fit her circumstances and express her individuated soul.3
Daniel Keys and Barry Schwartz write that a complete theory of rational deci-
sion making “must consider the very broadly construed consequences of a deci-
sion. That is, it must consider short- and long-term consequences, consequences
to the self and to others, consequences that are central to the decision at hand,
and consequences that may be more peripheral. It must also consider conse-
quences of decisions for the character of the decision maker, as the effects on
character may have a significant impact on a host of future decisions.”4 The next
several chapters suggest how difficult it is for even very thoughtful people to
predict whether particular courses of action will serve their individual ends, let
alone those of various stakeholders.
The official policy of Bhutan is that “gross national happiness is more important
than gross national product,” because “happiness takes precedence over eco-
nomic prosperity in our national development process.”5
Bhutan’s constitution assesses government programs in terms of the economy,
culture, environment, and good governance. The government employs seventy-
two indicators in nine different domains: psychological well-being, ecology, health,
education, culture, living standards, time use, community vitality, and good gover-
nance. For example, indicators of psychological well-being include the frequencies
of prayer and meditation and of feelings of selfishness, jealousy, calm, compassion,
and generosity, and the absence of frustration and suicidal thoughts.6
People sometimes use “happiness” as a synonym for utility, but that is some-
what misleading. Many people have a conception of leading a “good” or “whole”
3. Mark G. Kelman, Law and Behavioral Science: Conceptual Overviews, 97 Nw. L. Rev.
1347, 1358 (2003).
4. Daniel J. Keys and Barry Schwartz, “Leaky” Rationality: How Research on Behavioral
Decision Making Challenges Normative Standards of Rationality, 2 Perspectives on
Psychological Science 162–80 (2007).
5. Orville Schell, “Frontline World, Bhutan”—The Last Place, May 2002, http://www.
pbs.org/frontlineworld/stories/bhutan/gnh.html.
6. Seth Mydans, Thimphu Journal: Recalculating Happiness in a Himalayan Kingdom,
New York Times, May 7, 2009, http://www.nytimes.com/2009/05/07/world/asia/07bhutan.
html
346 problem solving, decision making, and professional judgment
Patrick has a serious phobia about flying. The very anticipation of flight creates an
almost physical sense of dread. As a result, he avoids business and pleasure trips
that require flying, and he drives extraordinarily long distances, all at tremendous
professional and personal costs. Patrick believes that statistics show that flying is
safe compared to driving and, indeed, to many other activities he engages in. He
acknowledges that the phobia is irrational. Is it irrational for him to take account
of his fear of flying in making travel decisions?
The following sections note some other sources of utility that do not seem
reducible to purely material ends.
12.1.1.b Existence and Option Value Considerable research has gone into
measuring the material benefits of ecosystems—for example, as sources of pure
drinking water. But some environmental legislation, such as the Endangered
Species Act, treats species, habitats, and natural landscapes as having an exis-
tence value independent of any human uses—for example, the value attributed to
the protection of an endangered species in a remote part of the world that most
people will never encounter. A cousin of existence value is the option value of a
natural resource—the value one assigns to the possibility of using it sometime
in the future. For example, you may have no plans to visit a wilderness area but
nonetheless place some value on the possibility of visiting it in the future. We’ll
say more about how economists assess these values in the discussion of cost-
benefit analysis in Section 12.3.
12.1.1.c Affection In a charming mind experiment, Christopher Hsee and
Howard Kunreuther demonstrate what they call the affection effect.11
On a visit to Europe, Helen and Laura each bought a painting for $100 and
shipped it back to the United States, each purchasing a $100 shipping insur-
ance policy. The paintings were damaged beyond repair, but claiming the
insurance will require a time-consuming trip to the shipping company. Helen
loved her painting a great deal. Laura thought hers was just OK. Who is more
likely to spend the time collecting the insurance compensation?
From a purely economic point of view, there is no difference between the
choices facing Helen and Laura: $100 is $100, whatever its source and rationale.
However, Helen, who has great affection for the painting, is more likely to take
11. Christopher K. Hsee and Howard C. Kunreuther, The Affection Effect in Insurance
Decisions, 20 Journal of Risk and Uncertainty 141 (2000). We have simplified the
scenarios.
348 problem solving, decision making, and professional judgment
the time to collect the insurance compensation than Laura. Hsee and Kunreuther
explain the difference in terms of what they call the “consolation hypothesis”:
Compensation has symbolic value. It is redemption for the lost object; it is a
token of consolation for the pain caused by the loss. The utility of a given
amount of compensation has two components: (a) the utility of that particular
amount of money, and (b) the utility of its symbolic consolation value. . . . The
consolation value, in turn, depends on one’s level of affection for the object.
The more affection one has, the more pain one experiences, and the more one
needs the consolation.
The main focus of this chapter is on decisions made in advance of knowing
just what the consequences will be—for example, whether one will suffer a loss.
But just as individuals anticipate the experience of regret, they can anticipate the
need for consolation. In another part of the experiment, Hsee and Kunreuther in
effect asked Laura and Helen how much they would pay (in advance) for insur-
ance to ship their paintings. Not surprisingly, Helen, who has strong affection
for her work of art, would pay significantly more than Laura—almost twice as
much. By the same token, people would be more willing to buy a warranty for a
beautiful used convertible than for a utilitarian used station wagon even if the
expected repair expenses and the cost of the warranty were identical.12
Suppose that your dog is dying of either Virus A or Virus B. They are equally likely,
but which one it is cannot be known. There are different medicines for Virus A
and Virus B, each equally effective for its intended target, that cost $1 and $20
respectively. They interact in a way that they cannot be given together. Which one
would you choose?
When the dog was described as friendly or faithful, 43 percent of respondents
would have opted for the $20 medicine and said that they would have felt better
paying more. The effects were significantly diminished when the dog was
described as unfriendly or unfaithful.
Christopher Hsee and Howard Kunreuther, The Affection Effect in Insurance
Decisions, 20 JOURNAL OF RISK AND UNCERTAINTY 141 (2000).
12.1.1.d The Allure of the Free Suppose you were offered the choice between
buying a Hershey’s Kiss for 1 cent and buying a delicious Swiss truffle, made
by Lindt, for 15 cents; which would you choose? Now imagine that instead of
costing 1 cent, the Hershey’s Kiss were free and the Lindt truffle cost 14 cents.
Would your preference be different? Research by Dan Ariely and his colleagues13
suggests that it would. When they sold Hershey’s Kisses and Lindt truffles to the
public for 1 cent and 15 cents respectively, 73 percent of their “customers” went
for the Lindt truffle, which, considering their usual retail prices, is a much better
deal. However, when they dropped the price of each chocolate by a penny, so the
Kiss was free and the truffle cost 14 cents, only 31 percent opted for the truffle.
The rest—a large majority—now chose the Kiss.
The relative value of the two chocolates had not changed, and nor had the
price difference. So why did people’s choices change? Ariely and his colleagues
suggest that we enjoy a benefit when it is untainted by costs, which take an affec-
tive toll that compromises our enjoyment of the benefit. (By the same token, we
may enjoy a vacation more if we de-couple the experience from its cost by paying
in advance.)
13. Kristina Shampanier, Nina Mazar, and Dan Ariely, Zero as a Special Price: The True
Value of Free Products, 26 Marketing Science 742–57 (2007); Dan Ariely, Predictably
Irrational: The Hidden Forces That Shape Our Decisions (New York: HarperCollins
2008).
14. Daniel J. Keys and Barry Schwartz, “Leaky Rationality: How Research on Behavioral
Decision Making Challenges Normative Standards of Rationality, 2 Perspectives on
Psychological Science 162–80 (2007). The authors acknowledge Deborah Frisch’s
article, Reasons for Framing Effects, 54 Organizational Behavior and Human Decision
Processes 399–429 (1993), for the concept of what they term “leakage.”
350 problem solving, decision making, and professional judgment
The “dictator” game controls for the proposer’s anticipation of the respondent’s
response. The dictator divides a fixed amount of money between herself and the
“recipient,” who cannot respond. When the game is played anonymously with
participants knowing nothing about the other player, most dictators keep the
entire sum or allot a paltry amount to the recipient.20 However, when dictators are
provided information about recipients—pictures or surnames, for example—they
15. Jonathan D. Casper, Tom Tyler, and Bonnie Fisher, Procedural Justice in Felony
Cases, 22 Law and Society Review 483–508 (1988).
16. Donna Shestowsky, Misjudging: Implications for Dispute Resolution, 7 Nev. L.J. 487,
493 (2007).
17. Colin Camerer and Richard H. Thaler, Anomalies: Ultimatums, Dictators and
Manners, 9 The Journal of Economic Perspectives 210 (1995).
18. Elizabeth Hoffman, Kevin A. McCabe, and Vernon L. Smith, On Expectations
and the Monetary Stakes in Ultimatum Games, 25 International Journal of Game
Theory 289–301 (1996).
19. Alvin E. Roth, Vesna Prasnikar, Masahiro Okuno-Fujiwara, and Shmuel Zamir,
Bargaining and Market Behavior in Jerusalem, Ljubljana, Pittsburgh, and Tokyo: An
Experimental Study, 81 American Economic Review 1068–95 (1991).
20. Elizabeth Hoffman, Kevin McCabe, Keith Shachat, and Vernon Smith, Preferences,
Property Rights, and Anonymity in Bargaining Games, 7 Games and Economic Behavior
346–80 (1994); Elizabeth Hoffman, Kevin McCabe, and Vernon Smith, Social Distance and
Other-Regarding Behavior in Dictator Games, 86 American Economic Review 653–60 (1996).
choices, consequences, and trade-offs 351
increase their giving.21 The dictator’s behavior is also affected if she adverts to
being observed. In an experiment where players conducted the game on comput-
ers to ensure anonymity, dictators gave more when the background of the com-
puter screen featured a robot with human eyes.22 Commenting on the ultimatum
game, Martin Nowak suggests that objecting to unfairness is a way of asserting
one’s reputation: “Rejecting low offers is costly, but the cost is offset by gaining
the reputation of being somebody who insists on a fair offer.”23 In the dictator
game, perhaps the image of the observer reminds the dictator of her reputational
interests.
be affected by framing the same item in different terms.24 People given a yogurt
described as having “only 5% fat” found it richer tasting than the same food
described as “95% fat-free.” They thought that 7UP in a yellow bottle tasted more
like lemon than the same drink in a green bottle.
Deborah Frisch asserts that “framing has an effect on decisions because it has
an effect on experience.”25 But does framing actually change people’s experiences
or just provide independent information about the utility of a product or experi-
ence? That is, do people think 7UP in a yellow bottle tastes more like lemon
because the color of the bottle changes the way the drink tastes to them, or does
the bottle’s color provide them with independent information about how lemony
the drink is? Leonard Lee, Shane Frederick, and Dan Ariely probed this question
with a series of clever experiments involving beer.26 They approached patrons of
a pub on MIT’s campus and offered them two small free samples of beer. One
was ordinary Budweiser or Samuel Adams and the other was called “MIT Brew,”
which was prepared by adding a few drops of balsamic vinegar to ordinary beer.
Participants all drank the two small samples of beer and then were offered a
full glass of the beer of their choice. In the blind condition, participants tasted the
two samples and choose their free beer without knowing MIT Brew’s secret
ingredient. In the before condition, participants were told the ingredients in both
samples before tasting them, and in the after condition, they were told the ingre-
dients after they had tasted the samples but before they chose their free beer.
Although beer mixed with vinegar may sound unappealing, the majority of
participants in the blind condition preferred the MIT Brew to the unadulterated
beer. The majority of participants in the before condition, turned off by the pros-
pect of vinegar-laced beer, preferred the unadulterated beer to the MIT Brew.
What did people choose in the after condition? If our expectations (about the
taste of vinegar) provide independent information that influences our ratings of
utility, then it should not matter whether this information comes before or after
tasting the samples. The information should simply be combined with the taste
information to reach a conclusion about preference. If, on the other hand, our
expectations affect the experience of tasting the beer, then they should have a
stronger effect when independent information is presented before tasting than
after, when it is too late to affect the tasting experience. As it turned out, when
participants were told about the vinegar in the MIT Brew after tasting the sam-
ples, they still liked it as much as those in the blind condition and significantly
more than those in the before condition. Thus participants’ expectations actually
influenced their experience of the beer.
Lee and colleagues speculate that expectations may have affected people’s
experiences by biasing them to focus on the negative aspects of the complex expe-
rience of tasting beer. People may then have erroneously attributed those nega-
tive aspects of the experience to the presence of vinegar. (This is not unlike the
process involved in confirmation bias, discussed in Section 10.3.) When people
expected the beer to taste bad, they may have searched for evidence in the tasting
experience to confirm that expectation, ignoring evidence to the contrary.27
Another set of experiments, by Baba Shiv, Ziv Carmon, and Dan Ariely, indi-
cates that the price we pay for a product can affect the utility we get from it.28 In
a series of three studies, participants were given a number of difficult anagrams
to solve (for example, participants were given the letters TUPPIL and asked to
rearrange them to form a word—in this case PULPIT). Before doing the ana-
gram, participants were given a bottle of SoBe Adrenaline Rush (an iced tea drink
that claims on its package to improve mental functioning). Some of them were
asked to pay the full retail price for their bottle of iced tea ($1.89) while others
were given a heavily discounted price (89 ). The price participants paid affected
their subsequent performance on the anagram task. Those who paid less for the
iced tea consistently performed worse than those who paid full price. They had
lower expectations about how much of a boost they would get than those who
paid full price, and this led them to perform less well on the anagram task.
In another study, Shiv and colleagues also found that inflating expectations
about the effectiveness of the drink led to a significant increase in performance
on the anagram task. Some participants were told that numerous studies had
established that SoBe produced a “significant” boost in mental acuity, while
others were told that it only produces a “slight” boost. People experienced what
they were led to expect: those who were led to expect a “significant” boost experi-
enced just that, and those who were led to expect a “slight” boost had a slightly
deflated performance. Price, meanwhile, continued to play its role: those who
believed that SoBe produced a “significant” boost and paid full price did best on
the anagram task. Those who paid a discounted price did and believed that SoBe
only produced a “slight” boost to begin with did worse.
In a related experiment, participants were exposed to a series of painful elec-
tric shocks after being given a placebo pill that they were told was a new type of
painkiller.29 Some were told the painkiller’s retail price would be $2.50 per dose
while others were told it had been discounted from $2.50 to 10 cents. As you
27. See also Stephen J. Hoch and Young-Won Ha, Consumer Learning: Advertising and
the Ambiguity of Product Experience, 13 Journal of Consumer Research 221–33 (1986).
28. Baba Shiv, Ziv Carmon, and Dan Ariely, Placebo Effects of Marketing Actions:
Consumers May Get What They Pay For, 42 Journal of Marketing Research 383–93
(2005).
29. Rebecca Waber, Baba Shiv, Ziv Carmon, and Dan Ariely, “Paying More for Less
Pain,” working paper, MIT (2007).
354 problem solving, decision making, and professional judgment
might expect by now, participants who were told the pill was expensive reported
that they experienced significantly less pain than those who were told it was
inexpensive. This effect may depend on people not adverting to the inference
they are making based on price. In one of the iced tea studies, when participants
were asked how effective they thought the SoBe would be at boosting their per-
formance given the price they had paid for it, the effect of price on performance
was completely erased. (Recall the discussion of metacognition, in Section 9.6.2.)
Later in this chapter and in the next, we discuss other situations in which the
decision process may affect satisfaction with the outcome. Perhaps the most
pervasive of these involve the phenomenon of regret, closely followed by the
effects of choice overload—of being faced with a decision that involves too many
choices.
Disconnecting the decision process from the experience falls into what Daniel
Gilbert and Jane Ebert (2002) have called the “illusion of intrinsic satisfaction”—
people’s belief that “hedonic experiences [are] due entirely to the enduring intrin-
sic properties of their outcomes, as though the wonderfulness or awfulness they
are experiencing was always there ‘in the outcome’ waiting to be experienced,
and none of these properties were altered or induced by the mere act of making
the outcome their own.”30 Daniel Kahneman similarly writes that “a theory of
choice that completely ignores feelings such as the pain of losses and the regret
of mistakes is not just descriptively unrealistic. It also leads to prescriptions that
do not maximize the utility of outcomes as they are actually experienced.”31
30. Daniel T. Gilbert and Jane E. J. Ebert, Decisions and Revisions: The Effective
Forecasting of Changeable Outcomes, 82.4 Journal of Personality and Social Psychology
503–14 (2002).
31. Daniel Kahneman, A Perspective on Judgement and Choice, 58.9 American
Psychologist 697–720 (2003).
32. See Jon Elster and George Loewenstein, Utility from Memory and Anticipation, in
Choice Over Time (George Loewenstein and Jon Elster eds., New York: Russell Sage
Foundation 1992).
choices, consequences, and trade-offs 355
experience and the forward effect of contemplating the future. Whether through
memory or anticipation, the nonprimary emotions can result from:
• the consumption effect, where present positive or negative emotions are
based on reliving positive or negative experiences from the past or antici-
pating them in the future (e.g., remembering a lovely concert or dreading
a trip to the dentist), or
• the contrast effect, where one’s present emotions are affected by the contrast
of one’s present primary experience with one’s recalled or anticipated expe-
rience in the past or future (e.g., remembering what it was like to have a
well-paying job when currently unemployed).
“While each of the Wosileskis had experienced their rare moments of happiness,
John thought that they might have been better off had they not; that the moments of
happiness served only to accentuate despair. Had John . . . never known those excru-
ciatingly beautiful seconds with Valentine Kessler, had Mrs. Wosileksi, twenty-four
years ago, not walked down the aisle as a grateful bride and hopeful mother-to-be,
had Mr. Wosileski not, those same twenty-four years ago minus a few months,
known a surge of pride and love at the sight of his newborn, had he not twice—-
twice!—bowled a perfect game, then the Wosileskis might not have understood that
they were unhappy now. With nothing to compare it to, a dreary life would’ve been a
flatline, which is something like a contented one. But they did know rapture and so
the loss of it, the fleetingness of it, rendered the sadness that much more acute.”
—BINNIE KIRSHENBAUM, AN ALMOST PERFECT MOMENT 180 (2004).
I love that goodnight kiss that I reached the stage of hoping it would come as late
as possible, so as to prolong the time of respite during which Mamma would not
have appeared.” As Elster and Loewenstein say: “On the one hand, the narrator
wants to be able to look forward to the goodnight kiss (a forward consumption
effect). On the other hand, he doesn’t want to be in a position in which the kiss
has already occurred (a backward contrast effect).”
Proust’s desire to postpone his mother’s kiss, and thus savor its prospect,
deviates from people’s general tendency to be impatient (captured in the typical
model of discounting discussed in Section 13.6)—the preference to enjoy a good
experience sooner rather than later. On the negative side, people sometimes
prefer to hasten experiencing a bad event rather than dread its prospect over a
period of time. In an experiment by Gregory Berns and his colleagues,33 partici-
pants chose to receive an electric shock sooner rather than later, and some pre-
ferred to receive a more severe shock now rather than delay a smaller one. As
with the instances of deferring enjoyment, the authors suggest that “waiting
enters the utility function separately from the outcome.” Hastening even an
unpleasant outcome affords relief from dread.34
12.1.3.b The Uncertain Relationship Between Momentary and Recollected
Experience (1) We experience events at the moment they occur, (2) we remember
events in retrospect, and (3) we predict how we are likely to feel about similar
events in the future. There can be significant differences between momentary
and recalled experience. For example, a study of students’ spring-break vacations
indicated that their recollections were more positive than their experience at the
time. The students’ prediction of future experiences was based largely on the
remembered experience.35 As Dave Barry notes, “The human race is far too
stupid to be deterred from tourism by a mere several million years of bad
experiences.”36
Daniel Kahneman and Jason Riis write:37
When we are asked “how good was the vacation,” it is not an experiencing self
that answers, but a remembering and evaluating self, the self that keeps score
33. Gregory S. Berns et al., Neurobiological Substrates of Dread, 312 Science 754–58 (2006).
34. Id. 754, 757.
35. Derrick Wirtz, Justin Kruger, Christie Napa Scollon and Ed Diener, What to Do on
Spring Break? The Role of Predicted, On-line, and Remembered Experience in Future Choice,
14 Psychological Science 520 (2003). Interestingly the students recollected more nega-
tive as well as positive moments than they experienced at the time, and overestimated the
intensity of the recalled experience. The authors speculate that one tends not to remember
the (many) moments of relatively neutral affect.
36. Dave Barry’s Only Travel Guide You’ll Ever Need (New York: Ballantine 1991).
37. Daniel Kahneman and Jason Riis, Living and Thinking About It: Two Perspectives on
Life, in The Science of Well Being (Felicia A. Huppert, Nick Baylis, and Barry Keverne
eds., Oxford: Oxford University Press 2005).
choices, consequences, and trade-offs 357
and maintains records. Unlike the experiencing self, the remembering self is
relatively stable and permanent. It is a basic fact of the human condition
that memories are what we get to keep from our experience, and the only
perspective that we can adopt as we think about our lives is therefore that of
the remembering self. For an example of the biases that result from the
dominance of the remembering self, consider a music lover who listens raptly
to a long symphony on a disk that is scratched near the end, producing a
shocking sound. Such incidents are often described by the statement that the
bad ending “ruined the whole experience.” But in fact the [momentary] expe-
rience was not ruined, only the memory of it.
The recollection of painful experiences appears to follow a quite robust rule,
called the peak-end phenomenon. Figure 12.1 shows the pain of two patients, as
reported every ten minutes, while they were undergoing colonoscopies (before
the use of the most recent sedatives). Patient A’s procedure lasted less than ten
minutes, and included one period when he reported a pain level of 7, and one
with an ending pain level of 8 on a 10-point scale. Patient B’s procedure lasted
twenty-five minutes, with one period at a pain level of 8 and several with pain
levels ranging from 4 to 6,and with the last few minutes being relatively painless.
Taking the amount of the pain and its duration into account, it seems apparent
that Patient B had a more painful experience than Patient A. Yet, when asked
about the experience retrospectively, Patient B reports having had a less painful
procedure.
In this and similar experiments—e.g., immersing one’s hand in freezing
water—people’s retrospective evaluation follows the peak-end rule: they report
the painfulness of the event as the arithmetic mean of the most painful moment
and the last moments of the experience.38
It is not surprising that retrospective assessments of past affective experi-
ences differ from affect at the moment of the experience. Like other memories,
retrospective assessments are not simply “remembered”; they are constructed.
What is so striking about the peak-end phenomenon is its apparent duration
neglect, which violates the assumption of temporal monotonicity—that adding
moments of negative affect should make the experience worse.
The most obvious explanation for the peak-end phenomenon is that the peak
and end of an experience may be the two events most available to one’s memory.
Barbara Fredrickson also suggests that, like the representative heuristic, it is an
example of judgment by prototype, where a “global evaluation of a set is domi-
nated by the evaluation of its prototypical element or exemplar.” Analogous to
ignoring base rates in identifying Jack’s occupation (Section 8.5.4), people
38. See generally Barbara L. Fredrickson, Extracting Meaning from Past Affective
Experience: The Importance of Peaks, Ends, and Specific Emotions, 14 Cognition and
Emotion 577 (2000). The phenomenon manifests itself in rats as well as humans.
358 problem solving, decision making, and professional judgment
Patient A
6
Pain intensity
0
0 10 20
Time (minutes)
Patient B
6
Pain intensity
0
0 10 20
Time (minutes)
ignore the duration of the event and focus on the prototypical moments of peak
and end.
Fredrickson proposes the alternative (but not inconsistent) explanation that
peaks and ends are particularly meaningful aspects of experience. Peak discomfort
choices, consequences, and trade-offs 359
the difference between belief-based generic judgments (‘I enjoy my kids’) and
specific episodic reports (‘But they were a pain last night’).”42
42. Daniel Kahneman, Alan B. Krueger, David Schkade, Norbert Schwarz, and Arthur
A. Stone, A Survey Method for Characterizing Daily Life Experience: The Day Reconstruction
Method (DRM), 306 Science 1776–80 (2004).
43. Betsey Stevenson and Justin Wolfers, “Economic Growth and Subjective Wellbeing:
Reassessing the Easterlin Paradox,” Institute For the Study of Labor Discussion Paper No.
3654 1–80 (August 2008).
choices, consequences, and trade-offs 361
Utility
Wealth
between the two government positions may be more important to her than the
$10,000 difference between the two higher paying positions.
We’ll return to these concepts below, but definitionally, the expected value
(EV) of Christine’s choice is her expected net salary. You could compute it just
using a calculator and without knowing anything about her individual prefer-
ences. Her subjective satisfaction is the expected utility (EU) of the choice. The
declining marginal utility of money (and some other goods as well) is an impor-
tant category in which EU deviates from EV.
12.1.4.c Psychophysical Numbing
One death is a tragedy; a million is a statistic.
—Joseph Stalin
I am deeply moved if I see one man suffering and would risk my life for him. Then
I talk impersonally about the possible pulverization of our big cities, with a hundred
million dead. I am unable to multiply one man’s suffering by a hundred million.
—Albert Szent-Gyorgi
Participants in a study were given a scenario in which a foundation was pre-
pared to make a $10 million grant to a medical institution to implement a new
treatment to reduce deaths from disease. They were told about three institutions
concerned with three different diseases with respective annual fatalities of
15,000, 160,000, and 290,000, and asked the minimum number of lives the insti-
tution must save to merit the grant. Two-thirds of the respondents required that
more lives be saved as the size of the at-risk population increased, requiring that
the institution dealing with the largest disease group save eleven times more
people than the institution addressing the smallest disease group.44
44. Quoted in David Fetherstonhaugh, Paul Slovic, Stephen Johnson, and James
Friedrich, Insensitivity to the Value of Human Life: A Study in Psychophysical Numbing, 14
Journal of Risk and Uncertainty 28 (1997).
362 problem solving, decision making, and professional judgment
47. See Varda Liberman and Lee Ross, Idiosyncratic Matching and Choice: When Less Is
More, 101 Organizational Behavior & Human Decision Processes (2006); Itamar
Simonson, Ziv Carmon, and Suzanne O’Curry, Experimental Evidence on the Negative
Effect of Product Features and Sales Promotion on Brand Choices, 13 Marketing Science 23
(1994); Ran Kivetz and Itamar Simonson, Earning the Right to Indulge: Effort as a
Determinant of Customer Preferences Toward Frequency Reward Programs, 39 Journal of
Marketing Research 155 (2002); Ran Kivetz and Itamar Simonson, The Role of Effort
Advantage in Consumer Response to Loyalty Programs, 40 Journal of Marketing Research
454 (2003).
364 problem solving, decision making, and professional judgment
3. Once an unexpected event occurs and people have a relatively intense emo-
tional reaction, they tend to make sense of the event, quickly and auto-
matically.
4. When people make sense of an event, it no longer seems surprising or
unexpected, and as a result they think about it less and it produces a less
intense emotional reaction. The process of sensemaking “ordinizes” events
in a way that robs them of their emotional power.
The easier is it for people to make sense of an experience, the more quickly
they return to the baseline. With respect to negative events in particular, psy-
chologists have posited a psychological immune system that protects individuals
from an overdose of gloom53— analogous to the well-documented processes of
physical homeostasis that cause most life forms and complex systems to main-
tain their internal equilibrium. In addition to other adaptive mechanisms, ratio-
nalization can often play an important role. The easier it is to rationalize a bad
experience (breaking up with a romantic partner, not getting a job), the easier it
is to recover.
We tend to mitigate regret over an outcome by reconstruing our responsibil-
ity for its cause. As between blaming oneself for missing a plane by a few min-
utes (“if only I had left the house earlier”) and tending to cast the blame on
others (“if only they hadn’t shut the doors sooner than the departure time”),
most people opt for the latter.54 We also reconstrue our intentions and desires,
for example, responding to a job rejection by thinking “that wouldn’t have been
a good firm to work for anyway.” Indeed, the anticipation of regret may lead to
anticipatory reconstrual: “I don’t know how well my interview went from their
point of view, but I’ve got to say that I wasn’t particularly impressed with
them.”55
People are in principle capable of pursuing their ends—whatever they may be—
in a rational manner. We will tentatively define a rational decision-making pro-
cess as one that tends to achieve the decision maker’s objectives—that is, the
quality of the process is ultimately measured in terms of its consequences.58 A
good process is one that works (most of the time). This seemingly tautological
premise is more complicated that it appears.
First, even the best possible decision-making processes do not always lead
to the desired consequences. For example, a decision maker may lack the
56. See generally Eddie Harmon-Jones and Judson Mills, eds., Cognitive
Dissonance: Progress on a Pivotal Theory in Social Psychology (American
Psychological Association (APA), 1999).
57. The literature is summarized on Hazel Markus’s home page, http://www-psych.
stanford.edu/∼hmarkus/
58. Deborah Frisch and Robert T. Clemen, Beyond Expected Utility: Rethinking
Behavioral Decision Research, 116 Psychological Bulletin 46 (1994).
choices, consequences, and trade-offs 367
59. These include money, tangible goods, physiological and psychological capacities,
social relationships. See Reid Hastie and Robyn M. Dawes, Rational Choice in an
Uncertain World 18 (Thousand Oaks, CA: Sage Publications, 2001).
60. Paul Slovic, Melissa L. Finucane, Ellen Peters, and Donald G. MacGregor, Rational
Actors or Rational Fools? Implications of the Affect Heuristic for Behavioral Economics 31
Journal of Socio-Economics, 329–42 (2002).
368 problem solving, decision making, and professional judgment
are tagged to varying degrees with affect. In the process of making a judg-
ment or decision, people consult or refer to an “affect pool” containing all the
positive and negative tags consciously or unconsciously associated with the
representations. Just as imaginability, memorability, and similarity serve as
cues for probability judgments (e.g., the availability and representativeness
heuristics), affect may serve as a cue for many important judgments. Using
an overall, readily available affective impression can be far easier—more
efficient—than weighing the pros and cons or retrieving from memory many
relevant examples, especially when the required judgment or decision is
complex or mental resources are limited. This characterization of a mental
short-cut leads to labeling the use of affect a “heuristic.” . . .
Like other heuristics, the affect heuristic can conduce to good decision making,
but it also has the potential to distort the process. Whether intuition and affect
play a constructive role in decision making depends on the nature and context of
the decision at hand.61
61. See Kenneth R. Hammond, Human Judgment and Social Policy: Irreducible
Uncertainty, Inevitable Error, Unavoidable Injustice 60 (New York: Oxford
University Press 1996).
62. These roughly map on to the correspondence and coherence theories of truth in
philosophy. See, e.g., “The Stanford Encyclopedia of Philosophy,” http://plato.stanford.
edu/entries/truth-coherence/.
choices, consequences, and trade-offs 369
64. Philip E. Tetlock, Coping with Trade-offs: Psychological Constraints and Political
Implications, in Political Reasoning and Choice Elements of Reason: Cognition, Choice,
and the Bounds of Rationality (Arthur Lupia, Matthew D. McCubbins, and Samuel L.
Popkin eds., Berkeley: University of California Press; Cambridge, UK: Cambridge University
Press, 2001); Alan P. Fiske and Philip E. Tetlock, Taboo Tradeoffs: Constitutive Prerequisites for
Political and Social Life, in Political Psychology: Cultural and Cross Cultural
Foundations (Stanley A. Renshon and John Duckitt eds., New York: NYU Press 2000).
65. Richard L. Revesz And Michael A. Livermore, Retaking Rationality: How
Cost-Benefit Analysis Can Better Protect the Environment and Our Health
78–79 (New York, Oxford University Press 2008) [hereafter, Retaking Rationality].
66. For a sustained critique, arguing that cost-benefit analysis is fundamentally flawed,
see Frank Ackerman and Lisa Heinzerling, Priceless: On knowing the Price of
Everything and the Value of Nothing (New York: The New Press, 2004); See also
Steven Kelman, Cost-Benefit Analysis: An Ethical Critique, 5 J. Gov’t & Soc. Reg. 33 (1981).
67. Retaking Rationality 13.
68. On the national and international level, the issue arises with respect to whether to
subject the perpetrators of massive crimes against citizens to criminal punishment or to
choices, consequences, and trade-offs 371
Scarcity is a simple fact of the human condition. To have more of one good
thing, we must settle for less of another. Claiming that different values are
incommensurable simply hinders clear thinking about difficult tradeoffs.
Notwithstanding their public pronouncements about incommensurability,
even the fiercest critics of cost-benefit analysis cannot escape such tradeoffs.
For example, they do not . . . get their brakes checked every morning. The
reason, presumably, is not that . . . auto safety does not matter, but that they
have more pressing uses of their time. Like the rest of us, they are forced to
make the best accommodations they can between competing values.69
Policy makers (and others) use various strategies to avoid explicit trade-offs
between constitutively incommensurable values.70 When they cannot avoid a
conscious or public trade-off, they engage in buck-passing, procrastination,
and obfuscation.71 And when responsibility for a decision cannot be avoided,
they tend to “spread the alternatives,” “playing down the strengths of the to-be-
slighted value and playing up the strengths of the to-be selected value.”72
Albert Einstein suggested: “Make everything as simple as possible, but not sim-
pler.” Thus, a decision maker should consider all those but only those factors that
could significantly affect her objectives or interests. In identifying the criteria rel-
evant to siting the wastewater treatment plant, Christine Lamm has tried to be
complete. Sometimes, however, she may lump a number of criteria together. For
example, environmental impact comprises a number of factors, but it may not be
worth her time and energy to consider each one separately, rather than make an
overall assessment.
***
choices, consequences, and trade-offs 373
This concludes our first look at the axioms of subjective expected utility theory.
We consider some axioms relevant to decision making under risk in Section
15.1. Before moving on, we note two points about SEU.
First, SEU provides a generalized framework for rational choice. But evolu-
tionary psychologists such as Leda Cosmides, John Tooby, and Gerd Gigerenzer
have argued that much individual decision making is not based on general prin-
ciples of rationality, but is modularized to address specific problems in particu-
lar domains. They use the term ecological rationality to refer to the success of a
process in producing outcomes in a particular environment or context.73 Granted
that individuals often depart from the norms of SEU, however, there is consider-
able doubt whether decision making is as highly contextual as these social scien-
tists claim, let alone to what extent modularized decision making leads to
successful outcomes.74 In any event, lawyers and policy makers are often called
upon to make decisions that cut across many domains.
Second, most applications of expected utility theory focus on outcomes. But
earlier in the chapter we considered how the very process of choosing can produce
gains or losses of utility for the decision maker. This phenomenon can confound
decision makers, or at least confound understanding whether particular choices
are rational from the decision maker’s point of view. We will return to this matter
in the next several chapters.
73. See Leda Cosmides and John Tooby, Origins of Domain Specificity: The Evolution of
Functional Organization, in Mapping the Mind: Domain Specificity in Cognition
and Culture 85, 89–90 (Lawrence A. Hirschfeld and Susan A. Gelman eds., Cambridge,
UK: Cambridge University Press, 1994); Gerd Gigerenzer, Peter M. Todd and ABC
Research Group, Simple Heuristics That Make Us Smart (New York: Oxford University
Press, 1999).
74. Jerry Fodor, The Mind Doesn’t Work That Way: The Scope and Limits of
Computational Psychology 58–61 (Cambridge, MA: The MIT Press, 2000); Mark G.
Kelman, The Heuristics Debate: Its Nature and Its Implications for Law and
Policy (Oxford University Press, forthcoming 2010).
75. See Office of Management and Budget (OMB), “Economic Analysis of Federal
Regulations Under Executive Order 12866,” Jan. 11, 1996, http://www.whitehouse.gov/
374 problem solving, decision making, and professional judgment
President Bill Clinton in 1993, requires federal agencies to “assess all costs and
benefits of available regulatory alternatives, including the alternative of not
regulating.”
The most difficult and contested applications of CBA involve risk, a matter
that we discuss in Chapter 15. Even when risk is not a major factor, however,
CBA can present difficulties in monetizing benefits and in making distributional
decisions (for example, where the benefits flow partly to an individual benefi-
ciary of a program and partly to taxpayers). And even when the calculation is
reasonably clear, there may be reasons not to follow the logic of CBA: For better
or worse, politics, ideology, and issues of fairness and rights may trump CBA.
But CBA almost always provides a useful norm or reference point for decision
making.
The longitudinal analysis of the Perry Preschool program in Michigan pro-
vide a classic example of CBA. This precursor of Head Start had low student-
teacher ratios; all the teachers were certified in both early childhood and special
education, and they made weekly visits to every family.76 The program produced
long-term gains in school attendance, academic achievement, graduation rates,
and earnings.
In constant discounted dollars, the economic return to society of the program
was more than $17 per dollar invested ($258,888 on an investment of $15,166
per student). Seventy-five percent of that return accrued to the general public
($12.90 per dollar invested), and 25 percent to each participant ($4.17 per
dollar invested). While net benefits derive from many sources, such as
increased earnings and taxes paid, with Perry Preschool students having sig-
nificantly fewer lifetime arrests and fewer arrests for violent crimes than stu-
dents not in the program, a significant share of the public benefits come from
the avoidance of these significant costs to society.77
Thus, even when limiting the benefits to those returned to taxpayers, the pro-
gram had a benefit-cost ratio of almost 13 to 1. Economists at the Federal Reserve
Bank in Minneapolis estimated that the Perry Preschool program generated a 16
percent rate of return on investment, of which 12 percent was a public return.78
79. See Paul Brest, Hal Harvey, & Kelvin Low, Calculated Impact, Stanford Social
Innovation Review, Winter 2009.
80. For more on discount rates, see Section 13.6.1.
376 problem solving, decision making, and professional judgment
value of the benefits based on the probability of success; it discounts its own role
in generating these benefits by 50 percent to account for the fact that other
donors, both private and public, support Bob’s Jobs. Thus, Robin Hood attri-
butes $4.6 million of the program’s benefits to its philanthropic investment.
Calculating costs. What about the costs? The grant to Bob’s Jobs was $200,000.81
Dividing $4.6 million by $200,000 results in a net benefit of $23 per $1 invested.
That is, for each dollar that Robin Hood spends on Bob’s Jobs, trainees and their
families gain $23—a pretty good return on investment.
Robin Hood has made similar calculations for its other poverty programs,
including a program to help people manage budgets, bank accounts, and loans
(net benefit = $1.90) and a clinic that deals with asthma, hepatitis, and cancer
(net benefit = $12).
Although Robin Hood considers cost-benefit analysis in its grant-making, it
does not rely on this measure exclusively. The foundation recognizes that the
metrics are imprecise, with each CBA calculation depending on complex and
uncertain empirical assumptions. Thus Robin Hood continues to test its metrics
against the informed intuitions of program officers and experts in the field. At
the same time, the foundation presses its staff to justify their intuitions against
the numbers generated by analysis.
81. Robin Hood does not include the administrative costs of making the grant, both
because these costs are not large and because they do not differ significantly from those
of the other poverty-fighting programs to which it will compare Bob’s Jobs.
choices, consequences, and trade-offs 377
$16.3 million.83 Nonetheless, some scholars advocate, on moral grounds, the use
of a universal VSL for all peoples, regardless of their economic status or willing-
ness to pay. On the other hand, others, including Cass Sunstein, assert that
policy makers should establish specific, local VSLs for communities and coun-
tries. Sunstein argues that VSLs should be assessed within the population where
regulations force the beneficiaries to bear the costs. If an inflated VSL is used,
poorer citizens may be forced to “spend” their money to reduce a workplace risk
when they would prefer to put that money to other uses.84
12.3.3b Value of Statistical Life Years (VSLY) VSL treats the lives of the
very young and very old as equally valuable. Under the VSLY approach, benefits
to individuals are assessed based on their estimated remaining years of life.
VSLY values are derived from VSLs drawn from age-specific populations.85
Michael J. Moore and W. Kip Viscusi, proponents of this approach, note that
“in the case of fatalities, a young person loses a much greater amount of lifetime
utility than does an older person,” and thus assert that it doesn’t make sense
to use the same value for an elderly person with a remaining five-year life
expectancy as for a 25-year-old.86 Critics of VSLY term it the “senior death
discount.”
Revesz uses ozone emission standards to demonstrate the difference between
the VSL and VLSY approaches.87 Assume for the sake of argument that elderly
individuals with existing respiratory problems would be the most likely to bene-
fit from stricter ozone regulations, and that these individuals have an average life
expectancy of five years. Under a VSLY calculation, the benefit of a stricter regu-
lation, using a $180,000 life-year value, would be $900,000. By assessing benefits
at $900,000 per individual instead of at the $6.3 million VSL, the total estimated
regulatory benefit is reduced by 85 percent.
The VSLY approach assumes a decreasing relationship between willingness
to pay and age that is proportional to remaining life expectancy. This empirical
assumption remains contested despite considerable research.88 Some studies
83. Id.135.
84. Cass R. Sunstein, Laws of Fear: Beyond the Precautionary Principle 170
(New York: Cambridge University Press 2005). This distributional problem is further con-
founded when the costs for this population are subsidized or covered by a grant: in this
case, people “will be helped even if risk reduction is based on an excessive VSL.”
85. See Joseph E. Aldy and W. Kip Viscusi, Adjusting the Value of Statistical Life for Age
and Cohort Effects, April 2006, RFF DP 06–19.
86. Michael J. Moore and W. Kip Viscusi, The Quantity-Adjusted Value of Life, 26 Econ.
Inquiry 269 (1988).
87. Retaking Rationality 78–79.
88. Id. 81–82.
choices, consequences, and trade-offs 379
suggest that WTP peaks around age fifty;89 but elderly citizens may be willing to
pay enormously for risk reductions that prolong life. Because both VSL and
VSLY “may change with age, life expectancy, anticipated future health, income,
and other factors,”90 the U.S. Office of Management and Budget has encouraged
federal agencies to conduct analysis using both approaches.91
12.2.2 c Quality-Adjusted Life Years (QALYs) QALYs incorporate a coeffi-
cient that measures “quality of life” and can be used to assess the efficiency of
health care policies and procedures. Consider an example by Dr. Fritz Allhoff:
“If treating a patient would lead to a life expectancy of 20 years with a high
quality of life (e.g., hypertension, which only requires daily pills and rarely
manifests negative symptoms), coverage for this patient’s treatment should
presumably have priority over treatment for a patient who would have a life
expectancy of 20 years with a comparatively low quality of life (e.g., diabetes,
which requires daily injections and often manifests symptoms such as neu-
ropathy, blindness, etc).”92
This “quality of life” coefficient ranges from 0 to1,valuing a healthy year of life
at 1 and a year where one suffers under a disease or disability at some fraction of
1. The QALYs assigned to a particular reduction in mortality, then, are based on
the expected number of years of life gained adjusted by their perceived quality of
life. Proponents of this method argue that this system distributes finite resources
as efficiently and fairly as possible.
The perceived quality-of-life coefficient, which has a large impact on the final
assigned value, is typically determined through contingent valuation, asking
“healthy people, ex ante, to evaluate various health states.”93 But healthy
individuals may fail to take adaptation into account and thus underestimate the
quality of life of disabled persons, leading to disproportionately low QALY
values. Additionally, as Revesz notes, the “empirical work comparing how will-
ingness to pay relates to QALYs confirms there is no simple dollar-per-QALY
89. Thomas J. Kniesner, W. Kip Viscusi, and James P. Ziliak, Like-Cycle Consumption
and the Age-Adjusted Value of Life, 5 Contributions, Economic Analysis and Policy 1524
(2006).
90. James K. Hammitt, Risk in Perspective: Valuing “Lives Saved” vs. “Life-Years Saved,
Harvard Center for Risk Analysis, March 2008, Volume 16, Issue 1.
91. Id.
92. Fritz Allhoff, The Oregon Plan and QALY’s, Virtual Mentor, Policy Forum, February
2005, Volume 7, Number 2.
93. Retaking Rationality 90.
380 problem solving, decision making, and professional judgment
conversion rate.”94 Although valuations of about $50,000 are often used,95 a study
by the Harvard Center for Risks Analysis reveals that willingness to pay per
QALY fluctuates depending on “both the severity and duration of the illness.”96
The State of Oregon was a leader in using QALYs to prioritize treatments
after deciding to make fewer Medicaid-funded services available to a larger
number of people. In 1989, a commission determined “quality-of-life” coeffi-
cients based on numerous community meetings and phone surveys.97 The com-
mission assigned QALYs to 709 procedures, providing coverage only for those
ranked higher than 587. More specifically, “treatments that prevented death with
a full chance of recovery were ranked first, maternity care was ranked second,
treatments that prevented death but did not guarantee full recovery were ranked
third, and treatments that led to minimal or no improvements in quality of life
were ranked last. For example, “Diagnosis: severe or moderate head injury,
hematoma or edema with loss of consciousness; Treatment: medical and surgi-
cal treatment’ was ranked at the top of the list and ‘Diagnosis: mental disorders
with no effective treatment’ . . . was ranked near the bottom.”98
Inevitably, the rankings produced some seemingly arbitrary and politically
vulnerable choices. For example, “acute headaches” were ranked higher than
some treatments for AIDS or cystic fibrosis.99 The plan was originally rejected
by Health and Human Services, but approved in 1993 under the Clinton
administration after some changes in methodologies and safeguards against
discrimination.100
Britain’s National Health Service, which provides 95 percent of the nation’s
care, also uses QALYs to ration health care. The National Institute for Health
and Clinical Excellence (NIHCE), which approves treatments and drugs for the
Health Service, determined that, in effect, six months of good quality life is worth
about $22,750. It automatically approves drugs that provide six months of good
quality life for $15,150 or less.101 Dr. Andreas Seiter, a senior health specialist at
the World Bank, notes that “all the middle-income countries—in Eastern Europe,
Central and South America, the Middle East and all over Asia—are aware of
NIHCE and are thinking about setting up something similar.”102
In Metropolitan Edison Co. v. People Against Nuclear Energy, 460 U.S. 766
(1983), involving the operation of a nuclear power plant near a residential
area the Nuclear Regulatory Commission, the Court was asked to review
the Commission’s assessment of adverse psychological affects under the
National Environmental Policy Act. Justice William Rehnquist noted that
“any effect on psychological health is an effect on health,” but he won-
dered how one could distinguish between “genuine” adverse psychologi-
cal effects that are caused directly by a policy from claims that are rooted
in discontent.
101. Michael D. Rawlins and Anthony J. Culyer, National Institute for Clinical Excellence
and Its Value Judgments, BMJ 2004 Jul 24; 329 (7459):227–9. This has “led many companies
to offer the British discounts unavailable almost anywhere else.” Gardiner Harris, The
Evidence Gap: British Balance Benefit vs. Cost of Latest Drugs, New York Times, Dec. 2, 2008.
102. See id.
103. OMB, Circular A-4, “Regulatory Analysis: Memorandum to the Heads of Executive
Agencies and Establishments” 3 (Sept. 17, 2003).
382 problem solving, decision making, and professional judgment
are often not assigned valuations (or “shadow prices”).104 She offers these catego-
ries with respect to social programs targeting youth:
• Outcomes rarely or never monetized: child/youth behavioral/emotional out-
comes (e.g., behavior problems, school suspensions/expulsions, mental
health outcomes), child/youth cognitive outcomes (e.g., IQ scores), K–12
grades, school attendance, school engagement, general health status of
children/youth, contraceptive use, adult parenting measures, marriage and
divorce, adult mental health, adult reproductive health, employment,
income, and poverty status.
• Outcomes without clear consensus on monetization: achievement test scores,
high school graduation, college attendance, teen pregnancy, tobacco use,
alcohol abuse, illicit drug use, crime and delinquency, and child abuse and
neglect.
• Outcomes with more-established methodology for monetization: special educa-
tion use, grade retention, transfer payments, and other means-tested
programs
Even if one can assign values to such benefits, questions of judgment remain
about which to include in CBA. These are appropriately questions not for eco-
nomic experts, but for policy makers.105
One study indicated that the regulation of carbon monoxide emissions from
motor vehicles to improve air quality had an enormous ancillary benefit of reduc-
ing cancer risks, thus saving an average of 25,000 lives per year—whereas other
health benefits had been estimated only at 212 to 515 lives annually.108 And
health or safety regulations may have what Revesz calls an “attentiveness effect,”
inducing people to become more conscious of safety issues beyond what is
mandated.109
Policy makers often only consider unintended costs. However, the OMB rec-
ommends that agencies account for ancillary benefits of regulations as well,110
and Revesz argues that a complete cost-benefit analysis requires as much.111 In a
2007 decision, the U.S. Court of Appeals for the Ninth Circuit invalidated the
National Highway Traffic Safety Administration (NHTSA)’s fuel economy stan-
dards for light trucks, noting that while NHTSA had accounted for the negative
consequences of the policy on employment and automobile sales, it failed to
include the ancillary beneficial effects on climate change. The court said that
“NHTSA could not put a thumb on the scale by undervaluing the benefits.”112
Office of Water, Constructed Wetlands for Wastewater Treatment and Wildlife Habitat
(91993), EPA 832-R-93-005, available at http://www.epa.gov/owow/wetlands/construc/.
108. See M. Shelef, Unanticipated Benefits of Automotive Emission Control: Reduction in
Fatalities by Motor Vehicle Exhaust Gas, 146/147 Sci. Total Envtl. 93 (1994).
109. Retaking Rationality 175.
110. OMB, Circular A-4, “Regulatory Analysis: Memorandum to the Heads of Executive
Agencies and Establishments” 3 (Sept. 17, 2003).
111. Retaking Rationality 145.
112. Center for Biological Diversity v. National Highway Traffic Safety Administration
(NHTSA), No. 06-71891, 13,871 (9th Cir. Nov. 15, 2007).
113. Cass R. Sunstein, Laws of Fear: Beyond the Precautionary Principle, supra.
114. Frank Ackerman and Lisa Heinzerling, supra. See also Mark Sagoff, The
Economy of the Earth: Philosophy, Law, and the Environment (Cambridge Studies
in Performance Practice).
384 problem solving, decision making, and professional judgment
at stake.”115 Consider James Hammit’s March 2003 letter to the editor of the
New York Times:116
The Office of Management and Budget’s interest in applying cost-benefit
analysis to homeland security measures is laudable. Even if precise quantifi-
cation of the benefits to security and harms to civil liberties is impossible,
analytic consideration of the trade-offs is wise.
As John Graham, director of regulatory affairs at O.M.B., says, “Simply
identifying some of these costs will help understand them and get people to
think about alternatives that might reduce those costs.”
Cost-benefit analysis should also have been used to evaluate the potential
war in Iraq. The benefits of ousting Saddam Hussein should be measured
against the costs of casualties, waging war, reconstruction, retaliatory terror-
ism and increased anti-Americanism. Such analysis would prompt us “to
think about alternatives that might reduce those costs.”
Neoclassical economics and the axioms of expected utility theory posit a model
of the individual as homo economicus. In Gary Becker’s words, all human behav-
ior can be viewed as involving participants who maximize their utility from a
stable set of preferences and accumulate an optimal amount of information and
other inputs in a variety of markets.1
The next several chapters inquire whether and when decision makers deviate
from this model. This chapter begins with a compilation of characteristics of a
decision that may lead to suboptimal outcomes, and then discusses how the
choice of decision-making strategies can affect outcomes. The bulk of the chap-
ter focuses on people’s relationships with their own future selves—their beliefs
about how today’s decisions will affect their futures, and limits on their capacity
to plan and implement plans to improve their future well-being. Chapter 14 con-
siders how the outcomes of decisions can be affected by the ways that choices are
framed. Chapter 15 considers the effects of framing and other matters when
decisions are made in conditions of risk or uncertainty. Chapter 16 continues the
discussion of risk in situations of high affect.
suboptimal decisions. This raises the preliminary question of how one would
know whether a decision maker had made a suboptimal decision. The following
characteristics do not necessarily lead to poor decisions, but they all reflect poten-
tial hazards for optimal decision making:
1. The decision is based on incorrect data or the incorrect analysis of data—
whether the result of biases (e.g., the availability or representativeness heu-
ristic), poor predictions of one’s own future desires or what it will take to
achieve them, or simple miscalculation.
2. The decision violates one of the axioms of expected utility, or rational
choice. (See Section 12.2.2). For example, it violates the principle of domi-
nance or transitivity.
3. The decision maker poorly integrates utilities from the decision-making
process with utilities from the outcome of the process—for example, giving
excessive weight to the utility of avoiding confronting a stressful choice
compared to the enduring consequences of making a poor choice.
4. The decision is sensitive to the way the issue is framed and is made in a con-
text where the framing is highly variable or manipulable. (See Chapter 14.)
5. The decision is sensitive to affect (i.e., emotion) and is made in a context
where affect is highly variable or manipulable, or where present affect is
likely to be very different from affect when experiencing the consequences
of the decision. (See Chapter 16.)
6. The decision maker was subjected to undue social influence. (We postpone
consideration of this subject to Chapter 17.)
We will examine some of these phenomena and other factors that can affect
the decision-making process, often for the worse.
2. The French word bricolage means “doing odd jobs.” Borrowing from Claude Levi-
Strauss’s use of the term, a bricoleur is someone who improvises and uses any means or
materials that happen to be lying around in order to tackle a task. Claude Levi-Strauss,
The Savage Mind 19 (Chicago: The University of Chicago Press, 1966).
complexities of decision making 387
then, that decision makers faced with taboo trade-offs often use choice-avoidance
strategies like putting off the decision indefinitely or delegating the responsibil-
ity of deciding to someone else.8 The emotional difficulties of making taboo
trade-offs—as well as the risk that others will condemn the person who consid-
ers them—can thus shape the strategy that decision makers select.
For example, participants in a study were asked to assume that they were
music majors purchasing a music dictionary in a used book store. In the joint-
evaluation condition, they were told that there were two music dictionaries
for sale:
• one dictionary was “like new” and had 10,000 entries;
• the other had a torn cover and had 20,000 entries.
On average, the participants were willing to pay $19 for the first and $27 for
the second. But when participants were told that there was only one music dic-
tionary in the store, those presented with only the first dictionary were willing to
pay $24 and those presented with only the second were willing to pay $20: a torn
cover is easy to evaluate without comparison, while the number of entries is dif-
ficult to evaluate. This phenomenon of preference reversals seems robust in a
variety of different contexts. Table 13.1 provides some examples where people’s
preferences were reversed in separate and joint evaluations.1011
table 13.1 easy and hard to evaluate attributes
Computer programmer 4.9 GPA vs. 3.0 GPA Number of computer programs written
job applicants10 in past 2 years: 10 vs. 70
Eye surgeons for laser Harvard MD Degree Number of similar successful
surgery11 vs. Iowa MD Degree procedures: 80 vs. 300
Especially when the values of attributes are not readily monetizable, people
are often influenced more by proportions than absolute values. For example,
Paul Slovic and his colleagues found that, in separate evaluations, people sup-
ported an airport-safety measure expected to save 98 percent of 150 lives at risk
more strongly than a measure expected to save 150 lives (period).12 They explain:
“Saving 150 lives is diffusely good, hence only weakly evaluable, whereas saving
98% of something is clearly very good because it is so close to the upper bound
on the percentage scale, and hence is readily evaluable and highly weighted
in the support judgment.” They report that “subsequent reduction of the
10. Id.
11. Brian J. Zikmund-Fisher, Angela Fagerlin, and Peter A. Ubel, Is 28% Good or Bad?
Evaluability and Preference Reversals in Health Care Decisions, 24 Medical Decis. Making
142 (2004). Evaluation was on a scale of 0 (extremely bad choice) to 10 (extremely good
choice) rather than WTP.
12. Paul Slovic, Melissa Finucane, Ellen Peters, and Donald G. MacGregor, Rational
Actors Or Rational Fools: Implications Of The Affect Heuristic For Behavioral Economics, 31
Journal of Socio-Economics 329 (2002).
390 problem solving, decision making, and professional judgment
percentage of 150 lives that would be saved to 95%, 90%, and 85% led to reduced
support for the safety measure but each of these percentage conditions still gar-
nered a higher mean level of support than did the save 150 lives condition.”
Suppose that, in the 1990s, there were two camps for Rwandan refugees in
neighboring Zaire. One camp holds 250,000 refugees and the other 11,000.
Cholera is rampant in both camps. A nongovernmental organization can pro-
vide enough clean water to save the lives of a total of 4,500 refugees suffering
from cholera in one or the other camp. Which one would you chose? From a
moral or policy point of view, the number of refugees in the camps makes no
difference,13 and 42 percent of the participants in a study said so. However, 44
percent would have directed the resources to the smaller camp, where a larger
proportion of lives would have been saved.14 Slovic observes that number of lives
saved, standing alone, appears to be poorly evaluable, but becomes clearly evalu-
able and important in side-by-side comparisons.
13. Actually, one could imagine variables, such as preventing riots or contagion, that would
affect the policy choice. But such factors did not play a role in the participants’ decisions.
14. In another set of questions focusing on only one camp, one group of participants
was told that prior aid had met 5 percent of the camp’s water needs, and that the new aid
would bring the total to 7 percent; another group was told that prior aid met 95 perecnt of
the needs and the new aid would bring the total to 97 percent. Most participants would
have directed aid to the latter.
15. This section draws heavily on Eldar Shafir, Itamar Simonson, and Amos Tversky,
Reason-based Choice, 49 Cognition 11 (1993).
complexities of decision making 391
Now suppose that Adele delegates the decision to Barbara, who, rather than
using a subjective linear model, looks at the files one by one, intuitively sorting
them in “yes” or “no” piles. Do you think she will reach the same result as Adele
did? Consider the role of the availability heuristic, discussed in Section 9.6.
Suppose that Adele gives the identical files to Clara and David and asks Clara
to produce a list of the names of candidates to invite to interview and David to
produce a list of those who should be sent a rejection letter. Who is likely to have
more candidates in the running? Experiments suggest that David will leave more
candidates in the running than Clara. Resumes and application letters tend to
favor a candidate’s strong points rather than weaknesses. Clara will only select
candidates who stand out among the rest, while David will find fewer negative
characteristics on which to reject the candidates.16
Along these lines, Eldar Shafir and his colleagues conducted an experiment
in which participants were given the following scenario: Imagine that you serve
on the jury of an only-child sole-custody case following a relatively messy divorce.
The facts of the case are complicated by ambiguous economic, social, and emo-
tion considerations, and you decide to base your decision entirely on the charac-
teristics shown in Table 13.2.
Parent A Parent B
Notice that Parent A is average in all respects, while Parent B has some strong
and weak points.
When asked to whom they would award sole custody, the majority of partici-
pants (64 percent) chose Parent B to receive custody. Yet when a different group
of participants was asked to whom they would deny sole custody, the majority (55
percent) chose the opposite outcome and denied custody to Parent B.
Based on this and similar experiments, the researchers conclude that
when people are basing decisions on reasons, “the positive features of options
(their pros) will loom larger when choosing, whereas the negative features
16. Vandra L. Huber, Margaret A. Neale, and Gregory B. Northcraft, Decision Bias and
Personnel Selection Strategies, 40 Organizational Behavior and Human Decision
Processes 137 (1987).
392 problem solving, decision making, and professional judgment
of options (their cons) will be weighted more heavily when rejecting.”17 But this
phenomenon violates the axiom of invariance (Section 12.2.2.b): the way in which
the question is framed should not, by itself, affect the outcome.
13.3.2.b Preference Reversals Between Prices and Choices Suppose you are
presented with two gambles:
• Gamble A: 0.9 chance to win $8
• Gamble B: 0.2 chance to win $30
Suppose that you are asked to choose between the two gambles? Now suppose
you are asked to price each gamble—e.g., suppose that you own tickets entitling
you to each, at what price would you would be willing to sell the tickets? On aver-
age, people asked to choose select gamble A, which has a much higher probabil-
ity of winning. But when asked to price each gamble, they demand more for B.18
There are two explanations for this violation of invariance: the compatibility
effect and the prominence effect.19 Under the compatibility effect, when subjects
are thinking in terms of dollars, as they do when asked to set a price, they are
more likely to evaluate their response in terms of dollars. Under the prominence
effect, choice invokes reasoning in a way that pricing doesn’t—people feel a need
to explain the choice at least to themselves—and the gamble with the higher
probability is easier to explain.
Such preference reversals have been manifested in other situations,
including people’s valuations of reducing the risks from chemical products,
improvements in consumer goods, and improvements in the environment.20
We will say a word more about this in a discussion of contingent valuation in
Section 14.6.
13.3.2.c The Effect of Giving Reasons on the Quality of Choices and
Satisfaction with Them Using reasons as a strategy can make decision making
easier and more transparent. However, excessive reliance on reasons and
deliberation has its own drawbacks. A number of experiments by Timothy
Wilson and his colleagues suggest that giving reasons for one’s preferences
17. Eldar Shafir, Choosing versus Rejecting: Why Some Options Are Both Better and Worse
than Others, 21 Memory & Cognition 546–56 (1993).
18. Amos Tversky and Richard H. Thaler, Anomalies: Preference Reversals, 4 Journal
of Economic Perspectives 193–205 (1990).
19. Paul Slovic, Dale Griffin, and Amos Tversky, Compatibility Effects in Judgment
and Choice, in Insights in Decision Making: A Tribute to Hillel J. Einhorn 5–27
(Robin M. Hogarth ed., Chicago, IL: University of Chicago Press, 1991); Amos Tversky,
Schmuel Sattath, and Paul Slovic, Contingent Weighting in Judgment and Choice, 85
Psychological Review 371–84 (1988).
20. Julie R. Irwin, Paul Slovic, Sarah Lichtenstein, and Gary H. McClelland, Preference
Reversals and the Measurement of Environmental Values, 6 Journal of Risk and
Uncertainly 5–18 (1993).
complexities of decision making 393
Similar experiments have been conducted with respect to the choice of jams,
college courses, and other items—with similar results. With respect to the effects
of reason giving, Timothy Wilson and Jonathan Schooler write:22
Forming preferences is akin to riding a bicycle; we can do it easily but cannot
easily explain how. Just as automatic behaviors can be disrupted when people
analyze and decompose them, so can preferences and decisions be disrupted
when people reflect about the reasons for their feelings. We suggest that this
21. See Timothy D. Wilson and Jonathan W. Schooler, Thinking Too Much: Introspection
Can Reduce the Quality of Preferences and Decisions, Journal of Personality and Social
Psychology 181 (1991); Timothy D. Wilson et al., Introspecting About Reasons Can Reduce
Post-Choice Satisfaction, 19 Personality and Social Psychology Bulletin 331 (1993).
22. Timothy Wilson and Jonathan Schooler, Thinking Too Much: Introspection Can
Reduce the Quality of Preferences and Decisions, 60 J Pers. Soc. Psychol. 181–192 (1991).
394 problem solving, decision making, and professional judgment
can occur as follows. First, people are often unaware of exactly why they feel
the way they do about an attitude or an object. When they reflect about their
reasons, they thus focus on explanations that are salient and plausible. The
problem is that what seems like a plausible cause and what actually deter-
mines people’s reactions are not always the same thing. As a result, when
asked why they feel the way they do, people focus on attributes that seem like
plausible reasons for liking or disliking the stimulus, even if these attributes
have no actual effect on their evaluations.
. . . [U]nder some circumstances . . . people will focus on reasons that
imply a different attitude than they held before and will adopt the attitude
implied by these reasons. . . . [P]eople often do not have a well-articulated,
accessible attitude and thus do not start out with the bias to find only those
reasons that are consistent with an initial reaction. They conduct a broader
search for reasons, focusing on factors that are plausible and easy to verbalize
even if they conflict with how they felt originally.
[W]e suggest that reflecting about reasons will change people’s attitudes
when their initial attitude is relatively inaccessible and the reasons that are
salient and plausible happen to have a different valence than people’s initial
attitude.
In short, because what you can readily articulate doesn’t necessarily reflect
what you deem most important, articulating reasons for your preferences can
lead to choices inconsistent with your actual preferences. Does articulating rea-
sons interfere with judgments more generally? Research by John McMakin and
Paul Slovic suggests that it depends on the kind of judgment being made. In one
study, McMakin and Slovic asked participants to estimate how their peers would
react to a series of advertisements. As with judging one’s own preference for
posters, this task requires a holistic, intuitive approach. As in Wilson and
Schooler’s study, participants who were asked to articulate reasons for their esti-
mate provided less accurate estimates than those not asked to articulate reasons.
In a second study, however, McMakin and Slovic had participants make judg-
ments that required analytic rather than intuitive thinking, such as estimating
the length of the Amazon river. For this task, generating reasons increased the
accuracy of estimates.
Why would the first judgment task have been better suited for intuitive think-
ing, while the second demand more analytic thinking? The complexity and
ambiguity of the task may provide part of the answer. Posters and advertise-
ments vary in countless ways that affect preferences; identifying the relevant
dimensions and keeping all in mind simultaneously is a complex task with
ambiguous criteria. By contrast, estimating the length of the Amazon is a prob-
lem that, while difficult, does not involve multiple, ambiguous dimensions.
Is it possible that holistic, intuitive thinking may lead to better judgments for
more complex tasks while analytic, reasons-based thinking may work better for
complexities of decision making 395
less complex tasks? Some evidence for this proposition comes from work by Ap
Dijksterhuis and his colleagues.23 In one study, participants tried to choose the
best car from a selection of descriptions that varied on either twelve dimensions
(a relatively complex task) or on only four dimensions (a relatively simple task).
The descriptions were crafted so that, in each condition, one car was objectively
superior to the others. The researchers instructed half of the participants to spend
four minutes thinking carefully about their judgment, and the other half to spend
four minutes doing a task that distracted them from thinking about the matter.
Results showed that participants told to think carefully made more accurate judg-
ments about the cars when the task was simple than when it was complex. By
contrast, participants who were distracted from generating reasons for their judg-
ment were more accurate when the task was complex than when it was simple.
It would be a mistake to conclude that one should always eschew reasoned
analysis of complex decisions in favor of intuition. First, one might question the
generalizability of these findings to the sorts of consequential decisions facing
policy makers, who often have longer to deliberate than the participants in these
studies. More fundamentally, as we see throughout this book, “gut feelings” can
be biased by emotions unrelated to the decision at hand, or by prejudices and
stereotypes that can only be mitigated by deliberate reflection.
Still, the research reviewed in this section indicates that reason-based decision
making does not always lead to optimal judgments. As Eldar Shafir’s study about
custody awards demonstrates, the way a question is framed can affect the kinds
of reasons that one generates. Additionally, when we grapple with complex prob-
lems, the reasons we generate may indicate more about what is easy to articulate
than about what is most relevant to the judgment at hand. Without ignoring
the ability of reasoned analysis to enhance judgment, then, we should exercise
caution in over-relying on those reasons that come most easily to mind.
There are two tragedies in life. One is not to get your heart’s desire; the other is to get it.
—George Bernard Shaw
Vronsky meanwhile, although what he had so long desired had come to pass, was not
altogether happy. He soon felt the fulfillment of his desires gave him only one grain
of the mountain of happiness he had expected. This fulfillment showed him the
eternal error men make in imagining that their happiness depends on the realiza-
tion of their desires.
—Leo Tolstoy, Anna Karenina (1887)
23. Ap Dijksterhuis, Maarten W. Bos, Loran F. Nordgren, and Rick B. van Baaren, On
Making the Right Choice: The Deliberation-Without-Attention Effect, 311 Science 1005 (2006).
396 problem solving, decision making, and professional judgment
It seems that the linear model used by Christine Lamm in the wastewater
treatment plant problem might be able to overcome some of the shortcomings
of the reason-based choice strategy we have just reviewed. However, decision
makers almost never have perfect information relating to a decision—
especially in conditions of risk and uncertainty. In Chapter 15, we will discuss
how to deal with uncertainties about objective matters—for example the weather
or, in Lamm’s case, problems involving acquiring the site near Edenville’s
airport.
Even more difficult to determine are uncertainties about subjective matters—
for example, how the residents of north Edenville will actually experience having
the wastewater treatment plant in their neighborhood.
The following sections explore ways that people may make systematic
errors—akin to the cognitive biases discussed in earlier chapters—in predicting
how they will feel about things in the future. In addition to considering their
relevance to individual decision making, it will be interesting to ask how a policy
maker should take them into account.
Your actual experience of the consequences of the decision is termed the
experienced utility of a decision. At the time you make a decision, however, you
can only predict the decision’s effect on experienced utility. Ideally, your
predicted utility is perfectly congruent with experienced utility. Unfortunately,
as we all know from personal experience, they can diverge. The giant-size box
of popcorn you bought when entering the movie theater no longer seems
so appealing when you’ve eaten half of it. The dress that seemed to make a
daring fashion statement in the store sits in your closet after being worn
only once. While “the ability to predict the intensity of one’s future likes and
dislikes appears to be an essential element of rational decision making,” the task
of predicting one’s future feelings and tastes “turns out to be surprisingly
difficult.”24
We separate these issues into two related parts. The first concerns what Daniel
Gilbert and Timothy Wilson have termed impact bias—mispredictions of how
you will feel in the future about a current occurrence or event, assuming
that your basic preferences have not changed in the interim.25 The second
involves what George Loewenstein and his colleagues have termed projection
bias—mispredictions of your future preferences.26
24. Daniel Kahneman and Jackie Snell, Predicting a Changing Taste: Do People Know
What They Will Like, 5 Journal of Behavioral Decision Making 187 (1992).
25. Daniel T. Gilbert, Elizabeth C. Pinel, Timothy D. Wilson, Stephen J. Blumberg, and
Thalia P. Wheatley, Immune Neglect: A Source of Durability Bias in Affective Forecasting, 75
Journal of Personality and Social Psychology 617–38 (1998).
26. See George Loewenstein and Erick Angner, Predicting and Indulging Changing
Preferences, in George Loewenstein, Daniel Read, and Roy Baumeister eds., Time and
complexities of decision making 397
Legal Note
Your client, a historian, enraged by claims that a critic has made that portions
of the book were plagiarized, wants to sue the critic for defamation. Your tell
your client that he has a fairly good chance of ultimately winning a small
judgment but that, in addition to financial costs, litigants tend to incur heavy
emotional costs during the process, which often outweigh the satisfactions
of victory. However, your client is focused on vindication.
See Jeremy Blumenthal, Law and the Emotions: The Problems of Affective Forecasting,
80 IND. L.J. 155 (2005).
• Assistant professors predicted that they would be less happy during the
first five years after being turned down for tenure, and happier during the
first five years after gaining tenure, than they actually were.
• Dental patients, and especially those who were anxious about their appoint-
ments, overpredicted the amount of pain they would experience.
• People being tested for HIV or for (unwanted) pregnancy overpredicted
misery from a positive result and elation from a negative result.
• People overpredict their happiness if their favored teams or political candi-
dates wins and their unhappiness if they lose.
Most of these predictions involve errors in what Gilbert, Wilson, and their
colleagues have termed affective forecasting. Predictions can err in several
respects.
Although we are usually correct in predicting whether an emotion will have
positive or negative valence, we may oversimplify situations that evoke complex
emotions by focusing on only positive or negative affect—e.g., students looking
forward to graduation with joy and pride, overlooking sadness about leaving
friends and apprehension about their futures.28 We also sometimes mispredict
the specific positive or negative emotions we will experience—e.g., women who
predicted that they would respond with anger to sexually harassing questions in
a job interview tended instead to respond with fear.29
More generally, as illustrated by Figure 13.1,30 we often mispredict the inten-
sity and duration of emotional experiences. The remainder of this subsection
explores the possible causes of problems in impact bias—“the tendency to over-
estimate the enduring impact that future events will have on our emotional
reactions.”31
13.4.1.a Focalism One reason that we fall prey to the impact bias is that our
affective forecasts tend to place disproportionate emphasis on whatever our
attention is focused on at the moment. Such focalism can lead to systematic
errors in affective forecasting because features that are particularly salient when
predicting an event may be joined and even swamped by a host of other factors
when experiencing the event. This phenomenon is reinforced by the fact that
“predictions of future feelings inevitably involve some type of stylized represen-
tation of future events. Mental images of future vacations and holidays, for
example, typically do not include features such as rain, mosquitoes, and rude
28. Jeff T. Larsen, A. Peter McGraw, and John T. Cacioppo, Can People Feel Happy and
Sad at the Same Time?, 81 Journal of Personality and Social Psychology 684–96 (2001).
29. Julie A. Woodzicka and Marianne LaFrance, Real versus Imagined Reactions to
Sexual Harassment, 57 Journal of Social Issues, 15–30 (2001).
30. Timothy D. Wilson and Daniel T. Gilbert, Affective Forecasting, in Advances in
Experimental Social Psychology, Vol. 35 345–411 (Mark P. Zanna ed., San Diego, CA:
Academic Press, 2003).
31. Id. Much of the following discussion is drawn from Wilson and Gilbert’s excellent
review essay.
complexities of decision making 399
Predicted
Experienced
Emotional
intensity
service people . . .”32 People may overpredict the benefits of moving to a location
with nicer weather, not realizing that weather is just one of many factors affect-
ing their happiness—and that the weather is not always nice. When asked to
predict two months in advance how happy they would feel after a game if their
team won, football fans predicted a strong impact lasting a number of days.33 In
fact, although their happiness rose right after the Saturday game, it declined to
their normal baseline by the following Monday. In a parallel study that both
demonstrates the mechanics of focalism and suggests corrective action, partici-
pants made more realistic predictions about their happiness when they were
asked to write a “prospective diary” describing what particular activities they
would engage in on the Monday following the game.
Focalism also biases our affective forecasts when we focus our attention on
the immediate consequences of an event rather than on its effects in the longer
term. As Daniel Kahneman writes, we evaluate “an entire extended outcome by
evaluating the transition to it. For example, errors in predicting the well-being of
paraplegics may reflect people’s treating the tragic event of becoming a paraplegic
as a proxy in evaluating the long-term state of being a paraplegic.”34 And paraple-
gics spend much of their time in activities that are not compromised by their
disability.
Focalism is also encouraged by some decision-making processes, such as
elimination by aspects (See Section 4.3), that cancel out attributes that are shared
across alternatives and focus on those attributes that differ. Elizabeth Dunn and
her colleagues asked students to predict their happiness in living in various dor-
mitories, whose physical features differed considerably more than the social
aspects of dorm life.35 The students predicted their happiness or unhappiness
based solely on the former; yet, a year later, it turned out that their happiness was
not at all correlated with physical features, but was significantly correlated with
social life. When forecasting which of several options will make them happiest,
people tend to ignore the features that the options share (all dorms have active
social lives), while giving undue weight to features that make each option unique
(all dorms had different architecture).
Following Kahneman and Tversky,36 Dunn and colleagues called this
special case of focalism the isolation effect, because decision makers isolate the
unique features of each option. Focusing on the differences among competing
alternatives is not a poor decision strategy as such, as long as one takes account
of this phenomenon: “It would be quite reasonable to focus on location when
choosing between [college] houses that varied a great deal on this dimension
but not on others. When imagining a life in a specific house with a terrible
location, however, it would be unwise to expect lasting misery while neglecting
the happiness derived from living with one’s wonderful roommates.”
13.4.1.b Adaptation and “Ordinization” Neglect A second reason why the
impact bias occurs is that people tend not to consider the psychological pro-
cesses that keep our feelings and moods close to an emotional set-point. Processes
of hedonic adaptation blunt the effects of both good and bad events. (See Section
12.1.5) Affective reactions are generally most intense at the onset and diminish
over time. After experiencing setbacks, people usually get on with life, some-
times with reduced aspirations. Indeed, intensely disappointing experiences
tend to trigger proportionally stronger psychological coping mechanisms—such
as rationalizing and blaming others for one’s errors—than moderately disap-
pointing ones. Gilbert and Wilson call this the “psychological immune system.”
They term people’s tendency to underpredict these homeostatic tendencies
ordinization neglect.37
Predicting our emotional response to future events is no simple task. Yet,
given that we all accumulate years of experience forecasting and experiencing
emotional events, why do we continue to make forecasting errors without learn-
ing from our past mistakes? Gilbert and Wilson identify three reasons: (1) it is
35. Elizabeth W. Dunn, Timothy D. Wilson, and Daniel T. Gilbert, Location, Location,
Location: The Misprediction of Satisfaction in Housing Lotteries, 29 Personality and Social
Psychology Bulletin 1421 (2003).
36. Daniel Kahneman and Amos Tversky, Prospect Theory: An Analysis of Decision
Under Risk, 47 Econometrica 263–91 (1979).
37. Daniel T. Gilbert, Matthew D. Lieberman, Carey K. Morewedge, and Timothy D.
Wilson, The Peculiar Longevity of Things Not So Bad, 15 Psychological Science 14 (2004).
complexities of decision making 401
Legal Note
A healthy client asks you to prepare an advance health directive, specifying that life-
sustaining measures be withheld or withdrawn if she has an illness that is so severe
that there is no reasonable prospect that she will recover and be able to live without
continuing life-sustaining measures. What concerns would you raise with her?
See Jeremy Blumenthal, Law and the Emotions: The Problems of Affective Forecasting,
80 IND. L.J. 155 (2005).
Students were given a choice among six snacks, one of which would be eaten in
each of three successive class sessions. Some were required to choose their snacks
for all three days on the first day of the study. Others could choose the snack on
the day it was to be eaten. Those making all three choices in advance chose a sub-
stantially more varied group of snacks than those making choices day by day.40
These are all examples of how people mispredict changes of their tastes, pref-
erences, or behaviors over time or under different circumstances. George
Loewenstein and Eric Angner describe a variety of processes in which prefer-
ences can change:
• Habit formation, where consuming a substance or having an experience
increases one’s preference for it.
• Satiation, where consuming a substance or having an experience decreases
its marginal utility. (Satiation is not necessarily a continuous function; you
may suddenly stop enjoying an experience or commodity one day.)
• Refinement, where people sometimes expose themselves to experiences in
art, literature, or music to improve their appreciation of them.
40. Itamar Simonson, The Effect of Purchase Quantity and Timing on Variety Seeking
Behavior, 27 Journal of Marketing Research 150–62 (May 1990).
complexities of decision making 403
13.4.3 The Consequences of the Impact and Projection Biases and the
Possibility of Becoming More Accurate
The point of the preceding discussion is not that we usually mispredict our own
future feelings and preferences, but that we may do so in some situations that
have important consequences. These errors are not inevitably dysfunctional.
41. George Loewenstein and Daniel Adler, A Bias in the Prediction of Tastes, 105
Economic Journal 929–37 (July 1995).
42. Loewenstein and Angner, Predicting and Indulging Changing Preferences, in Time
and Decision.
43. John Mack Frargher, Daniel Boone: The Life and Legend of an American
Pioneer (Holt Paperbacks Press, 1993).
404 problem solving, decision making, and professional judgment
13.5 positive illusions: i’m above average, and my future looks rosy
We base important decisions not only on our forecasts of how various outcomes
would make us feel, but also on our judgments about how likely these outcomes
are to occur, as well our likelihood of contributing to bringing them about. For
example, an undergraduate might decide whether or not to apply to a competitive
law school based on her assessment of her academic skills, as well as her intu-
itions about her likelihood of acceptance. Yet these judgments tend to be
44. Loewenstein, O’Donoghue, and Rabin, Projection Bias in Predicting Future Utility,
Quarterly Journal of Economics 1209 (2003).
45. Id. The authors report on a study in which only 10 percent of healthy respondents
predict that they would accept a grueling course of chemotherapy to extend their lives by
three months, compared to 42 percent of current cancer patients.
46. George Loewenstein and David Schkade, Wouldn’t It Be Nice?: Predicting Future
Feelings, in Hedonic Psychology: Scientific Approaches to Enjoyment, Suffering,
and Well-Being (New York: Russell Sage Press, 1998).
47. See Jeremy A. Blumenthal, Emotional Paternalism, 35 Fla. St. U. L. Rev. 1 (2007).
complexities of decision making 405
systematically biased in ways that create positive illusions about oneself.48 In particu-
lar, people tend to be overly optimistic about their futures, and overestimate their
skills in a variety of domains. We consider each of these observations in turn.
13.5.1 Overoptimism
In When Every Relationship Is Above Average: Perceptions and Expectations of
Divorce at the Time of Marriage,49 Lynn Baker and Robert Emery asked couples
applying for marriage licenses to estimate the percent of couples in the United
States who marry today who will get divorced at some time in their lives. The
average estimate was 50 percent—a fairly accurate prediction based on current
statistics. When asked about the likelihood that they would ever be divorced, vir-
tually no one thought so. Earlier, we mentioned a study that showed that entre-
preneurs who were realistic about the general failure rate for new ventures were
overconfident about the success of their own ventures. By the same token, “most
people expect that they will have a better-than-average chance of living long,
healthy lives; being successfully employed and happily married; and avoiding a
variety of unwanted experiences such as being robbed or assaulted, injured in an
automobile accident, or experiencing health problems.”50
In an article on what they term the planning fallacy, Roger Buehler, Dale
Griffin, and Michael Ross describe some mega-projects that have taken much
more time and money to complete than originally projected—such as the Denver
International Airport, Boston’s Central Artery/Tunnel project (the “Big Dig”)—
and many more quotidian ventures, such as term papers, honors theses, and
finishing shopping for Christmas presents. The authors wonder at the ability of
people “to hold two seemingly contradictory beliefs: Although aware that most of
their previous predictions were overly optimistic, they believe that their current
forecasts are realistic.”51 Does this describe the reader’s personal experience?
Although overoptimism often involves predictions of one’s own future, a
study of physicians’ predictions of the life expectancy of terminally ill patients
indicated a strong bias toward overoptimism, with only 20 percent being accu-
rate (within 33 percent of actual survival), 17 percent being overly pessimistic,
52. Nicholas A. Christakis and Elizabeth B. Lamont, Extent and Determinant of Error in
Doctors’ Prognoses in Terminally Ill Patients: Prospective Cohort Study, 320 British Medical
Journal 469 (2000).
53. Armor and Taylor, supra.
54. See Section 8.2.3.
55. Armor and Taylor, supra.
56. George Loewenstein, Samuel Issacharoff, Colin Camerer, and Linda Babcock,
Self-Serving Assessments of Fairness and Pretrial Bargaining, 22 Journal of Legal Studies
135 (1993); Leigh Thompson and George Loewenstein, Egocentric Interpretations of Fairness
complexities of decision making 407
optimistic scenarios. Having people think about alternative scenarios is not suc-
cessful in countering overoptimism. The favored scenario provides too strong an
anchor. Although there is not much experimental evidence, Buehler, Griffin, and
Ross suggest that the most promising strategy with respect to personal projects
may be to draw people’s attentions to their past experiences in completing similar
projects, in effect tempering optimistic scenarios with actual personal base rates.
well as downsides.64 She writes that overoptimism usually “is not a Panglossian
whitewash that paints all positive events as equally and commonly likely and all
negative events as equally and uncommonly unlikely. Rather, unrealistic opti-
mism shows a patterning that corresponds quite well to the objective likelihood
of events, to relevant personal experiences with events, and to the degree to
which one can actively contribute to bringing events about. Positive events are
simply regarded as somewhat more likely and negative events as somewhat less
likely to occur than is actually the case.” More broadly, Taylor argues that these
moderate illusions create self-confidence that motivates people to persevere in
tasks that they might not otherwise attempt:
[P]eople who hold positive illusions about themselves, the world, and the
future may be better able to develop the skills and organization necessary to
make their creative ideas and high level of motivation work effectively for
them. They seem more able to engage in constructive thinking. They can tie
their illusions concretely to the project at hand, developing a task-oriented
optimism and sense of control that enable them to accomplish more ambi-
tious goals. They are more likely to take certain risks that may enable them to
bring their ventures to fruition. And they seem better able to choose appropri-
ate tasks, gauge the effort involved, and make realistic estimates of their need
to persevere. Moreover, by developing the ability to postpone gratification,
they are able to commit themselves to a longer term task that may involve
sacrifices and delayed rewards along the way.”
64. Shelley E. Taylor, Positive Illusions: Creative Self-Deception and the Healthy
Mind (New York: Basic Books, 1989).
410 problem solving, decision making, and professional judgment
65. David A. Armor, Cade Massey, and Aaron M. Sacket, Prescribed Optimism: Is It
Right to Be Wrong About the Future?, 19 Psychological Science 329–31 (2008).
complexities of decision making 411
$8,300
$8,200
Future Value with 1 year return
$8,100
$8,000
$7,900
$7,800
$7,700
$7,600
Rate of return less than 6.667%: Rate of return more than 6.667%:
Choose to wait 1 year for $8000. Take $7500 now, invest for 1 year.
$7,500
0% 1% 2% 3% 4% 5% 6% 7% 8% 9% 10% 11% 12%
1 year risk-free rate of return
where X is the amount of the future payment, r is the annual rate of return, and
n how many years in the future you will receive the payment. To say that a
discount rate is high means that you anticipate a high rate of return if you have
the funds today and are able to invest them.
Hard work often pays off after time, but laziness always pays off now.
—Author unknown
412 problem solving, decision making, and professional judgment
If choosing the alternative with the greater net present value usually seems
the sensible strategy, there are obvious exceptions. For example, suppose that
the interest rate is 6.7 percent, and
• the choice is between: $7400 today and $8000 a year from now. Even though
the $8000 is a great deal, taking the smaller amount now would allow you to
purchase a $7400 painting that you covet, which will not be available in a year;
or
• the choice is between $8000 today and $8000 a year from now. You are a
compulsive gambler and know you’ll blow any money you now have in Las
Vegas, but you’re about to enroll in a 12-step program.
These examples suggest that intertemporal preferences cannot be subject to
global criteria of rationality (like those of the axioms of subjective expected utility
theory), but rather depend on an individuals’ own tastes and plans. That said,
when people deviate significantly from the standard model or act as if the dis-
count rate were extraordinarily high, it raises a warning flag that they may not be
acting in their own best interests.
The standard model of discounted utility, first proposed by Paul Samuelson
in 1937, assumes that people discount at a constant rate over time, with the dif-
ference in utility between receiving $100 today and $200 a year from now being
proportional to the difference in utility between receiving $100 in ten years and
$200 eleven years from now. As shown in the solid line in Figure 13.3, this is an
exponential function.
1.2
1 Exponential
Value of Util Delayed by x Years
Hyperbolic
0.8
0.6
0.4
0.2
0
0 5 10 15 20 25 30 35 40 45 50
Delay (Years)
66. Some recent experimental evidence suggests that this may not be due to the differ-
ence between the present and future as such, but to the fact that people tend to have a
greater discount rate for periods of shorter duration whenever they occur. See Daniel
Read, Subadditive Intertemporal Choice, in Time and Decision 301.
67. Samuel M. McClure, David I. Laibson, George Loewenstein, and Jonathan D.
Cohen, Separate Neural Systems Value Immediate and Delayed Monetary Rewards, 306
Science 503–07 (Oct 15, 2004).
68. Howard Kunreuther, Ayse Onculer, and Paul Slovic, Time Insensitivity for Protective
Investments, 16 Journal of Risk and Uncertainty 279 (1998).
69. George Ainslee and Nick Haslam, Hyperbolic Discounting, in Choice Over Time
61 (George Loewenstein and Jon Elster eds., New York: Russel Sage Foundation, 1992).
70. George Ainslie and Vardim Haendel, The Motives of Will, in E. Gottheil, K.
Druley, T. Skolda and H. Waxman (eds.), Etiologic Aspects of Alcohol and Drug
Abuse 61 (Springfield, IL: Charles C. Thomas, 1983).
414 problem solving, decision making, and professional judgment
heart disease and cancer, excessive drinking, and drug dependence are
impeded by hyperbolic discounting.71
• Giddens also argues that “hyperbolic discounting is one of the main factors
explaining the lazy attitude most people have towards the threats posed by
global warming. Surveys show that the majority now accept both that cli-
mate change is real and dangerous and that it is created by our own behav-
iour. However, the proportion that has made any significant behaviour
change is very low. The implications are disturbing. Consciousness-raising
and green taxes, even if carefully thought out and organised, may have only
marginal effects—and they might be widely resisted even then.”72
In short, we “tend to grab immediate rewards and avoid immediate costs that
our ‘long-run selves’ do not appreciate.”73 Procrastination, in which postponing
inevitable costs often leads to increased costs (the paper poorly done at the last
minute, or the brief filed a day after the deadline) is a common example.
Addictions are another example. George Loewenstein notes that “visceral fac-
tors, such as the cravings associated with drug addiction, drive states (e.g.,
hunger, thirst, and sexual desire), moods and emotions, and physical pain . . .
[can] cause people to behave contrary to their own long-term self-interest, often
with full awareness that they are doing so.”74
There is evidence that children’s ability to defer gratification predicts
their success later in life. Walter Mischel conducted an experiment in which
four-year-olds had a marshmallow placed in front of them, and were given
a choice between eating it immediately or waiting a few minutes and getting
Personal Note
You are a very busy associate at a law firm. The firm’s pro bono coordinator
asks if you could commit to doing some work on a project that will require
about five hours a week for a number of weeks. You reply that this isn’t a
good time, but volunteer to do the work in several months—even though
you’ll be at least as busy then as you are now.
71. Anthony Giddens, This Time It’s Personal, The Guardian, Jan. 2, 2008, http://
www.guardian.co.uk/commentisfree/2008/jan/02/thistimeitspersonal.
72. Id.
73. Ted O’Donoghue and Matthew Rabin, Doing It Now or Later, 89 American
Economic Review 103 (1999).
74. George Loewenstein, Out of Control: Visceral Influences on Behavior, 65
Organizational Behavior and Human Decision Processes 272 (1996).
complexities of decision making 415
“The belief that a person should weight all utility the same, regardless of its tem-
poral position, implicitly assumes that all parts of one’s future are equally parts
of oneself; that there is a single, enduring, irreducible entity to whom all
future utility can be ascribed. However, some philosophers—most notably,
Derek Parfitt—deny this assumption. They argue that a person is nothing
more than succession of overlapping selves related to varying degrees by
physical continuities, memories, and similarities of character and interests.
On this view, the separation between selves may be just as significant as
the separation between persons, and discounting one’s ‘own’ future utility
may be no more irrational than discounting the utility of someone else.”
—Shane Frederick, Time Preference and Personal Identity, in TIME AND DECISION 89
the marshmallows used various strategies of this sort. An observer of video foot-
age reports: “Some cover their eyes with their hands or turn around so that they
can’t see the tray. Others start kicking the desk, or tug on their pigtails, or stroke
the marshmallow as if it were a tiny stuffed animal.78
13.6.2.b Self-Control, Habits, and Routines We can develop personal rules,
such as “no desserts” or “twenty minutes on the treadmill every morning.” Self-
control strategies often rely on creating a feeling of fear or guilt when faced with
the temptation to be avoided. In George Loewenstein’s and Ted O’Donoghue’s
words they “immediatize the delayed costs in the form of immediate [negative]
emotions.” However, using dieting as an example, Loewenstein and O’Donoghue
75. Yuichi Shoda, Walter Mischel, and Philip K. Peake, Predicting Adolescent Cognitive
and Self-Regulatory Competencies from Preschool Delay of Gratification: Identifying Diagnostic
Conditions, 26 Developmental Psychology 978–86 (1990).
76. George Ainslee and Nick Halsam, Self Control, in Choice Over Time, supra.
77. We have given different names to some of the strategies.
78. Jonah Lehrer, Don’t!: The Secret of Self Control, The New Yorker, May 18, 2009.
416 problem solving, decision making, and professional judgment
observe that the immediate costs are only justifiable in terms of the long-term
benefits, which (realistically) are often not achieved.79
Perhaps a more promising strategy than pure self-control is to develop habits
or routines that guard against self-defeating behavior. Although habits must be
learned over time through repetition, a program of research by Peter Gollwitzer
suggests that people can often obtain the benefits of good habits immediately by
engaging in a very simple cognitive exercise.80 In a series of studies, Gollwitzer
and his collaborators asked people with specific goals (such as writing a school
paper over their holiday break or performing a breast self-examination in the
subsequent month) to come up with a specific plan for the implementation of
those goals. Such “implementation intentions” involved specifying when, where
and how participants intended to pursue their goals. By creating “instant habits”
and making goal pursuit automatic when the right circumstances present them-
selves implementation intentions increase the likelihood that people will accom-
plish their goals. For example, if your implementation intention is to “order
fruit” for dessert at a restaurant, you are more likely to automatically order fruit
when the waiter hands you the dessert menu.
The motivations to exercise self-control are complex. As Drazen Prelec and
Rodnit Bodner note, “the gap between a single act of dietary self-sacrifice and its
benefits—slimness, leading to beauty, health, and virtue—is not only temporal
but also causal. A single low-calorie salad does not cause slimness. . . . Rather, the
benefits or costs are properties of a long-run policy or lifestyle, and an integral
policy or lifestyle cannot be obtained with a single decision. . . .”81 The authors
posit that postponing gratification produces a diagnostic reward—the pleasure
of learning something about yourself, in this case, about your willpower.
These inferences about a person’s own character and prospects are “a source of
immediate pleasure and pain, which contributes to his/her ability to exercise
self-control.”82 More broadly, we tend to see our lives in terms of narratives,
where deferring gratification today leads to better outcomes over time.
13.6.2.c Extrapsychic Mechanisms We can make decisions now that substi-
tute for, or significantly aid, the exercise self-control in the future. The classic
79. George Loewenstein and Ted O’Donoghue, We Can Do This the Easy Way or the
Hard Way: Negative Emotions, Self-Regulation, and the Law, 73 Chicago Law Review 183
(2006).
80. Peter M. Gollwitzer, Implementation Intentions: Strong Effects of Simple Plans, 54
American Psychologist 493–503 (1999).
81. Drazen Prelec and Ronit Bodner, Self-Signaling and Self-Control, in Time and
Decision 277.
82. Baumeister and Kathleen D. Voss analogize the exercise of willpower to physical
endurance, which is subject to exhaustion. Emotional distress undermines self-regulation
by changing people’s priorities to focus on feeling good in the immediate present. Roy
Baumeister and Kathleen D. Voss, Willpower, Choice, and Self-Control, in Time and
Decision 201. In any event, the resources that contribute to self-control are limited.
complexities of decision making 417
example is Ulysses having his sailors tie him to the mast so he could not respond
to the Sirens’ temptations. Keeping the refrigerator free of caloric snacks, or
leaving one’s credit card at home, are more mundane examples. Thomas Shelling
describes a drug addiction clinic in which patients subject themselves to self-
extortion: Patients write a self-incriminating letter confessing their drug addic-
tion, which will be mailed to an appropriate person (e.g., a physician’s letter
is addressed to the state board of medical examiners) if he or she is caught
taking drugs.83 The Web site, http://www.stickk.com/ invites people to make
“Commitment Contracts” for health and other personal goals, contracting to pay
charities (including causes they abhor) if they don’t meet milestones in achiev-
ing their goals.
Studies by Dan Ariely and Klaus Wertenbroch show how students use extra-
psychic mechanisms to counter the temptation to procrastinate. Students in
three sections of a marketing class at MIT were assigned three short papers to be
written during the semester. In one section, students were told that they could
hand in the papers anytime they wanted as long as all three papers were in by the
last day of class. In the second section, the professor imposed fixed, evenly-
spaced deadlines such that students had to complete one paper by the end of
each third of the course, and would be penalized for handing a paper in late. In
the third section, students were given the opportunity to set their own deadlines
for the papers. They could choose any deadlines they wanted as long as all three
papers were in by the end of the quarter, but once they had chosen their dead-
lines, they had to stick to them or they, too, would be penalized.
When given the opportunity to set their own deadlines, most students did not
choose the schedule that would have given them the most flexibility: setting
the deadlines for all three papers as late as possible. From a purely economic
perspective, this would have been the schedule with the greatest utility, since
you can always complete the papers earlier than the deadline. So why give up
flexibility, and risk being penalized for a late paper? The students were self-aware
enough to realize that if they weren’t forced to complete at least one of the papers
early in the semester, they were likely to procrastinate and be stuck trying to
write all three of them at the last minute.
It turned out that students who chose their own deadlines received
higher grades, on average, than students who were given no deadlines at all.
However, students whose professors imposed evenly-spaced deadlines did
even better.
Forestalling climate change is one of a number situations where the benefits from
costs paid today may redound to the benefit of future generations. If people often
discount in planning for their own lives, should they do so for the lives of future
generations? Noting that you would almost certainly prefer $1000 now to $1000
in twenty years, Cass Sunstein asks: “Is a life twenty years hence worth a fraction
of a life today?”84
The Office of Management and Budget (OMB) suggests that policy makers
“consider generational discounting in the same way as individual discounting,
but with an explicit discussion of the intergenerational concerns.”85 Alternatively,
OMB suggests “discounting for future generations at a slightly lower rate.”86 But
some scholars argue that generational discounting is radically different from indi-
vidual discounting, and that it is morally objectionable because it does not give
equal consideration to those living in the future.87 They argue for a zero rate of
discounting for intergenerational decisions.88 While we flag this complex and con-
troversial issue, it is beyond the scope of the book.
8485868788
14.1 prospect theory, the endowment effect, and status quo bias1
Neoclassical economics and its cousin, expected utility theory, posit that rational
choices maximize one’s overall wealth or, more generally, one’s overall utility.
Subject to Bernoulli’s observation (in Section 12.1.4) that people often display a
diminishing marginal utility of additional wealth, neoclassical economics, and
expected utility theory posit that a penny gained has as much positive value as a
penny lost has negative value. In fact, however, people systematically violate this
norm when evaluating changes from the status quo. Amos Tversky and Daniel
Kahneman formalized this aspect of the psychology of decision making in pros-
pect theory, which is illustrated by the S-shaped curve in Figure 14.1.
In this chapter, we discuss prospect theory in circumstances where risk or
uncertainty is not a significant factor. The value function described by prospect
theory has three important characteristics.
1. Individuals do not think of the outcomes of decisions in terms of their
overall wealth (as neoclassical economics holds). Rather, they consider out-
comes in terms of gains and losses from some reference point.
2. Individuals are loss averse. The slope of the of the curve for losses is steeper
than the slope for gains. This means that people experience losses from the
reference point as greater than equivalent gains. Loss aversion is captured
by the phrase that “losses loom larger than gains.”
1. This section draws heavily on Daniel Kahneman, Jack L. Knetsch, and Richard H.
Thaler, The Endowment Effect, Loss Aversion, and Status Quo Bias, in Choices, Values,
And Frames (Daniel Kahneman And Amos Tversky, eds., Cambridge, UK: Cambridge
University Press, 2000) 159. (2000), and Russell Korobkin, The Endowment Effect and Legal
Analysis, 97 Northwester University Law Review 1227 (2003).
420 problem solving, decision making, and professional judgment
Value
Losses Gains
3. Both the gain and loss functions display diminishing sensitivity. They are not
straight lines. Rather, the gain function is concave downward and the loss
function is concave upward.
Two behavioral economists’ explanation of the so-called equity premium puzzle
provides a nice example of loss aversion. The puzzle is why people are overin-
vested in low-yield treasury bills compared to equities. A dollar in stocks in 1926
would have been worth over $1800 at the close of the twentieth century, while a
dollar invested in treasury bills would have been worth only about $15. Ordinary
degrees of risk aversion do not account for so large a disparity. Shlomo Benartzi
and Richard Thaler explain the puzzle in terms of what they call myopic loss aver-
sion: Many investors examine their portfolios’ performance on a daily, if not
more frequent, basis, and tend to reset their reference point each time they look.
If losses loom larger than gains, the pain of a dollar’s loss is greater than the
pleasure of a dollar’s gain. Thus the excessive risk aversion caused by frequent
evaluation of their portfolios prevents investors “from adopting a strategy that
would be preferred over an appropriately long time horizon.”2
As this example suggests, there is no objectively correct reference point. An
individual’s conscious or unconscious choice of a reference point in a particular
situation is subjective, context-sensitive, and labile. Two people—for example,
two parties negotiating a contract—may have different references points for
exactly the same situation or transaction. Whatever reference point one chooses,
though, a potential loss is more unattractive than an equivalent potential gain is
attractive.
2. Shlomo Benartzi and Richard H. Thaler, “Myopic Loss Aversion and the Equity
Premium Puzzle,” NBER Working Paper No. W4369. May 1993, available at SSRN: http://
ssrn.com/abstract=227015.
complexities of decision making: frames 421
The example also illustrates that people often choose the status quo as a refer-
ence point from which to evaluate change. As a consequence of loss aversion, all
things considered, people like things the way they are more than they like change.
This general phenomenon is called the status quo bias. We begin by describing
an important subset of status quo bias, the endowment effect, which captures the
fact that people often demand more to give up a good or entitlement than they
would be willing to pay to acquire it. The endowment effect arises in many con-
texts and often confounds policy making.
For example, contingent valuation (discussed in Section 14.6) is a common
method, used in cost-benefit analysis and the assessment of damages, for plac-
ing a value on environmental goods or harms: affected individuals are surveyed
to determine how much they would be willing to pay (WTP) to gain an environ-
mental good or how much they would be willing to accept (WTA) to suffer the
loss of that good. Contingent valuation is premised on the assumption that an
individual’s WTP and WTA for a good are identical. But it turns out that WTP
and WTA often differ by large amounts. Russell Korobkin reports:
In one notable study, 2000 duck hunters were surveyed about the value they
would place on protecting a wetland from development. Hunters were willing
to pay $247 per person per season, on average, for the right to prevent devel-
opment to make hunting viable, while they would demand, on average, $1044
dollars each to give up an entitlement to hunt there. . . . [Another] survey
found that a sample of residents of a region of the southwest would pay $4.75
per month, on average, to maintain 75 miles of air visibility in the face of
potential pollution that would reduce visibility to 50 miles. In contrast, how-
ever, people drawn from the same pool and told that they collectively enjoyed
the right to prohibit the pollution, reported that they would demand $24.47
per month before being willing to permit the pollution.3
In a classic demonstration of the endowment effect,4 half the students in a
classroom—the sellers—were handed decorated university mugs (costing $5 at
the university bookstore) and given a questionnaire that stated: “You now own
the object in your possession. You have the option of selling it if a price, which
will be determined later, is acceptable to you. For each of the possible prices
below, indicate whether you wish to sell your object and receive this price; or
keep your object and take it home with you.” The participants indicated their
decision for prices ranging from $0.50 to $9.50 in increments of 50 cents.
Students who had not received a mug—the buyers—received a similar question-
naire, informing them that they would have the option of receiving either a mug
5. Amos Tversky and Daniel Kahneman, Loss Aversion in Riskless Choice: A Reference-
Dependent Model, 106 Quarterly Journal of Economics 1039 (1991). In another experi-
ment, half the participants were given pens, and half were given a token redeemable for
an unspecified gift. Then, all of the participants were given a choice between the pen and
two chocolate bars. Fifty-six percent of those who already had the pen preferred the pen,
compared to 24 percent of the others. In a separate part of the experiment, when asked to
rank the attractiveness of a number of items, including the pens and chocolate bars, the
participants rated the chocolate bars as at least as attractive. “This suggests that the main
effect of endowment is not to enhance the appeal of the good one owns, only the pain of
giving it up.” Kahneman, Knetsch, and Thaler, supra. In yet another experiment involving
mugs and chocolate, participants were randomly given either a coffee mug or a large
Swiss chocolate bar and offered the opportunity to trade. Only 11 percent and 10 percent
of the participants, respectively, were willing to trade.
6. Jennifer Arlen, Matthew Spitzer, and Eric Talley, Endowment Effects within Corporate
Agency Relationships, 31 Journal of Legal Studies 1 (2002).
7. Edward McCaffrey, Daniel Kahneman, and Matthew Spitzer, Framing the Jury:
Cognitive Perspectives on Pain and Suffering Awards, 81 Virginia Law Review 1341 (1995).
complexities of decision making: frames 423
Very few switch.8 As a real-world analogue, consider the large number of indi-
vidual investors who will hold on to shares of stock when they would not pur-
chase the same stock at the same or even a lower market price. Consider whether
the various legal doctrines that reflect the saying that “possession is nine points
of the law,” or doctrines that give greater due process rights to someone already
holding a government job or benefit than one seeking it, may reflect the endow-
ment effect.
People may gain “psychological ownership” merely by having an item in their
physical possession. In a recent variation on the mug experiment, students were
asked to physically examine the mugs for different periods of time before bid-
ding on them in an auction. Those who held the mugs for longer period of time
bid more.9
As we mentioned, the endowment effect is a subset of the status quo bias—the
general phenomenon whereby people attach a value to the present state of the
world compared to alternative states, even when no property right or legal entitle-
ment is involved. For example, as of the 1990s, the standard auto insurance policy
in New Jersey did not provide a right to recover for pain and suffering for minor
injuries, though one could obtain that right by purchasing a higher priced policy.
In Pennsylvania, the standard policy did provide the right to recover, but one
could purchase a cheaper policy that relinquished the right. In a survey of resi-
dents of the two states, only 20 percent of New Jersey respondents and 25 percent
of Pennsylvanians indicated they would switch from their state’s default policy.10
Also, as we noted above, what one treats as a reference point is essentially a
subjective matter. Russell Korobkin conducted an experiment to explore whether
the parties negotiating a contract would treat a default contract term as the status
quo.11 Students played the role of lawyer negotiating a shipping contract on
behalf on an overnight delivery company. In one scenario, they were told that
there were two possible terms that could be used to deal with consequential dam-
ages for loss or delay of packages: (A) the delivery company would be liable only
for reasonably foreseeable damages, or (B) the delivery company would be liable
for all consequential damages, whether or not foreseeable. In another scenario
they were told that there were two possible terms for dealing with unforeseen
circumstances that make delivery commercially impracticable: (A) the delivery
company would be excused from its obligation to deliver, or (B) the delivery
company would be liable for damages. Participants were then given estimates of
the differential costs to their client (on a per-package basis) of performing its
obligations if the contract contained term B rather than A, and were then asked
to value term A relative to term B.
One-half of the participants were told that term A was the default rule and
were asked to provide their WTA price (how much their clients would need to be
paid per package) to agree to insert term B into the agreement to contract around
the default rule. The other half of the participants were told that term B was the
default rule and were thus asked to provide their WTP for inserting term A into
the contract in order to contract around the default rule. In both scenarios, WTA
significantly exceeding WTP; in other words, the strength of the participants’
preferences for the term that benefited their client was biased in favor of which-
ever term was identified as the default.
In Korobkin’s experiment, the contracting parties were induced to treat the
legal default rule as the status quo. However, Marcel Kahan and Michael Klausner
hypothesized that contracting parties might treat the terms embodied in a stan-
dard form contract as the reference point. “Although these parties have no formal
property (or other) rights in a standard term, the standard sets terms for an
expectational baseline . . .”12 Korobkin subsequently tested this hypothesis by
redesigning his experiments to inform the participants negotiating a contract
term about both the legal default rule and the term in the standard form con-
tract. They were told that the form term was the opposite of the legal default rule,
so that if they contracted around the form term, they would reestablish the
default term that the standard form circumvented. The participants’ preferences
turned out to be biased in favor of the form term rather than the default rule.
Korobkin concludes:
These results suggest that contract terms will be biased in favor of the status
quo, but that the reference point that identifies the status quo for contracting
parties is extremely dependent on context. The implication for legal policy is
that the content of contract law probably substantively affects the content of
contracts, but not all the time. The effect of contract law can be swamped by
other reference points, depending on the circumstances. The prescriptive
implications for contract lawyers seems more clear: “negotiators who are able
to define the status quo position, against which all proposed terms are judged,
are likely to enjoy an important bargaining advantage.”13
12. Marcel Kahan and Michael Klausner, Path Dependence in Corporate Contracting:
Increasing Returns, Herd Behavior and Cognitive Biases, 74 Washington University Law
Quarterly 347, 361 (1996).
13. Russell B. Korobkin, The Status Quo Bias and Contract Default Rules, 83 Cornell
Law Review (1998).
complexities of decision making: frames 425
14. Edward McCaffrey, Daniel Kahneman, and Matthew Spitzer, Framing the Jury:
Cognitive Perspectives on Pain and Suffering Awards, 81 Virginia Law Review 1341 (1995).
15. Charles R. Plott and Kathryn Zeiler, The Willingness to Pay/Willingness to Accept
Gap, The “Endowment Effect,” Subject Misconceptions and Experimental Procedures for
Eliciting Valuations, 95 American Economic Review 530–45 (2004).
426 problem solving, decision making, and professional judgment
but the thought and the preference might shift with evidence that the employer
has made enrollment automatic. With respect to savings, the designated
default plan apparently carries a certain legitimacy for many employees, per-
haps because it seems to have resulted from some conscious thought about
what makes most sense for most people. This interpretation is supported by
the finding that the largest effects from the new default rule are shown by
women and African-Americans. We might speculate that members of such
groups tend to be less confident in their judgments in this domain and may
have less experience in assessing different savings plans.
But there are many situations where the status quo or the fact that one is
endowed with an object provides no informational content.
The endowment effect may be affected by the phenomenon of projection bias
described in Section 13.4.2, which may lead to exaggerated predictions of feel-
ings of loss.19 On the other hand, a failure of empathy may lead buyers to under-
estimate the strength of the endowment effect on owners and thus cause
distortions in negotiations “because buyers will tend to underestimate owners’
reservation prices. By the same token, owners will tend to overestimate buyers’
reservation prices.”20
Another explanation for the endowment effect and status quo bias is the
desire to avoid regret—“the painful feeling a person experiences upon determin-
ing she could have obtained a better outcome if she had decided or behaved
differently.”21 Of course, people also rejoice when the actual outcome is superior
to what might have happened. But, consistent with loss aversion, the negative
emotion of regret tends to be stronger.
Consistent with a “regret avoidance” account of the endowment effect, Ziv
Carmon and Dan Ariely have shown that, when contemplating a sale, both
buyers and sellers tend to focus on what the sale would cause them to forgo in
determining what they think the price should be. Since buyers and sellers forgo
different things in a sale (sellers forgo the item that is being sold and buyers
forgo the money they spend in the sale), buyers’ WTP and sellers’ WTA are likely
to diverge. Carmon and Ariely asked students at Duke University to contemplate
a sale of tickets to a basketball game. They found that buyers’ WTP was mainly
19. George Loewenstein, Ted O’Donoghue, and Matthew Rabin, Projection Bias in
Predicting Future Utility, Quarterly Journal of Economics 1209 (2003).
20. Id. Leaf Van Boven, David Dunning, and George Loewenstein conducted an exper-
iment in which sellers endowed with coffee mugs were asked to estimate how much
buyers would pay, and buyers were asked to estimate how much sellers would charge.
Sellers predicted that buyers would pay $3.93, compared to buyers’ WTP of $1.85; buyers
predicted that sellers would demand $4.39, compared to sellers WTA of $6.37. Egocentric
Empathy Gaps Between Owners and Buyers: Misperceptions of the Endowment Effect, 79
Journal of Personality and Social Psychology 66 (2000).
21. Chris Guthrie, Better Settle Than Sorry: The Regret Aversion Theory of Litigation
Behavior, 1999 U. Ill. L. Rev. 43.
428 problem solving, decision making, and professional judgment
influenced by monetary considerations such as the base price of the tickets and
how heavily they were discounted. Meanwhile, sellers’ WTA was more heavily
affected by factors related to the experience of the game, including the atmo-
sphere of the stadium and the importance of the game.22
22. Ziv Carmon and Dan Ariely, Focusing on the Forgone: How Value Can Appear So
Different to Buyers and Sellers, 27 Journal of Consumer Research 360–70 (2000).
23. Although a decision maker cannot undo the behavior that she now regrets, a deci-
sion maker may learn from her own experience and avoid the regret-causing behavior in
the future. Moreover, there may be categories of cases where regret is so ubiquitous that
a counselor to the decision maker would predict the consequences.
complexities of decision making: frames 429
24. Marcel Zeelenberg, Jane Beattie, Joop van der Pligt, and Nanne K. de Vries,
Consequences of Regret Aversion, 65 Organizational Behavior and Human Decision
Processes 148 (1996).
25. By the same token, we especially regret bad outcomes from exceptions to our rou-
tines: “Mr. Adams was involved in an accident when driving home after work on his regu-
lar route. Mr. White was involved in a similar accident when driving on a route that he
only takes when he wants a change of scenery.” Eighty-two percent thought that Mr. White
would be more upset than Mr. Adams. Daniel Kahneman and Dale T. Miller, Norm Theory:
Comparing Reality to Its Alternatives 93 Psychological Review 136–53 (1986).
26. Victoria H. Medvec, Scott F. Madley, and Thomas Gilovich, Where Less Is More:
Counterfactual Thinking and Satisfaction Among Olympic Medalists, Journal of Personality
and Social Psychology 625 (1995).
430 problem solving, decision making, and professional judgment
27. This may also be due to the availability heuristic: taking actions is often more vivid
and memorable than doing nothing.
28. Daniel Kahneman and Amos Tversky, The Psychology of Preferences, 246 Scientific
American 246 (1982). See also Daniel Kahneman and Dale Miller, Norm Theory: Comparing
Reality to Its Alternatives, in Heuristics and Biases: The Psychology of Intuitive
Judgment 397, 348 (Thomas Gilovich, Dale Griffin, and Daniel Kahneman eds.,
New York: Cambridge University Press, 2002) [hereafter Heuristics and Biases].
29. Dale Miller and Brian Taylor, Counterfactual Thought, Regret, and Superstition: How
to Avoid Kicking Yourself, in Heuristics and Biases, at 367.
complexities of decision making: frames 431
will speed up and the new one will slow down. “To undertake a risky action
one easily imagines otherwise . . . is to tempt fate. . . . One is just ‘asking for it’
if one takes an action he or she knows, in retrospect, will be easy to imagine
otherwise.30 Daniel Keys and Barry Schwartz note the superstition that bullfight-
ers should not substitute for other matadors at the last minute as an example of
how regret in the decision process can “leak” into outcomes (See Section 12.1.2).
“This superstition may be rational, as the negative consequences of a matador
being gored include not just the physical damage to the matador but also the
regret, doubt, and other emotional suffering that can spread throughout the bull-
fighting community after a goring. A last-minute matador switch is a salient
event that tends to induce counterfactual thinking and increase susceptibility to
regret.”31
Actions can cause regret directly when they end in negative outcomes—as
when the decision to take a card makes a blackjack player “bust.” And actions
can also cause regret indirectly by changing the way one thinks about one’s past
behavior. Consider this example: Artie is about to take his first trip on Global
Airlines and is invited to join its frequent flier program. Ben is in the same posi-
tion, but Ben took a long trip on Global Airlines last month and didn’t take the
trouble to enroll in the frequent flier program then. Who is more likely to enroll
on this trip?
Research by Orit Tykocinski and Thane Pittman suggests that Artie is more
likely to enroll in the frequent flyer program than Ben.32 If he joined the frequent
flier program now, Ben would have to admit that his failure to join earlier had
been a mistake. His current action would make him regret his past inaction.
As the authors put it, “by quickly turning down the subsequent action opportu-
nity . . . one can avoid thinking about the inferior current opportunity that
triggers the perception of loss.” Unlike the blackjack players, who do not know
whether the next card will make them win or bust, Ben knows in advance that
joining the frequent flier program will have beneficial results. Yet they still avoid
taking these actions because doing so would make them regret their previous
procrastination.
In some circumstances, inaction is more likely to cause more regret than
action—even in the short run. Consider the dilemma of the soccer goalie as his
opponent swings his foot back for a penalty kick. Should the goalie move left,
right, or stay put? Although goalies are more than twice as likely to make a save
if they stay in the center of the goal, it turns out that they only rarely stay put. The
anticipation of regret may help explain these findings: professional goalkeepers
indicated that they would feel more regret if they let a penalty kick go by without
moving than if they had taken an ineffective action. Why would inaction cause
more regret in this case than action? The answer may relate again to the ease of
generating counterfactuals. Because goalkeepers are expected to act, action may
be less mutable than inaction. In other words, it may be easier to generate the
counterfactual “if only I’d jumped for the ball” than it is to think, “if only I’d
stayed put.”33 Perhaps action is the status quo for many athletes.
Whether actions or inactions beget more regret also seems to depend on the
time horizon. While in the short run people most regret actions, in the long run
they most regret failures to act.34 Research by Thomas Gilovich and Victoria
Medvec found that people’s most lasting regrets involved not failures to take
discrete actions, such as failing to sign up for the discounted Tuscany tour, but
rather patterns of inaction, such as not investing sufficiently in educational
opportunities or spending enough time with friends. Still, some discrete inac-
tions, such as failure to seize a romantic or business opportunity, can have last-
ing consequences. A popular web comic humorously illustrates this point with a
bar graph entitled “Number of Google Results For I _______ Have Kissed Her
(Or Him).” (See Figure 14.2.) The bar that fills in the blank with “should” reveals
almost nine times more hits than the bar for “shouldn’t.” 35
10,230
1,213
SHOULDN’T SHOULD
figure 14.2 number of google results for “i____ have kissed her (or
him). http://xkcd.com/458/ . reprinted by permission of xkcd.
33. Patricia Cohen, The Art of the Save, for Goalie and Investor, New York Times, March
1, 2008, Business Section, online edition.
34. Thomas Gilovich and Victoria H. Medvec, The Experience of Regret: What, When,
and Why, 102 Psychology Review 379 (1995).
35. XKCD webcomic, http://xkcd.com/458/.
complexities of decision making: frames 433
Gilovich and Medvec identify a number of factors that may diminish regret
about actions and increase regret about inactions over time:
• Elements that reduce the pain of regrettable actions so that the initial pain
of the regrettable action diminishes over time:
• Ameliorating the action. The pain of regrettable action is felt more
quickly and the action can often be ameliorated. (If you marry
Mr. Wrong, you can get divorced; if you fail to marry Mr. Right, he may
no longer be available.)
• Reducing dissonance, for example, by identifying silver linings (“At least
we had wonderful children”; “I learned a lot from making that bad
investment.”)
• Elements that bolster the pain of regrettable inactions.
• The passage of time makes one more confident that one could have
succeeded in the forgone action, thus making it easy to mentally “undo”
the inaction. (“We thought religious differences would get in the way of
our marriage, but we could have worked them out.”)
• While regrets about actions focus on bad things that actually happened,
regrets about inactions concern good things that might have happened—
an open-ended category that can grow in scope and importance with the
passage of time.
• Elements that differentially affect cognitive availability. Many regrets about
inaction involved unrealized ambitions and unfulfilled intentions that are
more available to our minds than goals that have been fulfilled.
36. Richard P. Larrick and Terry L. Boles, Avoiding Regret in Decisions with Feedback: A
Negotiation Example, 63 Organizational Behavior and Human Decision Processes
87–97 (1995).
434 problem solving, decision making, and professional judgment
will have no idea what the judge would have awarded had the case gone to trial.
In the “regret” jurisdiction, the judge informs the parties what he would
have awarded. By a large margin, the participants predicted that the parties—-
defendants as well as plaintiffs—would be more prone to settle in the traditional
jurisdiction, where they would be shielded from knowledge of the foregone out-
come and thus shielded from regret.37 Whether to encourage settlement or just
to avoid the costs of unnecessary adjudication, most real-world procedures shield
parties negotiating a settlement from learning what the specific foregone out-
come might have been.
Suppose that you are watching a movie and you find it to be boring and unlikely
to get any better. Do you nonetheless keep watching the movie? Suppose that
you paid $10 to see the movie. Does that affect your decision? In a study by
Deborah Frisch, 57 percent of people who said they would stop watching a free
movie said they would keep watching it if they had paid $10.41
This phenomenon of valuing sunk costs was replicated in an experiment
involving a campus theater company at the University of Ohio. Theatergoers
who were about to buy season tickets were randomly placed into groups that
either paid full price or received a discount. During the first half of the season,
those who paid full price attended many more plays than those who had received
the discount—though the rates of attendance converged for the second half of
the season.42
Economic theory holds that the price they paid for the ticket should not affect
ticket-holders’ decision whether to attend. Whether they paid full price or
received a discount, the cost of the ticket is “sunk” and can have no effect on the
utility of attending—as compared to whether or not it is snowing that evening,
or the ticket holder has a paper due the next day, or is or just isn’t in a theater-
going mood. But for those who paid full price, the loss of the money was more
salient than for those who got the discount.
The phenomenon of honoring sunk costs manifests itself in many personal,
government, and business decisions. For example:43
• Members of a health club with semi-annual dues made greater of use of
the facility in the month after they paid dues; use declined over the next five
months, until the cycle repeated itself.44
• A soldier on the first day of the Gulf War in 1991: “Finally, the day has
finally come. You’ve got to think logically and realistically. Too much
money’s been spent, too many troops are over here, too many people had
too many hard times not to kick somebody’s ass.”
• The parents of American troops killed in Iraq:45 “To say the war right now is
worthless or that it is without cause or unjust, would be saying that my son, for
all he had in him to go over there to preserve the freedom of his country, would
have died in vain.” “Don’t take our soldiers out of Iraq. If you do, it means
our loved ones will have died for nothing.” (This is not the universal view.
41. Deborah Frisch, Reasons for Framing Effects, Organizational Behavior and Human
Decision Processes, 54, 399–429 (1993), cited in Daniel J. Keys and Barry Schwartz, “Leaky”
Rationality: How Research on Behavioral Decision Making Challenges Normative Standards
of Rationality, 2 Perspectives on Psychological Science, 162–80 (2007) [hereafter Leaky
Rationality].
42. Hal R. Arkes and Catherine Blumer, The Psychology of Sunk Cost, 35 Organizational
Behavior 124 (1985).
43. Many of the quotations come from Reid Hastie and Robyn Dawes, Rational
Choice in an Uncertain World: The Psychology of Judgment and Decision
Making (Sage Publications: Thousand Oaks, CA, 2001) 32.
44. John T. Gourville and Dilip Soman, Payment Depreciation: The Effects Of Temporally
Separating Payments from Consumption, 25 Journal of Consumer Research 160 (1998).
45. New York Times, Sept. 9, 2004.
436 problem solving, decision making, and professional judgment
The mother of a soldier who died in Iraq said: “It pains me to hear that more
people should die because those people have died. That makes no sense. We
can honor them by having an intelligent, honest policy.”46)
• Senator Jim Sasser arguing for further investment in a project that would
be worth less than the amount of money necessary to finish it: “Completing
Tennessee-Tombigbee is not a waste of taxpayers’ dollars. Terminating the
project at this late stage of development would, however, represent a seri-
ous waste of funds already invested.”
• A businessman: “I’ve already invested so much in the Concorde airliner . . .
that I cannot afford to scrap it now.”
• Members of Congress arguing for the dollars and lives already spent on
NASA’s space shuttle program as a reason for continuing it.47
• The more people pay for advice from consultants or other experts with
respect to business or personal decisions, the more likely they are to follow
the advice.48
If taking account of sunk costs doesn’t make economic sense, is it therefore
always irrational? Consider:
• Avoiding regret. If the $500 treadmill you bought several months ago is sit-
ting idly in your family room, you may mentally kick yourself each time you
look at it and regret the purchase decision. Like all feelings, the sense of
regret is arational. Whether it leads to behaviors that are in your long-term
self-interest is a matter of chance. While anticipatory regret about the idle
treadmill may lead you to exercise more often, anticipatory regret about
wasting the nonrefundable airline ticket to a country that has had an upsurge
in terrorism may lead you to take a dangerous and anxiety-laden trip.
• Reputation. Individuals and organizations may rationally value their reputa-
tions for not making poor decisions, for adhering to commitments, and for
not being wasteful. Thus, even if a policy maker recognizes that it is subop-
timal to continue a project or activity, she may decide that it is in her or the
agency’s interest to obscure a poor decision by sticking with it49—or not to
be seen as a “flip-flopper.” We imagine that circumstances of this sort often
give rise to conflicts of interest between agents and principals—for
example, between the interests of a decision maker and the clients or con-
stituents she represents (who bear the costs of continuing a bad project).
• Sense of self, and self-education. “Reputation” refers to how other people think
of you. But you may also have a sense of yourself—a self-reputation—as a
person who is not wasteful and who follows through with plans and com-
mitments. Moreover, you may want to discipline yourself to be that kind of
a person. Keeping commitments that you wish you had not made may teach
you to be more cautious in the first place.50
• Precommitment. In Section 13.6.2 we discussed strategies through which
individuals could commit themselves to long-run valued behaviors in the
face of short-run temptations. Unlike Ulysses tying himself to the mast,
most of our precommitments are open to subsequent reneging. Giving
weight to sunk costs—e.g., the $500 treadmill sitting in your family room—
can be a psychic incentive for maintaining those commitments.51
• Respect for prior decisions and finality. Like judges’ respect for precedent,
sticking with earlier decisions even when they seem questionable may
reflect a combination of respect for their predecessors and the costs of con-
stantly revisiting earlier judgments. Even for personal decisions, individu-
als may believe that they do not have a better vantage point now than when
they initially made the decision.
• Meaning. The parents’ comments about loss of their children in Iraq sug-
gest that attending to sunk costs may give meaning to terrible events in our
lives—though sometimes at the cost of creating more terrible events.
Whether or not taking account of sunk costs sometimes makes sense for indi-
viduals, we are skeptical about its value to governments and private organiza-
tions. In any event, it is important to understand that a decision that takes
account of sunk costs inevitably compromises the pursuit of other goals by
diverting resources that might be used more productively or, occasionally, by
continuing to endanger people’s welfare and lives. We will return to the issue of
sunk costs when we discuss the escalation of commitment in Section 17.3.3.
Under classic economic theory, a decision maker views the money or other assets
she spends on, or gains from, a transaction in terms of their effect her overall
50. See Richard Epstein, Skepticism and Freedom: A Modern Case for Classical
Liberalism, ch. 8 (Chicago, IL: University of Chicago Press, 2003); Walton, supra.
51. Consider, though, whether taking account of the sunk costs to goad you into exer-
cising is best characterized as a “rational” decision, or whether it would be more accurate
to say that the overall strategy is (rationally) based on the prediction that you will (irratio-
nally) take account of sunk costs.
52. This section relies heavily on Richard H. Thaler, Mental Accounting Matters, in Choices,
Values, and Frames (Daniel Kahneman and Amos Tversky, eds., Cambridge, UK:
Cambridge University Press, 2000) 241 [hereafter, Mental Accounting Matters]. See also Richard
H. Thaler, Mental Accounting and Consumer Choice, 4 Marketing Science 199 (1995).
438 problem solving, decision making, and professional judgment
“Injuries, therefore, should be inflicted all at once, that their ill savor being less
lasting may the less offend; whereas, benefits should be conferred little by little,
that so they may be more fully relished.”
—MACHIAVELLI, THE PRINCE
14.4.2 Budgeting
Kahneman and Tversky posed essentially these questions to two randomly
selected groups in 1984.
• Imagine that you have decided to see a play presented by a campus theater
company and paid the admission price of $10 per ticket. As you enter the
theater, you discover that you have lost the ticket. Seats were not assigned,
and the ticket cannot be recovered. Would you pay $10 for another ticket?
• Now imagine the same scenario, except that you haven’t yet purchased the
ticket. As you enter the theater, you discover that you have lost a $10 bill.
Would you still pay $10 for a ticket to the play?
Only 46 percent of those who lost the ticket would purchase a new ticket,
while 88 percent of those who lost the $10 bill would purchase another one. Why
would this be? Kahneman and Tversky explain: “Going to theater is normally
viewed as a transaction in which the cost of the ticket is exchanged for the experi-
ence of seeing the play. Buying a second ticket increases the cost of seeing the
play to a level that many respondents apparently find unacceptable. In contrast,
the loss of the cash is not posted to the account of the play, and it affects the
purchase price of a ticket only by making the individual feel slightly less
affluent.”57
In another study, participants in one group were told that they had spent $50
for a ticket to a basketball game earlier in the week and asked if they would be
willing to buy a theater ticket. Participants in the second group were told that
they had paid $50 for a parking violation and asked if they would purchase the
theater ticket. Because the theater ticket was drawn from the “entertainment”
budget, people who had spent the $50 for a baseball ticket were less likely to buy
the theater ticket than those who paid $50 for the parking fine.
Dividing spending into budget categories serves two purposes. It can make
the trade-offs between competing uses of funds transparent and, to the extent
that certain accounts are “off limits,” it can assist in self-control. When budgets
are not treated as fungible, however, the existence of separate accounts can have
distorting economic effects.
57. Daniel Kahneman and Amos Tversky, Choices, Values, and Frames, 39 American
Psychologist 341 (1984).
58. Matthew D. Shapiro and Joel Slemrod, “Consumer Response to Tax Rebates,”
working paper, University of Michigan and NBER (2001).
59. Nicholas Epley, Rebate Psychology, New York Times, Op-Ed section, Jan. 1, 2008.
complexities of decision making: frames 441
60. Amos Tversky and Daniel Kahneman, The Framing of Decisions and the Psychology
of Choice, 211 Science 453 (1981); Richard H. Thaler, Toward a Positive Theory of Consumer
Choice, in Choices, Values, and Frames 269 (New York: Cambridge University Press,
2000).
61. France LeClerc, Bernd H. Schmidt, and Laurette Dube, Decision Making and Waiting
Time: Is Time Like Money?, 22 Journal of Consumer Research 256 (1995).
62. Mental Accounting Matters 256.
63. Mental Accounting Matters 267.
442 problem solving, decision making, and professional judgment
Look at this peanut butter! There must be three sizes of four brands of four consistencies!
Who demands this much choice? I know! I’ll quit my job and devote my life to choosing
peanut butter! Is “chunky” chunky enough or do I need extra chunky? I’ll compare ingre-
dients! I’ll compare brands! I’ll compare sizes and prices! Maybe I’ll drive around and see
what other stores have! So much selection and so little time!
—Bill Watterson, It’s A Magical World (1996) (Calvin & Hobbes)
64. Sheena S. Iyengar and Mark R. Lepper, Choice and Its Consequences: On the Costs
and Benefits of Self-Determination (2002), in Abraham Tessler, Diederik A. Stapel, and
Joanne V. Wood (eds), Self and Motivation: Emerging Psychological Perspectives)
(Washington, DC, American Psychological Association, 2002) 71, citing Lawrence M.
Friedman, The Republic of Choice (Cambridge: Harvard University Press, 1990).
complexities of decision making: frames 443
All things being equal, we think it is good to have options—the more that
are available, the more likely we’ll find one that’s a good fit with our interests,
needs, or objectives. But choices carry costs as well. As the number of choices
increase, decision making becomes more difficult, time consuming, and even
stressful.
People faced with many options tend to employ simplifying decision strate-
gies.65 For example, in studies of how people evaluate home housing purchases,
participants used compensatory procedures when choosing among three options,
but switched to a noncompensatory procedure when choosing among a dozen.66
As the number of housing options increased, the percentage of choices that
reflected the participants’ real preferences (as elicited in a separate process)
declined from 70 percent when considering five houses to 37 percent when con-
sidering twenty-five.67 A study of professional health care purchasers for large
corporations showed that the large majority used noncompensatory strategies in
choosing among health plans that had many different attributes.68
The remainder of this section considers another set of problems caused
by options, most of which involve (1) violation of the axiom of independence from
irrelevant alternatives, which holds that the relative ranking of a particular set of
options should not vary with the addition or deletion of other options, or (2) vio-
lation of the axiom of invariance, which holds that the order in which choices are
presented (or the way they are described) should not affect the outcome of the
decision. (See Section 12.2.2)
65. See generally John W. Payne, James R. Bettman, and Eric J. Johnson, The
Adaptive Decision Maker (New York: Cambridge University Press, 1993).
66. See Chapter 4 for the distinction between these procedures.
67. Richard W. Olshavsky, Task Complexity and Contingent Processing in Decision
Making: A Replication and Extension, 24 Organizational Behavior and Human
Performance 300 (1979); Naresh K. Malhorta, Information Load and Consumer Decision
Making, 8 J. Consumer Res. 419 (1982)—both summarized in Chris Guthrie, Panacea or
Pandora’s Box? The Costs of Options in Negotiations, 88 Iowa Law Review 601, 632–33
(2003) [hereafter, Guthrie, Options].
68. Judith H. Hibbard, Jacquelyn J. Jewett, Mark W. Legnini, and Martin Tusler,
Choosing a Health Plan: Do Large Employers Use the Data?, Health Affairs 172 (Nov.–
Dec. 1997). Also summarized in Guthrie, Options.
69. Guthrie, Options.
444 problem solving, decision making, and professional judgment
70. Lyle Brenner, Yuval Rottenstreich, and Sanjay Sood, Comparison, Grouping, and
Preference, 10 Psychology of Science 225 (1999).
complexities of decision making: frames 445
this example:71 Suppose that you are offered a choice between a recording of a
Beethoven symphony and Debussy piano works. You like the composers about
the same, so the probability of choosing one or the other is about 0.5. Now sup-
pose you are offered a choice among three recordings—the Debussy and two
versions of the Beethoven symphony. Is the probability of each choice now 0.33?
Regardless of the comparative excellence of the recordings, it’s closer to 0.25 for
each of the Beethoven’s and 0.50 for the Debussy. Most likely, you first consider
attributes that make you favor the Beethoven or Debussy recording at this time—
you’re in a German or French mood, or the sound of a symphony or of a piano
seems more appealing—and then, only if you choose Beethoven, do you com-
pare the two Beethoven discs.
Brenner et al. did some further experiments to see how grouping affected
people’s choices among options. Participants were asked to choose among four
items—for example four kinds of restaurants, four videotapes, or four gifts. In
each case, three of the items were grouped together and one was alone, and the
participants were asked to indicate their preference either for the solo option or
their choice of one of the three grouped options. For example, participants were
asked whether they would prefer to eat at (1) a seafood restaurant, or (2) their
choice of a Mexican, Italian, or Thai restaurant. Participants consistently deval-
ued options in the group of three restaurants compared to the solo item.
Apparently participants first compared the members of the group with each
other to choose a favorite, which they then compared to the option that stood
alone. “Comparative loss aversion implies that the greater degree of within-
group comparisons will sharply reduce the attractiveness of each of the grouped
options. However, because of the lesser degree of between-group comparisons,
the attractiveness of the lone option will not decrease as much.” The authors
observe that advertisers or salespersons can employ the grouping phenomenon
to manipulate consumers’ choices:
Consider choosing between three cars. One is a Japanese sedan, the second a
Japanese sports car, and the third is an American sports car. Our findings
suggest that the American car is more likely to be chosen when the cars are
grouped by country of origin (because it is the only American option) and less
likely to be chosen when the cars are grouped by body style (because it is in
the sports-car group). Similarly, the Japanese sedan is more likely to be chosen
when the cars are grouped by body style and less likely to be chosen when the
cars are grouped by country of origin. . . . [S]alespeople have been known to
emphasize the relative merits of Japanese and American cars, or the virtues
of sedans and sports cars, in an attempt to guide the consumer toward a par-
ticular purchase.
Along similar lines, Donald A. Redelmeier and Eldar Shafir asked physicians
to decide whether, after unsuccessfully trying a number of anti-inflammatory
medicines on a patient with osteoarthritis in his hip, to refer the patient for hip-
replacement surgery or try different medication. Physicians given the option to
prescribe one of several drugs that had not previously been tried on the patient
were more likely to choose surgery than those told that there was only one untried
drug.72
14.5.2.a The Downside of Having Too Many Choices Sheena Iyengar and
Mark Lepper have studied how consumers dealt with different numbers of
choices of food products.73 For example, customers entering an upscale food
market on different days encountered a display inviting them to taste either six
(limited choice) or twenty-four (extensive choice) different Wilkin & Sons pre-
serves. Anyone who approached the display received a $1 certificate toward the
subsequent purchase of a Wilkin & Sons preserve from the store’s regular
shelves. Both groups of customers actually tasted about the same number of
samples, but those who were offered the limited choice were much more likely
to purchase the product. A similar experiment, involving Godiva chocolates,
indicated that although participants enjoyed choosing from the extensive choice
options, those given the limited choice liked their chocolate better than those
who chose from the large sample. Those in the extensive choice group felt more
responsible for their choices, more dissatisfaction, and more regret about their
choices. Iyengar and Lepper speculate that they were relatively unsure about the
choice and “burdened by the responsibility of distinguishing good from bad
decisions.”
There is reason to believe that this phenomenon extends to more important
decisions than the choice of gourmet foods. Many employers offer 401(k) retire-
ment savings plans, some of which allow the employee to invest in just a few
funds, some in many. A large-scale survey across many industries indicated that
75 percent of employees participated in retirement savings plans that offered two
funds, while only 60 percent participated in plans with fifty-nine funds.74
72. Donald A. Redelmeier and Eldar Shafir, Medical Decision Making in Situations that
Offer Multiple Alternatives, 274 Journal of the American Medical Association 302 (1995).
73. Sheena S. Iyengar and Mark R. Lepper, When Choice Is Demotivating: Can One
Desire Too Much of a Good Thing?, 76 Journal of Personality and Social Psychology
995–1006 (2000).
74. How Much Choice Is Too Much: Determinants of Individual Contributions in 401K
Retirement Plans, in Developments in Decision-Making Under Uncertainty:
Implications for Retirement Plan Design and Plan Sponsors (O. S. Mitchell and S.
P. Utkus eds., Oxford: Oxford University Press, 2004).
complexities of decision making: frames 447
Barry Schwartz, Hazel Markus, Alana Connor Snibbe, and Nicole Stephens
describe a series of experiments concerning whether people of different socio-
economic status placed different values on choice.75 Among their conclusions:
Both preferences for, and the benefits of, choice might well vary across differ-
ent cultural contexts. . . . [W]hereas individual agency is an essential element
of the self-constructs of American individualists, it may be considerably less
relevant to the self-constructs of members of more collectivistic cultures,
characteristic of Asia and elsewhere.
Westerners . . . [may] possess a model of the self as fundamentally inde-
pendent. Such individuals strive for personal independence, desire a sense of
autonomy, and seek to express their internal attributes in order to establish
their uniqueness from others within their environments. . . .
By contrast, members of non-Western cultures . . . [may] possess an inter-
dependent model of the self and to strive for the super-ordinate goal of a
sense of interconnectedness and belongingness with one’s social in-group. . . .
Indeed, in some situations, the exercise of personal choice might even pose a
threat to individuals whose personal preferences could prove at variance with
of their reference group. Interdependent selves, therefore, might actually
prefer to have choices made for them, especially if the situation enables them
75
75. See Barry Schwartz, Hazel K. Markus, and Alana Connor Snibbe, Is Freedom Just
Another Word for Many Things to Buy, New York Times, Feb. 26, 2006. The SES-related
variable in the various studies included parents’ educational attainment and whether the
participants were in white- or blue-collar jobs.
448 problem solving, decision making, and professional judgment
76. Iyengar and Lepper, supra, citing Hazel Markus and Shinobu Kitayama, Culture
and the Self: Implications for Cognition, Emotion, and Motivation, 98 Psychological Review
224–53 (1991).
77. Jiwoong Shin and Dan Ariely, Keeping Doors Open: The Effect of Unavailability on
Incentives to Keep Options Viable, 50 Mgmt. Sci. 575, 575 (2004).
78. Daniel T. Gilbert and Jane E. J. Ebert, Decisions and Revisions: The Affective Forecasting
of Changeable Options, 82 Journal of Personality and Social Psychology 503 (2002).
complexities of decision making: frames 449
to buy. You pass by a store that is having a one-day clearance sale. Consider these
scenarios and the percentage of responses:79
Low conflict: The store offers a popular Sony player for $99—well below the
list price. Do you:
Buy the Sony player ($99) 66%
Wait until you learn more about the model 34%
High conflict: The store offers a popular Sony player for $99 and a top-of-the-line
AIWA player for $169—both well below the list price. Do you:
Buy the AIWA player ($169) 27%
Buy the Sony player ($99) 27%
Wait until you learn more about the models 46%
While the first decision seems easy, the second creates a mental conflict,
which can be avoided by deferring. But choosing in one case and deferring in the
other does not make sense in terms of a value-maximizing approach to decision
making, for the decision maker has precisely the same information about the
Sony player in both cases. But the phenomenon is easy to understand in terms
of reason-based decision making. While there are compelling arguments for
choosing the Sony when it is the only attractive product one is considering,
“choosing one out of two competing alternatives can be difficult: the mere fact
that an alternative is attractive may not in itself provide a compelling reason for
its selection, because the other option may be equally attractive. The addition of
an alternative may thus make the decision harder to justify, and increase the
tendency to defer the decision.”80
The phenomenon applies to public policy as well as personal decisions. In a
study, legislators were asked to recommend whether to close a hospital. Hospital
A was a 100-bed community hospital providing 200 jobs in a small urban area
which was also served by two other hospitals, which had a higher quality of care
and lower costs. Hospital B was a 250-bed teaching hospital providing 500 jobs
in a large urban area served by three other hospitals, which also had a higher
quality of care and lower costs. In both cases, the legislators were told that they
could also make “no recommendation.” When asked only whether to close
Hospital A, only 26 percent declined to make a recommendation. When asked
whether to close either Hospital A or Hospital B were options, 64 percent
declined to make a recommendation.81
The obvious solution to avoiding these odd changes of preferences is to com-
pare all three options (including the status quo of doing nothing) using a value-
based approach like the subjective linear model.
79. Amos Tversky and Eldar Shafir, Choice under Conflict: The Dynamics of Deferred
Decision, 3 Psychological Science 358–61 (1992).
80. Ibid.
81. Donald A. Redelmeier and Eldar Shafir, Medical Decision Making in Situations that
Offer Multiple Alternatives, 274 Journal of the American Medical Association 302 (1995).
450 problem solving, decision making, and professional judgment
82. Itmar Simonson and Amos Tversky, Choice in Contrast: Tradeoff Contrast and
Extremeness Aversion, 29 Journal of Marketing Research 281 (1992).
83. Simonson and Tversky. Only 2 percent of participants chose the inferior pen, thus
confirming its inferiority.
84. Mark Kelman, Yuval Rottenstreich, and Amos Tversky, Context-Dependence in
Legal Decision Making, 25 Journal of Legal Studies 287 (1996).
complexities of decision making: frames 451
85. Id.
86. Simonson and Tversky.
452 problem solving, decision making, and professional judgment
seventeen-year-old daughter from her first marriage say that the deceased had,
once again, tried to sexually molest her. The evidence also showed that the defen-
dant stood to inherit a large amount of money from her husband and that she
had been involved with another man for the last half year. One group of partici-
pants was given the choice (that jurors would have had in the District of Columbia)
between convicting her of manslaughter (homicide under extreme emotional
disturbance) or murder. The results:
• Manslaughter (extreme emotional disturbance) 47%
• Murder 53%
A second group was given the choice (that jurors would have had in California)
among convicting her of manslaughter, murder, or special circumstances
murder (committed for financial gain, or in an exceptionally heinous fashion, or
by the administration of poison). The results:
• Manslaughter (extreme emotional disturbance) 19%
• Murder 39%
• Special circumstances (financial gain, heinous, poison) 42%
But if the defendant was acting under emotional distress, she didn’t commit
murder at all, with or without special circumstances. As the authors note, the
choice between manslaughter on the one hand, and the murder verdicts on the
other, should not be sensitive to whether, if the defendant were acting deliber-
ately, she killed for financial gain, etc.
Another scenario involved the intentional killing of an off-duty police officer
working as a security guard at a shopping mall by an African American man he
suspected of burglary. When the defendant refused to be patted down, the guard
grabbed him, the guard shouted a racist epithet, and the defendant shot him.
Participants were informed that there were four possible verdicts:
• special circumstances murder—for the killing of a police officer acting in
the line of duty;
• murder;
• voluntary manslaughter—if the defendant was provoked or acting under
extreme emotional disturbance; or
• involuntary manslaughter—if the defendant subjectively, but unreason-
ably, believed he was entitled to defend himself.
Participants in one group were told that the judge had ruled, as a matter of
law, that there was no evidence that the defendant believed he was acting in self-
defense, so that they could not treat the homicide as involuntary manslaughter.
The results were:
• Special circumstances murder 12%
• Murder 57%
complexities of decision making: frames 453
such manipulations are common in the marketplace. They suggest that harm
befalls context-dependent consumers since someone manipulated into choos-
ing the option favored by another party, with her own set of interests, is less
likely to maximize his own well-being.
Although there is reason for skepticism whether people can be educated to
avoid context-dependent decisions, it seems plausible that eliminating explicit
irrelevant options will reduce context-dependent decisions. How and when that
can be done depends on, well, the context.
89. Richard Revesz and Michael Livermore, Retaking Rationality: How Cost
Benefit Analysis Can Better Protect the Environment and Our Health, 127
(New York, Oxford University Press, 2008).
456 problem solving, decision making, and professional judgment
90. Id.
91. Irwin et al., supra.
92. Daniel Kahneman and Shane Frederick, A Model of Heuristic Judgment, in Keith
Holyoak and Robert Morrison eds., Cambridge Handbook of Heuristic Reasoning
(New York: Cambridge University Press, 2005).
93. See Daniel Kahneman, Ilana Ritov, and David Schkade, Economic Preferences or
Attitude Expressions? An Analysis of Dollar Responses to Public Issues, in Choices, Values,
and Frames 642–672 (D. Kahneman and A. Tversky eds., New York: Cambridge
University Press, 2000); Christopher K. Hsee and Yuval Rottenstreich, Music, Pandas, and
Muggers: On the Affective Psychology of Value, 133 J. Experimental Psychol.: Gen. 23,
23–24 (2004).
complexities of decision making: frames 457
94. http://en.wikipedia.org/wiki/Contingent_valuation.
95. For a discussion of alternative procedures for assessing people’s values for non-
market goods, see Jonathan Baron, Thinking and Deciding (4th ed. Cambridge, UK:
Cambridge University Press, 2000).
This page intentionally left blank
15. decision making under risk
EU(E)= U(E)×P(E)
EU(E) is the expected utility of event E. U(E) is the subjective utility of the event.
P(E) is the probability of the event’s occurrence.
We begin with the simplifying assumption that the decision maker is risk
neutral. This means that the decision maker is indifferent between a payoff with
certainty and a gamble with the same expected payoff, for example, between
• a 100 percent chance to get $100, or
• a 50 percent chance to get $200 and a 50 percent chance to get nothing.
If a decision maker is risk neutral, then the expected utility (U(E)) of an event is
exactly the same as the expected value (V(E)) of the event—that is, its monetary (or
monetizable) value. Later, we address the fact that decision makers are often not
risk neutral, but may have different preferences or tolerances for risks in differ-
ent circumstances. Under these circumstances, expected utility and expected
value will diverge.
460 problem solving, decision making, and professional judgment
The axioms of rational decision making set out in Section 12.2.2 apply here as
well. We shall discuss their application to decision making in conditions of risk and
introduce some new axioms that bear on particular issues presented in this chapter.
Expected utility under the probability axiom. Underlying all the problems in this
chapter is the assumption—actually, another axiom of rational decision mak-
ing—that one can assign a probability to every possible outcome. In this context,
a probability is simply a degree of belief that the outcome will occur. (See Section
5.1.2b) Sometimes the decision maker may arrive at the probability through the
statistical analysis of a large sample. More often, however, she must use a sub-
jectivist, Bayesian approach and, particularly in the case of one-time decisions,
she can only hazard a guess about the probability. Be that as it may, every out-
come, in principle, has a probability somewhere between 0.0 and 1.0.
With the probability axiom as background, the essential assumptions about
completeness, dominance, and trade-offs that we discussed in Section 12.2.2 apply
to decisions under risk. To evaluate alternatives, you multiply the utility of each
possible outcome by the probability of its occurring to determine the expected
utility of that alternative. If the expected utility of alternative number 1 is greater
than that of alternative number 2, then the former dominates the other. Let’s
look at some decision tools that help guide decisions under risk.
Many decisions under conditions of risk involve causal chains that are not read-
ily modeled by a matrix of the sort that Christine Lamm used for the wastewater
treatment plant decision in Section 4.5. Two conventional means of modeling
them are dependency diagrams and decision trees. These tools actually serve two
purposes. First, quite apart from any quantitative analysis, they can assist in
structuring a decision—in laying out the sequence of the decision, identifying
uncertainties, specifying relationships between issues, and describing the out-
comes associated with different scenarios. Second, if one attaches probabilities
and values to the various nodes, they can compute expected value.
We begin this section with very simple examples showing how a dependency
diagram (also called an influence diagram) can help structure a decision and how a
decision tree can be used to compute expected value. (Actually, the dependency
diagram is an even more powerful computational tool than the decision tree, but the
computations are less transparent to the user.) A dependency diagram shows causal
links where an event or decision affects another event, decision, or outcome.
if rain is likely
carry umbrella
The differently shaped objects are called nodes. Your satisfaction or utility
(diamond-shaped node) depends on the relationship of two variables: your deci-
sion to carry (or not carry) an umbrella (rectangular node) and whether it rains
(oval node). You will be
• very dissatisfied if it rains and you do not have an umbrella, and
• mildly dissatisfied if you carry the umbrella around unnecessarily.
Your decision whether to carry an umbrella is based on the National Weather
Service (NWS) forecast.
The arcs, or lines, of the dependency diagram show the relations among these
elements. The presence of an arc indicates that there is some possibility that the
node pointed to is dependent on the node from which the arc originates. The
absence of an arc is, if anything, more informative than its presence: its absence
asserts that the nodes that are not connected are, in fact, independent. There is
no arc between the forecast and your satisfaction, for the only role of the forecast
is to inform your decision whether to carry an umbrella (which will affect your
satisfaction).
Were you to make the decision based on a combination of the NWS forecast
and your own forecast, the dependency diagram might look like Figure 15.2.
Your Forecast
if rain is likely carry
umbrella Rain or sun
This shows that both the NWS weather forecast and your own forecast rely on
what the weather has been in the past. It shows that you take account of the NWS
forecast, but that NWS doesn’t take account of yours. Since today’s actual weather
(rain or sun) also depends on past weather, we draw those arcs as well.
Even without using a dependency diagram quantitatively, it has value as a
checklist of all the factors that should affect your decision—and equally important
it identifies factors present on the landscape that do not affect your decision.
rain
don’t carry umbrella
sun
Both the decision nodes and the outcome nodes of a decision tree are mutu-
ally exclusive and collectively exhaustive (MECE). For example, we have posited
that there are only two possible outcomes—rain or sun—and if one occurs the
other cannot. But in fact, there are many other possible states of the weather,
ranging from snow, sleet, and hail to partly cloudy with occasional showers, and
we could represent all of them in the decision tree. A decision tree can have as
many outcomes as you wish—as many as may be relevant to the decision at
hand—but in aggregate they are MECE.
In a simple case like this one, the bare tree does not advance your understand-
ing of the decision. The tree’s power becomes more evident, however, when we
take a further step and add two sorts of numbers to the tree.
One set of numbers is the probability that it will (or will not) rain. Let us sup-
pose that the weather forecaster says there’s a 40 percent chance of rain, which
(following the conventions described in Chapter 5) we’ll write as 0.4. Because
rain and sun are MECE, the probability of sun is 0.6.
The second set of numbers indicates how satisfied or dissatisfied you will be with
the outcome. Let’s use a 100 point scale, and (since we’re talking mainly about dis-
satisfaction) we’ll treat 0 as your being perfectly happy if the sun shines and you
don’t have an umbrella, and –100 as being miserable if it rains and you don’t have
decision making under risk 463
an umbrella. You don’t like either rain or carrying umbrellas. Each independently
gives you a (dis)utility of -30. So if you are carrying the umbrella and it’s a sunny day,
your utility is down -30, and if it rains it’s down -60. (See Figure 15.4.)
rain
−60
carry umbrella 0.4
sun
−30
0.6
rain
−100
don’t carry umbrella 0.4
sun
0
0.6
For each branch of the tree, you calculate the expected utility by multiplying
the probability of rain (or shine) by the satisfaction of the outcome, and sum the
products as shown in Table 15.1.
table 15.1 solving the decision tree
Probability Outcome EU
−42
Sun 0.6 0 0
−40
The total expected utility of carrying an umbrella is -42, while the utility of not
carrying the umbrella is -40. What does the imply if you are risk neutral? Would
you in fact follow its implications?
working through the scenario of a trial. After all, while Evers needs the money,
the amount for which he would settle—his “reservation price”—will be influ-
enced by what he could gain (or lose) by litigating the case.
In predicting the outcome of the trial, these factors seem relevant:
• If the case goes to trial, Evers’ costs of litigation will be $100,000.
• Evers made extravagant oral representations about how the software would
perform, and he acknowledges to the lawyers that its performance is not
entirely what he said it would be.
• The contract disclaims all warranties of performance. It also has a so-called
“integration clause,” which says that the written contract is the sole agree-
ment between the parties, notwithstanding any oral representations.
• The law requires that warranty disclaimers be in a particular typeface, some-
what different from the disclaimer in the contract with Newport Records.
The judge might or might not regard the difference as immaterial.
• If the judge finds that the disclaimer’s typeface deviates materially from the
statutory requirements, he will instruct the jury to find the disclaimer effective
only if the president of Newport Records actually read and understood it.
• If the disclaimer is ineffective, and notwithstanding the integration clause,
the judge may permit the jury—as part of its determination of whether the
software performed “as promised”— to hear about Evers’ oral representa-
tions, in order to resolve an alleged ambiguity in the contractual descrip-
tion of the software’s function.
Trujillo suggests that it would be helpful to model these factors in a depen-
dency diagram, to be sure to understand how they combine to affect the outcome
of the trial, as shown in Figure 15.5.
With the dependencies thus clear, the lawyers construct the skeleton of a deci-
sion tree. Although they may refine this analysis later, the applicable contract law
indicates that Evers will either obtain a judgment for the $600,000 due under the
contract, or get nothing. This is shown by the outcome values on the right side
(see Figure 15.6). (Once we solve for the case on its merits, we will take into
account the $100,000 costs of litigating the case.)
Trujillo and Wilkins then discuss how to determine the likelihood of Evers’
prevailing (or losing). Rather than leap immediately to guessing probabilities, they
ask what reasons would lead the judge or jury to decide for or against Evers on
each issue. For example, whether the written disclaimer meets the statutory
requirements is a question of law to be decided by the judge, and they consider
the reasons why the judge might decide one way or the other, as shown in
Table 15.2.
If the judge determines that the disclaimer did not meet the statutory require-
ments because it was in the wrong typeface, the jury would determine whether
the president of Newport Records nonetheless understood that any warranty was
being disclaimed, as shown in Table 15.3.
outcome if
we litigate?
Reasons why the judge would find Reasons why the judge would find the
the written disclaimer does not meet written disclaimer meets the statutory
the statutory requirements requirements
After listing the pros and cons on the remaining issues, Trujillo and Wilkins
now begin to think hard about probabilities.1 They begin with qualitative judg-
ments, such as “likely” or “very likely,” but eventually realize they need to trans-
late these into numbers.
15.2.3.a Estimating probabilities. It is easy to estimate the probability of an
event’s occurring when there are sufficient instances that you can apply frequen-
tist statistics along the lines introduced in Chapters 5 and 6. But many decisions
faced by lawyers and policy makers are one-time events, susceptible only to a
subjective estimate. Trujillo’s and Wilkins’s strategy of articulating the reasons
why a particular outcome might or might not occur is often helpful background.
Here are some techniques—by no means mutually exclusive—for use in various
circumstances:2
1. See Marc B. Victor et al., Evaluating Legal Risks and Costs with Decision Tree Analysis
in Successful Partnering Between Inside and Outside Counsel ch. 12 (West Group
& ACCA 2000–2007).
2. See Douglas A. Wiegmann, “Developing a Methodology for Eliciting Subjective
Probability Estimates During Expert Evaluations of Safety Interventions: Application for
Bayesian Belief Networks,” final technical report, AHFD-05-13/NASA-05-4 (October 2005) 9
(Aviation Human Factors Division Institute of Aviation, University of Illinois at Urbana-
Champaign), http://www.humanfactors.uiuc.edu/Reports&PapersPDFs/TechReport/05-13.
pdf#search=%22probability%20wheel%22.
468 problem solving, decision making, and professional judgment
Reasons why the jury would find Reasons why the jury would find that
that Newport did not understand Newport understood the disclaimer
the disclaimer
• Ask how many times out of one hundred you would expect an event to
occur.
• Choose between a certain payoff (e.g., a settlement) or a gamble (litigate)
where the payoff depends on the probability in question, and adjust the
amount of the certain payoff until you are indifferent between the two
choices.
Of course, these elicitation methods depend on your actually having a (rea-
sonably good) estimate of the probability in your head. And they are subject to
the cognitive biases discussed in earlier chapters.
Trujillo and Wilkins are aware of the hazards of overoptimism and confirma-
tion bias (discussed earlier in the book). While they believe that their profes-
sional distance creates a degree of immunity from these biases, Wilkins suggests
that they err on the side of conservatism. But Trujillo responds that they would
do best just to be as realistic as possible—to avoid undervaluing Evers’ lawsuit.
(If Evers wants to be conservative, he can adjust his settlement position accord-
ingly.) The lawyers end up filling in the probabilities, as shown in Figure 15.7.
Finally, they roll back, or solve, the tree. The process involves multiplying the
outcome of each branch of the tree by the probabilities of achieving it, and adding
up the results. Go from top to bottom and right to left in Table 15.4.
Rather than do this manually, or even using Microsoft Excel, they use dedi-
cated software—in this case, a program called TreeAge, which happily performs
just as warranted. (See Figure 15.8.)
Based on their best estimates of the probability of winning on each issue, the
expected value of the case is $420,000. But it will cost $100,000 to litigate, so the
net value is reduced to $320,000.
judge finds our
warranty disclaimer
to be effective,
as a matter of law
{1} $600,000
.20
jury finds defendant
read and understood
that any warranty of
performance was
IF LITIGATE being disclaimed
{2} $600,000
.10
judge finds this in light of the contract’s
is an issue for the “integration clause,” jury finds software
jury, in light of judge excludes Evers’ performed as promised
incorrect typeface oral representations {3} $600,000
2/3
regarding performance
.80
jury finds defendant did jury finds software did
.75
not read or understand not perform as promised
the warranty disclaimer, {4} $0
1/3
What size settlement should Evers be willing to accept? Assuming for the
moment that he is risk neutral (more about that later), he might take into account
these considerations:
• If the case goes to trial, it is not likely to be resolved for a couple of years.
But Evers is eager to have funds to start a new business.
• Also, any sum received today is worth more than the same amount received
a couple of years in the future.3 (See Section 13.6.1).
• Evers’ lawyers estimate that their fee for their engaging in settlement nego-
tiations will be $5,000.
• And, of course, it’s possible that Newport Records will not agree to settle
for anything that Evers deems acceptable.
All things considered, Evers and his lawyers conclude that a settlement of
about $300,000 would be acceptable.
15.2.3.b Sensitivity analysis. If the disclaimer fails to meet the statutory
requirements, the evidence bearing on whether the president of Newport Records
read and understood the warranty disclaimer will be based on his deposition and
testimony at trial. Other than being shrewd cross-examiners, there is not much
that Evers’ lawyers can do to affect that.
But they may be able to affect the jury’s judgment about whether the software
performed as promised. Perhaps, if understood in the context of the reasonable
needs of a small business and prevailing standards for accounting systems, Evers
may not have egregiously overstated the capabilities of his product. But estab-
lishing the prevailing norms will require expert testimony. Based on preliminary
conversations, the expert they have in mind (a business school professor) would
likely conclude that under the prevailing standards the software materially met
the specifications. But she will charge $20,000 to do the analysis to establish this
and then testify.
Is it worth hiring the expert witness? That depends on how her testimony
will affect the value of the case. And that, in turn, depends on how much her
figure 15.8 solving the decision tree in newport records with a tree program.
472 problem solving, decision making, and professional judgment
testimony will increase the likelihood of persuading the jury on that issue. Right
now, the lawyers estimate this likelihood as being 0.67 if the judge excludes
Evers’ oral representations, and 0.33 if he allows them into evidence. The
lawyers believe that the expert’s testimony would increase the likelihood in the
former case to 0.80, and in the latter case to 0.20.
The decision tree makes it easy to determine whether it is worth the cost of
the expert witness. Solving the tree with these increased probabilities produces
an expected value of $470,400. Since this is an increase in case value of over
$50,000 for a cost of only $20,000, the investment in the expert witness would be
well worth her cost.
15.2.3.c The probability of different jury awards. We simplified the problem
by assuming that the jury would either award $600,000 or nothing. But it is easy
to imagine reasons that the jury would award Evers some amount in between.
For example, although he claimed that the software would do more than it did, it
performed reasonably well, and Newport Records was not justified in refusing to
pay anything—especially if it actually continued using the software. At this point,
you have the concepts and tools for factoring this in as well.
figure 15.9 the last word on expected value analysis. dilbert © scott
adams/dist. by united feature syndicate, inc. reprinted by permission.
So far, we have assumed that Evers is risk neutral in approaching his decision
whether to litigate or settle the case against Newport Records. But this needn’t be
the case. Evers may prefer to accept a settlement lower than the expected value of
the litigation, either because he wants funds now to start his new venture or just
as a matter of personal preference. Or he might wish to take a gamble on the
litigation even if the expected value happens to be less than that of a settlement.
There are essentially three conditions of risk tolerance (or risk attitude, or risk
preference). We’ll illustrate these with a simple example.
4. Gerd Gigerenzer, Peter M. Todd, and the ABC Research Group, Simple
Heuristics That Make Us Smart 9 (New York: Oxford University Press, 1999).
474 problem solving, decision making, and professional judgment
Suppose you are offered the choice between a certain $100 and a 0.5 chance
of winning $200 and a 0.5 chance of winning nothing. (This kind of choice,
referred to as a “gamble” or “lottery,” is used to determine a variety of individual
preferences involving risk.)
• If you are indifferent between the choices, you are risk neutral.
• If you would demand more than a $200 payoff, or a greater than 0.5 chance
of winning, before you accepted the gamble, you are risk averse. The addi-
tional amount of payoff you would demand is the risk premium.
• If you would take the gamble for less than a $200 payoff, or a smaller than
a 0.5 chance of winning, you are risk seeking.
If you are risk neutral in this situation, your (subjective) expected utility of the
certain $100 or the .05 chance of winning $200 is identical to the (objective)
expected value. But if you are risk averse or risk seeking, your expected utility
will diverge from expected value. Let’s consider your utility function if you are
risk averse.
Recall Bernoulli’s observation about the declining (subjective) marginal value
of money (Section 12.1.4.b). The subjective value of increasing your wealth from
$100 to $200 is greater than the subjective value of increasing your wealth from
$1100 to $1200. A graph of this “utility function” would be downwardly concave.
Bernoulli’s curve describes risk aversion as well as the declining marginal
utility of money: If you prefer a sure gain of $100 (lower on the curve) over a 0.5
chance of $200 (higher on the curve), then you are willing to pay a bit more (and
sacrifice some expected value) for the certainty. The extra amount you are willing
to pay is the risk premium. Risk aversion and the declining marginal utility of
money are two sides of the same coin. When we describe the declining marginal
utility of money, we say that increments to an already large amount bring less
utility than the equivalent increments to a small amount. When we describe risk
aversion, we say that a bird in the hand is worth two in the bush.
The same person may have a different risk preference in different situations
(e.g., finance, health and safety, social situations, and recreation).5 “Individuals
who exhibit high levels of risk-taking behavior in one content area (e.g., bungee
jumpers taking recreational risks) can be risk averse in other domains (e.g.,
financial).”6
And different people may have different risk preferences in the same situation.
If Alex and Andrea exhibit different risk preferences for flying on airlines with
poor safety records, it could be because they have different perceptions of the risk
or of the benefits from taking the risk, or differences in their willingness to take
on a (perceived) risk.7 While risk taking may be affected by the perceived benefit
of the activity,8 differences in risk preferences often result from different percep-
tions.9 (For example, entrepreneurs and people who enter crowded markets
that offer small chances of financial success tend to be highly optimistic.10) Apart
from its academic interest, the reasons for differences in risk preferences may
have practical consequences. For example, if you are negotiating in a situation
where the allocation of risks between the parties is an issue, understanding and
addressing the source of the other party’s risk preferences may help reach an
agreement.11
In any event, one cannot, a priori, describe any of these risk preferences as
more or less rational than the others. In many situations, risk preferences are as
subjective, and hence as arational, as preferences for flavors of ice cream.
Sometimes they make perfect sense in terms of your specific goals: If risk seek-
ing behavior seems irrational to you, suppose that you are in Topeka, with no
money in your pocket, and eager to take the next flight to New York to attend a
friend’s birthday party that night. A plane ticket costs $150, and a benefactor
offers you a certain $100 or a 0.5 chance of winning $150.
Even if risk preferences often are arational, we will see later in the chapter
that they can be affected simply by the way choices are framed. In these cases,
there is at least cause for concern that the frame-influenced preferences do not
always serve an individual’s best interests.
Policy Problem
Should policy makers be risk averse, risk seeking, or risk neutral? Does the answer
depend on the circumstances and, if so, what might they be?
In Section 14.1 we considered the status quo bias, including the endowment
effect, in decisions where risk is not at issue. Here, we return to the same
phenomenon in conditions of risk. In addition to the three features of prospect
theory noted above—
1. the value function is defined over gains and losses with respect to some
reference point;
2. individuals are loss averse (the slope of the curve for losses is considerable
steeper than the slope for gains);
3. both the gain and loss functions display diminishing sensitivity as one moves
away from the reference point;
—we examine two other features that are relevant to risky choices, particularly:
4. while the utility curve for gains with respect to the status quo is concave
(implying risk aversion), the utility curve for losses is convex (implying risk
seeking);
5. individuals overweight very low probabilities and underweight very high
probabilities, so that (reversing the phenomenon generally described by
the utility function) they tend to be risk seeking with respect to low probability
gains and risk averse with respect to unlikely losses.
The fact that the reference point separating gains and losses for a particular
decision is indeterminate and, indeed, manipulable, poses a severe problem for
expected utility theory. In what has become a classic experiment, Amos Tversky
and Daniel Kahneman gave one group of subjects the following problem (the
percentage of responses to each question is shown in parentheses):
Imagine that the United States is preparing for the outbreak of an unusual
virus that is expected to kill 600 people. Two alternative programs to combat
the disease have been proposed. Assume that the exact scientific estimates of
the consequences of the programs are as follows:
12. Amos Tversky and Daniel Kahneman, The Framing of Decisions and the Psychology
of Choice, 211 Science 453 (1981). David Mandel, Gain-Loss Framing and Choice: Separating
Outcome Formulations from Descriptor Formulations, 85 Org. Behavior and Human
Decision Processes 56 (2001), argues that the formulation of the problem may be ambig-
uous and that this and other factors may confound the framing effects.
13. Much of the following analysis is based on Chris Guthrie, Prospect Theory, Risk,
Preference, and the Law, 97 Nw. U. L. Rev. 1115 (2003).
478 problem solving, decision making, and professional judgment
to pay a certain settlement to the plaintiff or to gamble that further litigation will
reduce the amount they must pay. Thus, plaintiffs generally choose between
options that appear to represent gains, while defendants generally choose
between options that appear to represent losses.”
Economic theory predicts that their different perspectives will have no effect
on plaintiffs’ and defendants’ decisions whether to litigate or settle. Given
options with the identical expected value, a risk-neutral litigant would be indif-
ferent and a risk-averse litigant would prefer to settle. Overoptimism or the pur-
suit of nonmonetary values other than wealth (e.g., vengeance, vindication, or
stubbornness) might lead to suboptimal decisions to litigate, but theoretically
there is no reason to believe that these affect plaintiffs and defendants asym-
metrically.
In fact, however, plaintiffs tend to be risk-averse and defendants tend to be
risk-seeking. This is suggested by actual settlement patterns14 and is the robust
finding of many experiments. For example, Jeffrey Rachlinski asked students to
play the roles of parties in a copyright case.15 The plaintiffs had a choice
between:
• accepting the defendant’s $200,000 settlement offer, or
• going to trial with a 50 percent chance of winning $400,000 and a 50 per-
cent chance of winning nothing.
The defendants had a choice between:
• paying the plaintiff $200,000 in settlement, or
• going to trial with a 50 percent chance of losing $400,000 at trial and a 50
percent chance of losing nothing.
Although expected value of all of the options is identical, 77 percent of the plain-
tiffs preferred settlement, while 69 percent of the defendant chose to go to trial.
In another experiment, participants were assigned the role of the plaintiff or
defendant in litigation between a rural bed and breakfast inn (B&B) and an adja-
cent property owner.16 The B&B had unintentionally expanded to occupy a small
corner of the adjacent property, worth about $50. After unsuccessful negotia-
tions, the property owner sued for an injunction to require the B&B to remove
the encroaching structure. The judge can either order the B&B to remove the
14. Jeffrey Rachlinski, Gains, Losses and the Psychology of Litigation, 70 So. Cal. L. Rev.
113 (1996); Guthrie, Prospect Theory, supra; Russell Korobkin and Chris Guthrie, Psychology,
Economics, and Settlement: A New Look at the Role of the Lawyer, 76 Texas L. Rev. 77
(1997).
15. Rachlinski, supra.
16. Rachlinski, supra. Rachlinski did a number of versions of the experiment with dif-
ferent numbers.
decision making under risk 479
structure at considerable cost, or order the adjacent owner to sell the corner of
his property for about $50. If the judge orders removal of the structure, then
rather than tear it down, the B&B will offer the owner $100,000—an offer that
the owner is certain to accept. In other words, depending upon the judge’s deci-
sion, the plaintiff will win, and the defendant will lose, either $100,000 or $50.
The parties are informed that the judge assigned to the case is an adamant
defender of property rights who opposes forced sales of land and is very (70 per-
cent) likely to order removal of the buildings. It is now one day before trial. The
opposing party is willing to settle for $70,000.
In running this experiment with various stakes and probabilities, 81 percent
of the plaintiffs settled, while only 45 percent of the defendants settled. While the
defendants were mildly risk seeking, the plaintiffs were highly risk averse.
The reference point from which any decision maker assesses gains and losses
is subjective. An experiment by Russell Korobkin and Chris Guthrie shows how
a plaintiff’s reference point may change in almost identical cases.17 Participants
played the role of plaintiffs, who had been slightly injured and whose cars were
totaled in a car accident, and who had a choice between accepting a settlement of
$21,000 or going to trial with a 50 percent chance of an award of full damages of
$28,000 and a 50 percent chance of receiving only $10,000 (the insurance limit)—
an expected value of $19,000. The first group of subjects was informed that
$14,000 of the damages were for medical costs that had already been reimbursed
by health insurance and the other $14,000 was the full value of their car. The
second group was told that $4,000 were for (reimbursed) medical costs and the
remaining $24,000 was the value of their car. For subjects in the first group,
the settlement reflected a net gain and they were risk averse, overwhelmingly
preferring to settle. For subjects in the second group, the settlement represented
a loss, and they tended to choose to go to trial.
Not just partisans, but neutrals may be affected by how the issue is framed.
Chris Guthrie, Jeffrey Rachlinski, and Andrew Wistrich presented a version of the
copyright infringement case described above to judges presiding over settlement
conferences, who were asked whether to recommend settlement.18 They were
informed that there was a 50 percent chance that the plaintiff would prevail at trial
and win $200,000 and a 50 percent chance that she would not win anything at
all—and that it would cost each party $50,000 in attorneys’ fees to go to trial.
Half of the judges reviewed the case from the plaintiff’s perspective: Should
the plaintiff accept defendant’s offer to pay the plaintiff $60,000 to settle the
case? Those judges thus had a choice between
17. Russell Korobkin and Chris Guthrie, Psychological Barriers to Litigation Settlement:
An Experimental Approach, 93 Mich. L. Rev. 107 (1994).
18. Chris Guthrie, Jeffrey J. Rachlinski, Andrew J. Wistrich, Inside the Judicial Mind,
86 Cornell Law Review (May 2001).
480 problem solving, decision making, and professional judgment
Appendix A (Section 15.5) discusses how people’s different risk attitudes toward
gains and losses can violate the Dominance Axiom.
20. Maurice Schweitzer, Lisa Ordóñez, and Bambi Douma, Goal Setting as a
Motivator of Unethical Behavior, 47 Academy of Management Journal 422 (2004);
Maurice Schweitzer, Lisa, Ordóñez, and Bambi Douma, The Dark Side of Goal Setting:
The Role of Goals in Motivating Unethical Decision Making, 47 Academy of Management
Proceedings 422 (2002).
21. Chris Guthrie, Prospect Theory, Risk Preference & the Law, 97 Northwestern
University Law Review 1115, 1140 (2003).
22. Jeffrey J. Rachlinski, Gains, Losses, and the Psychology of Litigation, 70 S. Cal. L. Rev.
113, 118 (1996).
482 problem solving, decision making, and professional judgment
Two groups were asked whether they would reveal the documents before
going forward with settlement negotiations. The problem was presented to one
group in a “gains” frame: The company had expected to have to pay the parents
$5 million to settle the case and therefore thought that the case was “going well.”
The other group was given a “loss” frame: The company had originally expected
to pay the parents only $1 million and believed that the case was “going poorly.”
Only 12.5 percent of the subjects who saw the settlement as a gain chose to settle
without disclosing the documents, while 45 percent who saw the settlement as a
loss indicated that they would settle with disclosing the documents. By contrast
to the risk aversion manifested by those in gains condition, those in the loss
condition opted for the riskier—and ethically and legally problematic—course of
action rather than face a sure loss.
Suppose that Desmond and Edwina are predicting whether their respective cli-
ents will prevail as plaintiff in a civil suit. Desmond has no personal experience
with this sort of litigation, but is reliably informed that plaintiffs in this jurisdic-
tion win 10 percent of the time. Edwina handles many such cases as a plaintiff’s
attorney and wins 10 percent of the time. Who is more likely to predict that her
client will win? Consistent with the prospect theory, Desmond will overweight the
10 percent probability in making his prediction. However, Edwina may actually
underweight the probability. Her estimate is likely to be affected by the availability
of the most recent events in her experience—and, by the nature of probability, the
one win out of ten is not likely to be one of the most recent occurrences.24
24. Elke U. Weber, “Origins and Functions of Risk Perception, Center for the Decision
Sciences” (presented at NCI Workshop on Conceptualizing and Measuring Risk
Perceptions, February 13–14, 2003), http://dccps.nci.nih.gov/brp/presentations/weber.p
df#search=%22Weber%2C%20Origins%20and%20Functions%20of%20Risk%20Perce
ption%20nci%22.
25. Amos Tversky and Daniel Kahneman, The Framing of Decisions and the Psychology
of Choice, supra.
484 problem solving, decision making, and professional judgment
EV % Choosing
C 25% chance to win $300 75% chance to win nothing $75 42%
D 20% chance to win $450 80% chance to win nothing $90 58%
26. Stephen Breyer, Breaking the Vicious Circle: Toward Effective Risk
Regulation (Cambridge: Harvard University Press, 1995).
decision making under risk 485
Appendix B (Section 15.6) discusses the Certainty Effect and Violation of the
Independence Axiom.
EV % Choosing
27. Jonathan Baron, Thinking and Deciding 255 (3rd. ed. New York: Cambridge
University Press, 2001). The respondents apparently ignored the value of seatbelts in pre-
venting serious injuries.
486 problem solving, decision making, and professional judgment
choosing D, the majority of subjects in the two-stage game made the risk averse
choice of E.
Tversky and Kahneman label this phenomenon the pseudocertainty effect
because an outcome that is actually uncertain is weighted as if it is certain. The
framing of probabilities as a two-stage game encourages respondents to apply
cancellation: the event of failing to reach the second stage is discarded prior to
evaluation because it yields the same outcomes in both options.28 They suggest
that the choice between E and F has a greater potential for inducing regret than
the choice between C and D.
A simple experiment by Paul Slovic and his colleagues demonstrates how the
pseudocertainty effect can affect public health decisions. One group was told
that a disease will afflict 20 percent of the population, but that a vaccine will
reduce their risk of getting the disease by a half. A second group was told that
there are two strains of the same disease, that each will afflict 10 percent of the
population, and that a vaccine will entirely eliminate the risk of getting one strain
but have no effect on the other. Although each will reduce an individual’s risk of
getting the disease from 20 percent to 10 percent, many more people (57 per-
cent) in the second group than in the first group (40 percent) said they were
willing to be vaccinated.29
Tversky and Kahneman replicated the two-stage game described above in
these questions given physicians attending a meeting of the California Medical
Association:30
In the treatment of tumors there is sometimes a choice between two types of
therapies: (i) a radical treatment such as extensive surgery, which involves
some risk of imminent death, (ii) a moderate treatment, such as limited sur-
gery or radiation therapy. Each of the following problems describes the pos-
sible outcome of two alternative treatments, for three different cases. In
considering each case, suppose the patient is a 40-year-old male. Assume that
without treatment death is imminent (within a month) and that only one of
the treatments can be applied. Please indicate the treatment you would prefer
in each case.
28. Amos Tversky and Daniel Kahneman, The Framing of Decisions and the Psychology
of Choice, supra.
29. See Paul Slovic Baruch Fischhoff and Sarah Lichtenstein, Response Mode, Framing,
and Information Processing Effects in Risk Assessment, in New Directions for Methodology
of Social and Behavioral Science: The Framing of Questions and the Consistency
of Responses 21 (Robin M. Hogarth ed. San Francisco: Jossey-Bass, 1982).
30. Rational Choice in the Framing of Decisions, in Daniel Kahneman and Amos
Tversky, eds. Choice, Values and Frames 209 (New York: Cambridge University Press,
2000). The order of the cases has been changed to make it parallel to the preceding
examples in the text.
decision making under risk 487
CASE 1
Treatment A: certainty of a normal life, with an expected longevity of 18 years.
[65%]
Treatment B: 20% chance of imminent death and 80% chance of normal
life, with an expected longevity of 30 years. [35%]
CASE 2
Treatment C: 75% chance of imminent death and 25% chance of normal life,
with an expected longevity of 18 years. [32%]
Treatment D: 80% chance of imminent death and 20% chance of normal
life, with an expected longevity of 30 years. [68%]
CASE 3
Consider a new case where there is a 25% chance that the tumor is treatable
and a 75% chance that it is not. If the tumor is not treatable, death is immi-
nent. If the tumor is treatable, the outcomes of the treatment are as follows:
Treatment E: certainty of normal life, with an expected longevity of 18 years.
[68%]
Treatment F: 20% Chance of imminent death and 80% chance of normal
life, with an expected longevity of 30 years. [32%]
The three cases of this problem correspond, respectively, to the choices of
the bets A and B, C and D (Section 15.4.4), and E and F (Section 15.4.5) in the
preceding examples, and reveal the same pattern of preferences. The experi-
menters observe:
In case 1, most respondents make a risk-averse choice in favor of certain
survival with reduced longevity. In case 2, the moderate treatment no
longer ensures survival, and most respondents choose the treatment that
offers the higher expected longevity. In particular, 64% of the physicians who
chose A in case 1 selected D in case 2. This is another example of the certainty
effect.
The comparison of cases 2 and 3 provides another illustration of pseudo-
certainty. The cases are identical in terms of the relevant outcomes and
their probabilities, but the preferences differ. In particular, 56% of the phy-
sicians who chose D in case 2 selected E in case 3. The conditional framing
induces people to disregard the event of the tumor’s not being treatable
because the two treatments are equally ineffective in this case. In this frame,
treatment E enjoys the advantage of pseudocertainty: it appears to ensure
survival, but the assurance is conditional on the treatability of the tumor. In
fact, there is only a .25 chance of surviving a month if this option is
chosen.
488 problem solving, decision making, and professional judgment
15.5 appendix a
15.5.1 How People’s Different Risk Attitudes Toward Gains and Losses Can
Violate the Dominance Axiom
The axioms of completeness and dominance in conditions of uncertainty are
simply extensions of the axioms described in Section 12.2.2. Suppose a decision
maker prefers prospect (i.e., outcome) A to prospect B, and that he has a choice
between two lotteries, as shown in Figure 15.10. The first lottery, which we call
X, will result in prospect A with probability p, or prospect B with probability 1-p.
The second lottery, which we call Y, will result in prospects A or B with probabil-
ities q and 1-q, respectively. To satisfy the dominance axiom, the decision maker
will (1) prefer lottery X over lottery Y if and only if p > q; and (2) will be indifferent
between X and Y if and only if p = q.
A A
p q
X: Y:
1-p 1-q
B B
People’s different risk attitudes toward gains and losses can violate the domi-
nance axiom.
Consider the choice between these two lotteries:31
E. 25 percent chance to win $240 and 75 percent chance to lose $760, and
F. 25 percent chance to win $250 and 75 percent chance to lose $750.
When given to a group of subjects, everyone recognizes that F dominates E
and chooses F.
Now consider these two pair of concurrent choices.
Choose between:
A. A sure gain of $240, and
B. 25 percent chance to win $1000 and 75 percent chance to win nothing.
and
Choose between:
C. Sure loss of $750, and
D. 75 percent chance to lose $1000 and 25 percent chance to lose nothing.
31. Daniel Kahneman and Amos Tversky, Choices, Values, and Frames, in Choices,
Values, and Frames, supra at 1.
decision making under risk 489
As shown in Table 15.7, the choice of A and D has precisely the same expected
value as E, and the choice of B and C has the same value as F.
However, 73 percent of subjects choose A and D while only 3 percent chose B
and C,32 which dominates A and D: People are risk averse with respect to gains,
and risk seeking with respect to losses—and the violation of dominance was
stimulated by putting one part of the concurrent choice in the domain of losses
and the other in the domain of gains.
EV
A&D= E 25% chance to win $240 75% chance to lose $760 -$510
B&C= F 25% chance to win $250 75% chance to lose $750 -$500
15.6 Appendix B
10 blue, and 1 black chip. The equivalents of A and B, and C and D are shown in
Table 15.8.
Table 15.8 Allais’ Paradox
Chips
Note that choosing the red chip has no effect in the A-B choice; thus, the inde-
pendence axiom implies that one only need to consider the black and
blue chips. By the same token, choosing the red chip has no effect in the C-D
choice, and only the black and blue chips need to be considered. And note
that, with the red chips excluded, the lottery in A-B is precisely the same as that
in C-D.
Thus does the extraordinarily high utility placed on certainty cause a violation
of the independence axiom.
16. the role of affect in risky decisions
Anticipated outcomes
(including anticipated emotions)
Cognitive
evaluation
Subjective
probabilities Behavior Outcomes
(incl. emotions)
Feelings
Other factors,
e.g., vividness, immediacy,
background mood
5. Id. at 170.
6. Id. at 173–74 (paragraphs rearranged).
7. Id. at 174.
8. David H. Barlow, Anxiety and Its Disorders: The Nature and Treatment of
Anxiety and Panic (1988) (quoted in Loewenstein et al., supra note 2).
9. Damasio, supra at 193.
494 problem solving, decision making, and professional judgment
His experiments suggest that anxiety may be an important aspect of one’s capac-
ity to plan. For example, subjects were given a sum of play money and were
asked to repeatedly choose cards from among four decks, which had different
risks and payoffs. The subjects had to learn the rules that governed payoffs
through trial and error, and the rules were sufficiently complex that they were
difficult to learn them with certainty. Two of the decks provided consistent pay-
ments but unpredictably required the player to pay the experimenter a very large
sum. The other two decks provided smaller payments and unpredictably required
the player to pay a smaller amount. The expected value of choosing from the first
two decks was smaller than the latter.
Over time, members of a control group learned to avoid the two sets of cards
with higher risks and lower expected value. The feeling-impaired subjects con-
tinued to choose cards from the risk decks and went bankrupt. Members of the
control group hesitated before choosing the risky cards and they exhibited
heightened visceral reactions (e.g., change in skin conductance) before making
the choice. The feeling-impaired subjects did not hesitate and gave no physiolog-
ical indications before making the choice.
Loewenstein et al. note that “the lack of an emotional response does not nec-
essarily lead to poor decisions. . . . One could easily design an experiment where
the expected value of the high-risk deck (that contains some large losses) is actu-
ally higher than that of the low-risk deck. In this case, prefrontal damaged
patients would do better in the long run than nonpatients, because the fear in the
latter group would hinder them from choosing from the risky but higher expected
value deck.” Indeed, as we shall see below, emotions can often distort people’s
perception of or reaction to risk.
Damasio’s subjects had brain lesions that left their analytic capacities uncompro-
mised, but impaired their ability to associate emotions with the anticipated con-
sequences of their actions. People with and without similar lesions were given
this problem:
You have abandoned a sinking cruise ship and are in a crowded lifeboat that
is dangerously low in the water. If nothing is done it will sink before the rescue
boats arrive and everyone will die. However, there is an injured person who
will not survive in any case. If you throw that person overboard, the boat will
stay afloat and the remaining.
Damasio writes that “the action of biological drives, body states, and emotions
may be an indispensable foundation for rationality. . . . Rationality is probably
shaped and modulated by body signals, even as it performs the most sublime
distinctions and acts accordingly.”10 Emotions and reasoning exist in a delicate
balance, however, and emotions can sometimes overwhelm reasoning to our
detriment. The following pages focus mostly on risks involving human life and
safety, where emotions tend to run fairly high.
In Section 9.6 we saw that people often estimate the frequency of events (or
their probabilities) based on how easily examples come to mind. We saw that the
availability or vividness of events like shark attacks and the 9/11 terrorist attacks
affects people’s perception of risk. As David Myers notes, such perceptions have
behavioral consequences:
Even before the horrors of September 11th and the ensuing crash at Rockaway
Beach, 44 percent of those willing to risk flying told Gallup they felt fearful. . . . After
the four crashed airliners, and with threats of more terror to come, cancellations
understandably left airlines, travel agencies, and holiday hotels flying into the red.
. . . If we now fly 20 percent less and instead drive half those unflown
miles, we will spend 2 percent more time in motor vehicles. This translates
into 800 more people dying as passengers and pedestrians. So, in just the next
year the terrorists may indirectly kill three times more people on our high-
ways than died on those four fated planes.11
A more quotidian example is that purchases of flood and earthquake insur-
ance are high in the aftermath of a natural disaster, but decline as the vividness
of the event recedes from memory.
The easier it is to imagine a particular outcome, the more likely its occurrence
seems to be. Steven Sherman and his colleagues did an experiment in which stu-
dents were told about Hyposcenia-B, a (fictious) illness that was becoming preva-
lent on their campus. In the easy-to-imagine scenario, students were told that the
disease had concrete symptoms that most of them had probably experienced: low
energy level, muscle aches, severe headaches. In the difficult-to-imagine scenario,
the symptoms were less concrete: a vague sense of disorientation, malfunctioning
nervous system, and inflamed liver. Students given the former scenario thought
they were more likely to contract the disease than those given the latter. The
experimenters conclude that the effort required to imagine contracting the more
difficult-to-imagine symptoms led to a decrease in their availability.12
An easily imagined harm can lead people to violate the fundamental principle
that a subset cannot be greater than the larger set of which it is a part.13 A striking
example of this is people’s stated willingness to pay more for flight insurance for
losses resulting from “terrorism” than for flight insurance from all causes (includ-
ing terrorism).14 By the same token, in a between-subjects experiment, Eric
Johnson and his colleagues asked people how much they would pay for insurance
for hospitalization in four different scenarios. 15 The first two groups were respec-
tively asked their willingness to pay (WTP) for hospitalization insurance for any
reason, and for any disease or accident. The third group was asked their WTP for
insurance for any accident assuming that they already had insurance for disease,
and the fourth group was asked their WTP for insurance against any disease
assuming that they already had accident insurance. The mean WTP was:
Any reason $41.53
Any disease or accident $47.12
Any accident (followed by any disease) $69.55
Any disease (followed by any accident) $89.10
Isolating accidents and diseases makes vivid the individual imagine examples
of each, and thus leads to an irrational hierarchy of choices. The WTP is more
All the big money on an accident policy comes from [insuring] railroad accidents.
They found out pretty quick, when they began to write accident insurance, that the
apparent danger spots, the stops that people think are danger spots, aren’t
danger spots at all. I mean, people always think a railroad train is a pretty danger-
ous place to be, . . . but the figures show that not many people get killed, or even
hurt, on railroad trains. So on accident policies, they put in a feature that sounds
pretty good to the man that buys it, because he’s a little worried about train trips,
but it doesn’t cost the company much, because it knows he’s pretty sure to get
there safely. They pay double indemnity for railroad accidents.
—JAMES CAIN, DOUBLE INDEMNITY (1936).
13. Id. Recall the “Linda” problem in Section 8.2.3, where the representativeness heu-
ristic led to a similar error.
14. Id. People’s WTP for $100,000 insurance on a trans-Atlantic flight was $14.12 for
any act of terrorism, $10.31 for any nonterrorism related mechanical failure, and $12.03
for any reason.
15. Eric J. Johnson, John C. Hershey, and Howard Kunreuther Framing, Probability
Distortions, and Insurance Decisions, 7 Journal of Risk and Uncertainty, 35–52 (August
1993). Reprinted in: Choices, Values and Frames (Daniel Kahneman and Amos Tversky
eds., Cambridge, Eng.: Cambridge University Press, 2000).
the role of affect in risky decisions 497
In the public sphere, the aggravating factors can lead to high levels of con-
cern, or outrage. As Dr. Robert Scheuplein, former head of the U.S. Food and
Drug Administration’s Office of Toxicology, noted: “When risks are perceived to
be dread[ful], unfamiliar, uncontrollable by the individual, unfair, involuntary,
and potentially catastrophic, they are typically of great public concern, or high
outrage. When risks are perceived as voluntary, controllable by the individual,
18. Timur Kuran and Cass R. Sunstein, Availability Cascades and Risk Regulation, 51
Stanford Law Review 683, 709 (1999).
the role of affect in risky decisions 499
16.3.2 Familiarity
Familiar risks, such as driving, engender less dread or outrage than unfamiliar
ones, such as the explosion of a nuclear power plant, bioterrorism, or the D.C.
snipers’ attacks—even when the familiar risks are more likely to eventuate. This
may respond to the fact that familiar risks tend to recede to the background and
do not gain one’s attention. Cass Sunstein suggests that the unacceptability of
unfamiliar risks may be a consequence of loss aversion:21
People will be closely attuned to the losses produced by any newly introduced
risk, or by any aggravation of existing risks, but far less concerned with the
benefits that are foregone as a result of regulation. . . . The opportunity costs
of regulation often register little or not at all, whereas the out-of-pocket costs
of the activity or substance in question are entirely visible. In fact this is a
form of status quo bias. The status quo marks the baseline against which
gains and losses are measured, and a loss from the status quo seems much
more bad than a gain from the status quo seems good.
. . . [Loss aversion] places a spotlight on the losses introduced by some risk,
and downplay[s] the benefits foregone as a result of controls on that risk. . . .
Consider the emphasis, in the United States, on the risks of insufficient test-
ing of medicines as compared with the risks of delaying the availability of
those medicines. . . . For many people, the possible harms of cloning register
more strongly than the potential therapeutic benefits that would be elimi-
nated by a ban on the practice.
22. These and other factors are analyzed in Paul Slovic, The Perception of Risk
(London: Earthscan Publications, 2000) [hereafter, Perception of Risk].
23. Laws of Fear 37.
24. See Perception of Risk at 291.
25. See James P. Collman, Naturally Dangerous (Sausalito, CA: 2001). Collman
writes that organic foods, favored by many people on grounds of safety and health and
creating annual revenues of $4.5 billion in the United States alone, are “actually riskier to
consume than food grown with synthetic chemicals.” (31)
26. See Washington v. Glucksberg, 521 US 702 (1997) (distinguishing between right to
have life support withdrawn and physician-assisted suicide).
the role of affect in risky decisions 501
causes a harm is held liable, while someone who fails to prevent a harm usually
is not. The distinction is manifested in many moral intuitions as well.27 Among
other things, it reflects the generalization that it is easier to attribute causal
responsibility to action than inaction. From childhood, we are taught not to cause
direct harm, while the harms of inaction often seem indirect.
But if Hippocrates’ mandate, “First, do no harm,” captures a preference for
inaction where intervention may be counterproductive, the familiar injunction,
“Don’t just sit there. Do something!” also reflects a strong tendency.
Research suggests that people can have both an omission and an action bias,
depending on the circumstances and also on individual differences.
In a study modeled on the facts of the DPT vaccine, which causes serious
permanent neurological injury in one dose out of 310,000, Ilana Ritov and
Jonathan Baron found that many subjects preferred to subject their children to a
higher probability of dying from a disease that the vaccine would prevent, than a
smaller chance of dying from side-effects of the vaccine.28 Although a later study
by other Terry Connolly and Jochen Reb29 did not replicate the findings of omis-
sion bias with respect to vaccination, they concluded that anticipating regret was
nonetheless a significant factor in people’s decision whether to vaccinate: “inten-
tion to vaccinate was predicted by three measures: the respondent’s assessment
of the relative seriousness of the disease and the vaccine side-effects; her assess-
ment of the regret she would feel if vaccination turned out badly; and her assess-
ment of the regret she would feel if non-vaccination turned out badly.”
A later article by Baron and Ritov documents omission bias in various contexts,30
but also notes the existence of action bias. As in the case of the soccer goalie’s pen-
chant for action described in Section 14.2.2, a decision maker’s response may be
influenced by what is deemed “normal” behavior under the circumstance. By its
nature, a bias in either direction detracts from optimal decision making.
While the unfortunate outcome of an individual decision may result only in
personal regret, officials may incur public blame for bad policy decisions.
Officials in the Food and Drug Administration might well believe that criticisms
for delaying the approval of potentially life-saving drugs are mild compared to
the criticism of approving a drug that turns out to be deadly. One can think of
27. See Cass Sunstein, “Moral Heuristics,” University of Chicago Law & Economics,
Olin Working Paper No. 180 (2003).
28. Ilana Ritov and Jonathan Baron, Reluctance to Vaccinate: Omission Bias and
Ambiguity, in Behavioral Law and Economics 168 (Cass Sunstein ed., New York:
Cambridge University Press, 2000).
29. Terry Connolly and Jochen Reb, Omission Bias in Vaccination Decisions: Where’s
the “Omission”? Where’s the “Bias?, 91 Org. Behav. and Human Decision Processes
186–202 (2003).
30. Jonathan Baron and Ilana Ritov, Omission Bias, Individual Differences, and Normality,
94 Organizational Behavior and Human Decision Processes 74–85 (2004).
502 problem solving, decision making, and professional judgment
many other instances of omission bias by public officials. But there are also
apparent instances of action bias, perhaps including the Bush Administration’s
decision to invade Iraq after 9/11.
31. Jonathan J. Koehler and Andrew Gershoff, Betrayal Aversion: When Agents of
Protection Become Agents of Harm, 90 Organizational Behavior and Human Decision
Processes 244 (2003).
the role of affect in risky decisions 503
have the air bag that’s supposed to save my life be the cause of its termination.”
The authors consider and reject the possibility that the phenomenon is a result
of omission bias, since the participants were forced to actively choose between
the two safety devices.
32. Ali Alhakami and Paul Slovic, A Psychological Study of the Inverse Relationship
Between Perceived Risk and Perceived Benefit, 14 Risk Analysis 1085 (1994).
33. See Affect Heuristic, supra.
34. Howard Margolis, Dealing with Risk: Why the Public and the Experts Disagree
on Environmental Issues 124–125 (Chicago: University of Chicago Press, 1998).
35. Cass Sunstein, Worst-Case Scenarios 219 (Cambridge: Harvard University
Press, 2007) [hereafter, Worst Case Scenarios].
504 problem solving, decision making, and professional judgment
that when the U.S. Environmental Protection Agency (EPA) called attention to
the toxic risks of a mining town’s major industry, the citizens did not respond by
taking precautions, but rather by demonizing the EPA.
36. Dan M. Kahan, Paul Slovic, Donald Braman, and John Gastil, Fear of Democracy:A
Cultural Evaluation of Sunstein on Risk, 119 Harv. L. Rev. 1071 (2006).
the role of affect in risky decisions 505
Kahan, Slovic, et al. criticize Sunstein for, in effect, privileging experts’ per-
ceptions of risk over those of laypersons because he believes that laypersons are
more prone to the cognitive and affective biases described in this and the preced-
ing chapters. The critics doubt that policies concerning major environmental
and other issues involving risk can be adequately addressed only through cost-
benefit, or expected value, analysis.
Our own view is that issues of “cultural cognition” do play a role in people’s
perception of risk and that it is often difficult to separate cognitive and affective
biases from an individual’s values and tastes in determining his or her risk
preferences. However, as Sunstein notes in a response to his critics, most of the
risks that we face involve rather quotidian issues rather than divisive cultural and
political issues.37
Suppose you have an urn containing thirty red balls and sixty other balls that
are either black or yellow. You don’t know how many black or yellow balls there
are, but you know that the total number of black balls plus the total number of
yellow equals sixty. The balls are well mixed so that each individual ball is as
likely to be drawn as any other. You are now given a choice between two gambles,
as shown in Table 16.2a.
Gamble A Gamble B
You receive $100 if you draw a red ball You receive $100 if you draw a black ball
Also you are given the choice between these two gambles (in a different draw
from the same urn), as shown in Table 16.2b.
Gamble C Gamble D
You receive $100 if you draw a red or You receive $100 if you draw a black or
yellow ball yellow ball
Expected utility theory holds that you should prefer Gamble A to Gamble B if,
and only if, you believe that drawing a red ball is more likely than drawing a black
ball. Thus, there would be no clear preference between the choices if you thought
that a red ball was as likely as a black ball. Similarly it follows that you will prefer
Gamble C to Gamble D if, and only if, you believe that drawing a red or yellow
ball is more likely than drawing a black or yellow ball. If drawing a red ball is
more likely than drawing a black ball, then drawing a red or yellow ball is also
more likely than drawing a black or yellow ball. So, if you prefer Gamble A to
Gamble B, it follows that you will also prefer Gamble C to Gamble D. And, if
instead that you prefer Gamble D to Gamble C, it follows that you will also prefer
Gamble B to Gamble A. However, most people prefer Gamble A to Gamble B
and Gamble D to Gamble C.
Why is this paradoxical—or at least a logical error? A preference for Gamble A
to B implies that you believe that there are fewer black balls than red balls – i.e.,
fewer than 30 black balls, which entails more than thirty yellow balls. A preference
for Gamble D to C implies that you believe that there are more black balls than red
balls. So a preference for A and D implies that you believe the number of black
balls is less than the number of red balls and more than the number of red balls.
the role of affect in risky decisions 507
41. Cf. Robin Hogarth and Howard Kunreuther, Decision Making Under Uncertainty: The
Effects of Role and Ambiguity, in Decision Making and Leadership 189 (Frank Heller, ed.,
New York: Cambridge University Press, 1992).
42. Noting that even when frequentist statistics cannot produce “objective” probabili-
ties, people often assign subjective probabilities to events, some commentators have ques-
tioned whether uncertainty actually exists. Worst-Case Scenarios 159 ff. But when the
range of one’s subjective estimate of the probability of an event occurring is sufficiently
large, assigning a numerical probability seems pointless.
43. Hogarth and Kunreuther, supra. It appears that, in making judgments under uncer-
tainty a decision maker anchors on the given probability, but then adjusts the estimate
based on imagining other values it might have.
44. Merce Roca, Robin M. Hogarth, and A. John Maule, Ambiguity Seeking as a Result
of the Status Quo Bias, 32 Journal of Risk and Uncertainty 175 (2006).
508 problem solving, decision making, and professional judgment
p
B ~
1-p
46. Larry S. Temkin, Weighing Goods: Some Questions and Comments, 23 Phil. & Public
Affairs 350 (1994).
47. See Worst-Case Scenarios, supra.
48. Yakov Ben Haim, Cliff Dasco, and Barry Schwartz, “What Makes a Good Decision?
Robust Satisficing as a Normative Standard of Rational Decision Making” (unpublished
paper 2009).
510 problem solving, decision making, and professional judgment
Suppose that you are considering a number of entry-level job offers in a large
government agency, and that your criteria include (1) salary and other benefits,
(2) the quality of city life, (3) whether the work will be interesting and fulfilling,
and (4) the quality of support and mentorship you will receive. You can ascertain
the first two within a small margin of error, but the latter two are deeply uncer-
tain: they depend on what department or division you will be assigned to and on
the particular people you work for, which cannot be known in advance. Ben
Haim et al. propose that rather than attempting to determine which decision
course will yield the highest expected value, you should identify a threshold of a
satisfactory, or “good enough,” outcome and choose the job that is most robust
to uncertainty—that is most likely to produce a satisfactory outcome. Similarly,
robust satisficing would be a useful decision-making strategy for a man pre-
sented with a choice of treatments for prostate cancer, which have different
chances of curing the cancer and different likely side effects. Even if some of
these probabilities can be specified, your prediction of how side effects will affect
your future wellbeing fall into the realm of uncertainty.
The mathematics of robust satisficing is pretty complicated—more so than
for expected utility. Ben Haim, Dasco, and Schwartz suggest these (satisficing)
rules of thumb:
• Ask yourself what you need to achieve in order to be happy or satisfied with
the outcome of the decision. . . . Remember that high aspirations are more
vulnerable to error than modest ones.
• Distinguish between two very different attributes of each option. One attri-
bute of an option, which we call its nominal outcome, is the estimate of the
outcome of that option based on your best data and understanding. . . . The
other attribute of an option, called its robustness, is the degree of immunity
of the outcome to error in your estimates and understanding. The robust-
ness is a nonprobabilistic assessment of immunity to error. Nonetheless,
robustness may assess immunity to error in assessments of probabilities.
An option may be quite attractive based on your current understanding
(good nominal outcome), but may lead to highly undesirable results due to
only small errors (low robustness). Alternatively, an option may be both
nominally very attractive and very robust to error. Or, an option may nomi-
nally be only moderately attractive, but it may be very robust to error. These
two attributes—nominal outcome and robustness—are different and not
correlated with one another.
• Evaluate the nominal outcome of each option. This is relatively easy,
since you can ignore uncertainty and assume that your data and under-
standing are correct. Rank the options from highest to lowest nominal
outcome.
• Evaluate the robustness of each option for the particular quality of outcome
that you identified in the first step. This is more difficult, since you must ask:
the role of affect in risky decisions 511
How wrong can I be and still have the option yield a satisfactory outcome?
Listing contingencies (this could go wrong, that could go wrong, etc.) is one
approach. . . . Rank the options from highest to lowest robustness.
• Choose the most robust option. If the rankings according to nominal out-
come and robustness agree, then the most robust option will also be the
nominally most attractive option. If the rankings according to nominal out-
come and robustness disagree, then the most robust option may differ
from the nominally most attractive option.
• Step back and look carefully at your analysis. Does it make sense? . . . Ask
if your thought process fits together coherently, if it uses all relevant infor-
mation, if it avoids major assumptions or leaps of faith, and if it is free of
considerations that are not explicitly included.
49. Howard Kunreuther, Nathan Novemsky, and Daniel Kahneman, Making Low
Probabilities Useful, 23 J. Risk & Uncertainty 103 (2001).
50. Hypothesizing that the chemical plant risks would be more evaluable if partici-
pants were able to compare them to more familiar risks, like a car accident, the experi-
menters told subjects that the probability of injury in a car accident was 1/6000 per year
and asked them to evaluate a 1/650, 1/6300, and 1/68,000 chance of the plant’s release of
toxics. The differences among the groups were still negligible. However, when subjects
512 problem solving, decision making, and professional judgment
were given scenarios describing the contexts in which car accidents happened with rela-
tively and low and high probabilities (1/5900 chance of accident on icy mountain roads;
1/66,000 on flat desert highways), the differences were significant. The scenarios made
the risks more evaluable than the statistics alone. “There needs to be fairly rich context
information available for people to be able to judge differences between low probability
events. In particular, people need comparisons of risk that are located on the probability
scale and evoke people’s own feelings of risk.” Kunreuther, et al., supra.
51. Michael Jones Lee and Graham Loomis, Private Values and Public Policy, in Elke
Weber, Jonathan Baron, and Graham Loomis, Conflict and Tradeoffs in Decision
Making 205 (Cambridge, UK: Cambridge University Press, 2001).
the role of affect in risky decisions 513
52. Paul Slovic, Baruch Fischhoff, and Sara Lichtenstein, Accident Probabilities and Seat
Belt Usage: A Psychological Perspective, 10 Accident Analysis and Prevention 281 (1978).
53. Christopher K. Hsee and Yuval Rottenstreich, Music, Pandas, and Muggers: On the
Affective Psychology of Value, 133 J. Experimental Psychol.: Gen. 23, 23–24 (2004).
54. Jeremy A. Blumenthal, Emotional Paternalism, 35 Fla. St. U. L. Rev. 1, 23 (2007),
discussing Veronika Denes-Raj and Seymour Epstein, Conflict Between Intuitive and Rational
Processing: When People Behave Against Their Better Judgment, 66 J. Personality & Soc.
Psychol. 819, 823 (1994); Seymour Epstein and Rosemary Pacini, Some Basic Issues
Regarding Dual-Process Theories from the Perspective of Cognitive-Experiential Self-Theory, in
514 problem solving, decision making, and professional judgment
Dual-Process Theories in Social Psychology 462 (Shelly Chaiken and Yaacov Trope
eds., 1999).
55. Recall the suggestion in Section 8.6.1 that describing events in terms of frequen-
cies rather than probabilities could, under some circumstances, improve individuals’
intuitive Bayesian analysis.
56. Slovic, Monahan, and MacGregor, Violence risk assessment and risk communica-
tion: The Effects of Using Actual Cases, Providing Instruction, and Employing Probability
Versus Frequency Formats, 24 Law Hum Behav. 271–96 (2000).
57. Affect Heuristic, supra.
58. Kimihiko Yamagishi, When a 12.86% Mortality Is More Dangerous than 24.14%:
Implications for Risk Communication, 11 Applied Cognitive Psychology, 495 (1997).
59. Howard Kunreuther and Mark Pauly, Neglecting Disaster: Why Don’t People Insure
Against Large Losses, 28 J. Risk and Uncertainty 5 (2004).
60. Oswald Hoberet al., Active Information Search and Complete Information Presentation,
95 Acta Psychologica 15 (1997).
the role of affect in risky decisions 515
Boone’s older brother, Pete, is orienting Boone for his first day on the job at as an
insurance claim adjuster.
“These are claims,” his older brother explained, grabbing a stack of paper-
clipped and clamped wads of papers and forms from a bin outside the cubicle
marked IN. “Your job is to deny them.”
“I see,” Boone had said. “You mean, I sort through the claims and deny all
the fraudulent ones, right?”
His brother implored heaven for patience with a roll of his eyes, then sighed
a gust of wintry disgust. “The fraudulent claims were picked out downstairs by
high school graduates and denied three months ago. Anybody can deny a
fraudulent claim. You’re a college graduate. Your job is to find a way to deny
legitimate claims. . . .
“People who file claims believe that money will make them happy and will
somehow compensate them for their losses. This idea—that money makes
misfortune easier to bear—is an illusion that can only be enjoyed by those
who have not suffered an actual loss.
“The most terrifying thing about life is knowing that, at any moment, a
freak accident, violence, mayhem, a psychotic break, an addiction, a heart
attack, a sexually transmitted disease, cancer, an earthquake, or some other
act of God, or worse, can take all of your happiness away from you in the time
61. Elke U. Weber and Christopher Hsee, Cross-Cultural Differences in Risk Perception,
But Cross-Cultural Similarities in Attitudes Towards Perceived Risk, 44 Management
Science 1205 (1998); Ch. Schade, H. Kunreuther, and K.P. Kaas (2002), “Low-Probability
Insurance Decisions: The Role of Concern,” Discussion Paper Number 23, SFB 373,
Humboldt-Universität zu Berlin/Wharton Risk Center Working Paper Number 02-10-HK,
Wharton School, University of Pennsylvania, USA; Paul Slovic, Baruch Fischhoff, and
Sarah Lichtenstein, Facts and Fears: Understanding Perceived Risk, in Societal Risk
Assessment: How Safe Is Safe Enough? 181 (R. Schwing and W. A. Albers, Jr., eds.,
1980).
516 problem solving, decision making, and professional judgment
it takes you to pick up the phone and get the news. That’s why people buy
insurance, because they think it will protect them from catastrophes.
“But we are in the insurance business,” said Pete. . . . “We know there is no
protection from catastrophes. No matter what you do, there’s always a chance
that a catastrophe will come along, tear your heart out of your chest, and rub
it in your face.
“When you’re crawling on the bathroom floor sick with grief . . ., wondering
why God failed to give you the courage to kill yourself, a big check from the
insurance company looks like a swatch of wallpaper. You’re in a place money
can’t reach.
“So, insurance only works if catastrophe does not strike. . . . We don’t sell
protection. We sell peace of mind. For a premium, we agree to give the con-
sumer the illusion that money will protect him from every possible foreseeable
catastrophe. Once the premium is paid and before catastrophe strikes, the
consumer is free to wallow in the illusion that if something terrible happens
money will take the sting out of it. When a catastrophe actually occurs, the
illusion is shattered and there’s nothing to be done but drag yourself out of
bed every morning and get on with your life.”
“But if what you say is true,” asked Boone, “then you are charging people
thousands of dollars for . . . an illusion.”
“Exactly,” said Pete. “Peace of mind. The money is irrelevant. You probably
subscribe to the notion that insurance is a way to pool risk and share liability.
You think premiums should be based upon risk. Nothing could be more
wrong. Premiums should be based upon line thirty-one of your federal tax
return, adjusted gross income. Our objective is to charge the insured just
enough to make it hurt. We are looking for the financial pain threshold,
because only when it hurts does the insured really believe that he is obtaining
something of value, and, as I’ve shown, he is indeed obtaining peace of mind
for nothing more than money.”
RICHARD DOOLING, WHITE MAN’S GRAVE 25–27 (1994) Copyright © Richard Dooling
1994. Reprinted with permission.
set of actions that would more effectively reduce the risk.62 As Joseph Conrad
wrote in Nostromo, “Action is consolatory. It is the enemy of thought and the
friend of flattering illusions.”
62. Elke Weber, Perception and Expectation of Climate Change: Precondition for Economic
and Technological Adaptation, in Environment, Ethics, and Behavior: The Psychology
of Environmental Valuation and Degradation (M. H. Bazerman, D. M. Messick, A.
E. Tenbrunsel, and K. A. Wade-Benzoni eds., 1997).
the role of affect in risky decisions 517
The stronger the emotions, the more people tend to greatly underweight
probabilities or to ignore them altogether and focus only on the horrific, worst
case outcome. Cass Sunstein has coined the term probability neglect to describe
people’s departure “from the normative theory of rationality in giving excessive
weight to low-probability outcomes when the stakes are high” and giving low-
probability outcomes no weight at all when the risks are not vivid.
The phenomenon is illustrated by an experiment, based on an actual problem
that faced the Environmental Protection Agency. Sunstein asked law students to
indicate their willingness to pay (WTP) to reduce levels of arsenic in drinking
water to eliminate
• a cancer risk of 1/1,000,000;
• a cancer risk of 1/100,000;
• a cancer risk of 1/1,000,000 where the cancer was described in vividly grue-
some terms;
• a cancer risk of 1/100,000 also described in gruesome terms.
When given the unemotional description, people’s WTP increased signifi-
cantly as the probability of contracting cancer increased. Merely describing the
cancer in gruesome terms doubled people’s WTP to avoid the 1/1,000,000 risk,
but the tenfold increase of gruesome cancer risk did not greatly increase WTP.
Sunstein argues that probability neglect does not involve misestimating proba-
bilities based on their vividness or availability. Rather, emotion essentially
swamps considerations of probability.63
Laypersons’ intuitions about toxicology also manifest probability neglect.
Nancy Kraus, Torbjörn Malmfors, and Paul Slovic compared the basic attitudes
of professionals and laypersons toward toxic risks.64 A core assumption of toxi-
cology is that “the dose makes the poison,” meaning that there is a positive cor-
relation between the size of the dose and the likelihood of harm and that some
chemicals that are deadly in high concentrations are harmless in small amounts.
However, this view is not shared by the layperson—the “intuitive toxicologist”—
who tends to believe that “if large exposures to a chemical are harmful, then
small exposures are also harmful.” Specifically, laypersons believe:
• any exposure to a toxic chemical makes one likely to suffer adverse health
effects;
• any exposure to a carcinogen makes one likely to get cancer;
• the fact of exposure to a pesticide is the critical concern, rather than the
amount of exposure;
65. Paul Slovic, “If Hormesis Exists . . . Implications for Risk Perception and
Communication,” http://www.belleonline.com/newsletters/volume7/vol7-1/ifhormesi-
sexists.html. See also Paul Rozin and Carol Nemeroff, Sympathetic Magical Thinking: The
Contagion and Similarity “Heuristics,” in Heuristics and Biases, supra.
the role of affect in risky decisions 519
Such bimodal responses provide further support for the intuitive suggestion
that some risks are simply “off-screen,” whereas others, statistically not much
larger, can come “on-screen” and produce behavioral changes.66
The tendency to ignore some low-probability risks is partly due to the diffi-
culty of assessing them. As Sunstein notes, “a decision to disregard low-level
risks is far from irrational, even if it is based in whole or in part on emotions; we
lack the information that would permit fine-grained risk judgments, and when
the probability really is low, it may be sensible to treat it as if it were zero.”67 We
may also ignore risks because of our tendency to be unrealistically optimistic—
especially about our own futures68—and also for peace of mind. As Kai Erickson
writes: “One of the bargains men make with one another in order to maintain
their sanity is to share an illusion that they are safe, even when the physical evi-
dence in the world around them does not seem to warrant that conclusion.”69
Preserving sanity through illusion can have its costs, though. Consider the com-
placency about airline security before September 11, 2001, and the muted recep-
tion given the January 31, 2001, report of the U.S. Commission on National
Security, co-chaired by former Senators Gary Hart and Warren Rudman, which
warned of America’s vulnerability to terrorist attacks. The absence of availability
may lull people into complacency: What is out of sight is effectively out of mind.70
66. Cass Sunstein, Probability Neglect: Emotions, Worst Cases, and Law, 112 Yale L.J. 61,
75 (2002).
67. Id.
68. See Section 13.5. Bimodalism in this context may be a close cousin of the certainty
effect, discussed above.
69. Kai Erikson, Everything in its Path: Destruction of Community in the
Buffalo Creek Flood 234 (1976). See also George Akerlof and William Dickens, The
Economic Consequences of Cognitive Dissonance, in George Akerlof, An Economic
Theorist’s Book of Tales 123, 124–28 (1984).
70. Paul Slovic, Baruch Fischhoff, and Sarah Lichtenstein, Rating the Risks, 21
Environment 14 (1979).
71. Michael Jones Lee and Graham Loomis, Private Values and Public Policy, supra.
520 problem solving, decision making, and professional judgment
The preceding discussions have implications for the practice of cost-benefit anal-
ysis surveyed in Section 12.3.
the public, should bear the burden of proof.”81 There are many other versions of
the PP, but they all share the common element that it is worth high social oppor-
tunity costs to prevent the possibility of catastrophic harm.
In Worst-Case Scenarios, Cass Sunstein undertakes an extensive critique of the
PP, arguing that it is extremely vague, prone to biases of the sort described in
this book, and likely to be counterproductive to human welfare. Sunstein makes
these observations, among others:
• The PP makes a distinction between action and inaction that is at best
orthogonal to expected value. For example, the PP is frequently invoked to
halt the production of foods using genetically modified organizations, but
not to mobilize research and development to prevent a large meteor from
hitting the earth.
• The PP is invoked against particularly salient risks, based on the availabil-
ity heuristic, but not against risks that are not on people’s radar screens.
• By virtue of loss aversion, “people will be closely attuned to the potential
losses from any newly introduced risk . . ., but far less concerned about
future gains they may never see if a current risk is reduced.” Consider the
barriers to the introduction of new pharmaceuticals, which are on-screen,
compared to the illness and death that might be reduced by their introduc-
tion, which are off-screen.
• Prohibiting one harm under the PP may create equal or worse harms. For
example, in developing countries, bans on DDT have led to increases in
malaria, and bans on genetically modified organisms may increase deaths
through starvation.
Although Sunstein rejects most formulations of the PP, he proposes what he
calls an irreversible and catastrophic harm principle, under which “when a harm is
irreversible in the sense that restoration is very costly or impossible, special pre-
cautions may be justified; . . . it often makes sense to ‘buy’ an option to preserve
future flexibility; and . . . loss of cherished and qualitatively distinctive goods
deserve particular attention.”82
Without delving more deeply into this complex and hotly contested issue, two
things are reasonably clear to us. First, that the trade-offs between current ben-
efits and the risk of catastrophic harm cannot be left entirely to experts but, in a
democracy, must be the subject of debate and deliberation by legislators informed
by diverse civil society organizations. Second, public deliberation should be
informed by experts’ estimates of the scope and probability of harms, even if
they come with large margins of error.
Thus far, we have identified a variety of cognitive biases and heuristics that can
lead to incorrect empirical judgments and to decision making that does not max-
imize an individual’s utility. These include:
• the availability heuristic;
• the representativeness heuristic;
• social stereotyping;
• anchoring bias;
• neglect of base rates;
• confirmation bias;
• overconfidence;
• hindsight bias;
• overoptimism;
• self-serving bias
• misforecasting the effects of positive and negative experiences;
• framing effects—particularly the endowment effect and loss aversion;
1. See generally Scott O. Lilienfeld et al., Giving Debiasing Away: Can Psychological
Research on Correcting Cognitive Errors Promote Human Welfare?, 4 Perspectives on
Psychol. Sci. 390 (2009); Katherine L. Milkman et al., How Can Decision Making Be
Improved?, 4 Perspectives on Psychol. Sci. 379 (2009); Richard P. Larrick, Debiasing, in
Blackwell Handbook of Judgment and Decision Making 316 (Derek J. Koehler and
Nigel Harvey eds., Malden, MA: Blackwell Publishing, 2004).
524 problem solving, decision making, and professional judgment
• difficulties in dealing with too many choices and the framing and grouping
of choices;
• difficulties in evaluating and comparing risks.
Along the way, we have occasionally mentioned efforts to debias these
phenomena. Here we summarize what is known about debiasing techniques.
Informing people of the biases. Informing decision makers that they are
prone to particular biases has little or no effect in most circumstances.2
Incentives. The hope is that incentives will reduce error is based on the
assumption that they will motivate System 2 analysis and deliberation. But this
requires that people have the judgment and decision-making skills that the addi-
tional effort will mobilize.3 While incentives may stimulate the search for addi-
tional information, they have little or no effect in mitigating most cognitive
biases.4
Accountability. As with incentives, accountability to others can only reduce
biases if decision makers have the requisite skills. Even then, as we will discuss
in Section 19.11, the social pressures of accountability may induce biases of their
own.
Considering the alternative. One of the most robust and pervasive debiasing
techniques is to consider why one’s judgment may be wrong. Hindsight bias
can be counteracted by asking people to provide reasons why an event other than
the one that actually occurred might have occurred. Overconfidence and the
biased assimilation of information can be counteracted by asking people to pro-
vide counterarguments for their judgments. And the self-serving bias in litiga-
tion can be counteracted by asking parties to write down the weaknesses of their
own case.5 These techniques can sometime backfire, however: if people find it
difficult to think of reasons or examples on the other side, it may actually
strengthen their prior beliefs.6
Representing risks and other matters of likelihood in terms of frequencies rather
than probabilities. Under at least some circumstances, people will combine prob-
abilities more accurately when considering likelihoods as frequencies rather
than probabilities––for example, that an outcome occurs 25 out of 100 times
rather than 25 percent of the time. (See Section 8.6.1).
2. See Baruch Fischhoff, Debiasing, in Daniel Kahneman, Paul Slovic, and Amos
Tversky, Judgment Under Uncertainty: Heuristics and Biases 422 (1982).
3. See Colin Camerer and Robin Hogarth, The Effects of Financial Incentives in
Experiments: A Review and Capital-Labor-Production Framework, 19 J. Risk and
Uncertainty 7 (1999).
4. Camerer and Hogarth, supra; Larrick, supra.
5. Linda Babcock, George Loewenstein, and Samuel Issacharoff, Creating Convergence:
Debiasing Biased Litigants, 22 Law & Social Inquiry 914 (1997).
6. Larrick, supra.
conclusion 525
About half the practice of a decent lawyer consists in telling would-be clients that they are
damned fools and should stop.
— Elihu Root8
7. Milkman, supra.
8. Philip C. Jessup, Elihu Root 133 (New York: Dodd Mead, 1938).
9. The client autonomy model is similar in many respects to the client-centered
approach. For a leading defense of client autonomy, see Monroe H. Friedman,
Understanding Lawyers’ Ethics (1990).
10. For an account of the development of informed consent as a normative model of
doctor-patient decision making, and an extension of that model to the legal counseling
relationship, see Mark Spiegel, Lawyering and Client Decisionmaking: Informed Consent and
the Legal Profession, 128 U. Pa. L. Rev. 41 (1979). For a cognitive limitations critique of the
informed consent model of medical decision making, see Jon Merz and Baruch Fischhoff,
Informed Consent Does Not Mean Rational Consent, 11 J. Legal Medicine 321 (1990). See
also Donald A. Redelmeier, Paul Rozin, and Daniel Kahneman, Understanding Patient’s
Decisions: Cognitive and Emotional Perspectives, 270 J. of the American Medical
Association 72 (1993) (describing various cognitive biases and heuristics that influence
patient’s approach to making medical treatment decisions).
526 problem solving, decision making, and professional judgment
11. For a discussion of this issue, see David A. Binder, Paul B. Bergman, Susan
Price, and Paul R. Tremblay, Lawyers as Counselors: A Client-Centered Approach
(Eagan, MN: West. 2nd ed., 2004).
12. Anthony T. Kronman, The Lost Lawyer: Failing Ideals of the Legal
Profession 128–129, 131 (New York: Cambridge University Press, 1993).
conclusion 527
Legal scholars have begun to apply insights from this research to a wide vari-
ety of subjects in law and legal practice.13 Much of this literature seeks to identify
sources of bias that, for example, systematically distort clients’ decisions to com-
mence, continue, or settle litigation and that thus lead to suboptimal outcomes
in terms of their own interests.14 The hope is that lawyers can help clients avoid
or compensate for these biases and thus make better decisions. In other words,
lawyers might add value to client decision making not only by predicting legal
outcomes, but also by structuring and improving decision-making processes.
So, for example, Jeffrey Rachlinski observes that by describing settlement
offers to clients in ways that counteract framing effects, attorneys can improve
outcomes for the clients and for society in general.15 In an intriguing set of
experiments (described below), Russell Korobkin and Chris Guthrie suggest
that lawyers are more likely than nonlawyers to apply expected value analysis
13. See, e.g., Donald C. Langevoort and Robert K. Rassmussen, Skewing the Results: The
Role of Lawyers in Transmitting Legal Rules, 5 S. Cal. Interdisc. L. J. 375 (1997) (using
insights from social and cognitive psychology to explain why lawyers might overstate legal
risk in counseling business clients); Gary L. Blasi, What Lawyers Know: Lawyering Expertise,
Cognitive Science, and the Functions of Theory, 45 J. Legal Educ. 313 (1995) (describing the
development of legal expertise in terms of cognitive theory and the debiasing of judg-
ment); Jody Armour, Stereotypes and Prejudice: Helping Legal Decisionmakers Break the
Prejudice Habit, 38 Cal. L. Rev. 733 (1995) (applying social cognition theory to the prob-
lem of debiasing jury decision making); Donald C. Langevoort, Where Were the Lawyers?
A Behavioral Inquiry into Lawyers’ Responsibility for Client’s Fraud, 46 Vand. L. Rev. 75
(1993) (applying insights from social psychology and social cognition theory to explain
how lawyers may fail to detect client fraud); and Albert J. Moore, Trial by Schema: Cognitive
Filters in the Courtroom, 37 U.C.L.A. L. Rev. 273 (1989) (application of insights from
cognitive psychology to trial advocacy).
14. See, e.g., Mark Kelman et. al., Context-Dependence in Legal Decision Making, 25 J.
Legal Stud. 287 (1996) (demonstrating empirically that choice in legal decision-making
situations may be influenced by context effects); Jeffrey J. Rachlinski, Gains, Losses, and
the Psychology of Litigation, 70 S. Cal. L. Rev. 113 (1996) (demonstrating how framing
effects may influence client decisions to settle or continue litigation); Linda Babock, et al.,
Forming Beliefs About Adjudicated Outcomes: Perceptions of Risk and Reservation Values, 15
Int’l Rev. L. and Econ. 289 (1995) (influence of framing and prior expectancy effects on
settlement decision making); Russell Korobkin and Chris Guthrie, Psychological Barriers
to Litigation Settlement: An Experimental Approach, 93 Mich. L. Rev. 107 (1994) (illustrat-
ing how framing effects and equity seeking may influence settlement decision making);
Russell Korobkin and Chris Guthrie, Opening Offers and Out of Court Settlement: A Little
Moderation May Not Go A Long Way, 10 Ohio St. J. on Dispute Resolution 1 (1994)
(anchoring effects and the evaluation of settlement offers); George Loewenstein, et al.,
Self-Serving Assessments of Fairness and Pretrial Bargaining, 22 J. Legal Stud. 135 (1993)
(demonstrating the effects of overoptimism and other self-serving evaluative biases on
settlement decision making).
15. Jeffrey J. Rachlinski, Gains, Losses, and the Psychology of Litigation, supra at 170–73.
528 problem solving, decision making, and professional judgment
in deciding whether to settle or continue with litigation, and are less likely in
such situations to be influenced by certain cognitive biases.
This line of research poses a challenge to the traditional client-centered model
of legal counseling: If, at least in certain contexts, lawyers are less likely than
clients to be influenced by cognitive biases or the indiscriminate application of
simplifying heuristics, does this support introducing a degree of paternalism
into normative models of lawyer-client decision making?
Korobkin and Guthrie respond to this question by positing what they refer to
as a cognitive error approach to legal counseling.16 Under this approach, the appro-
priateness of a lawyer’s intervention depends on the judgments that underlie the
client’s decision. As Korobkin and Guthrie state:
The cognitive error approach to counseling . . . requires the lawyer to assess
whether an observed difference between the lawyer’s and client’s analysis of
decision options is due to the client’s cognitive error or is merely the manifes-
tation of differences in utility functions. If the difference is due to cognitive
error, the lawyer should attempt to change the client’s outlook. If the differ-
ence is the result of different preference structures, the lawyer should scrupu-
lously avoid any interference.17
With this distinction in mind, and with due modesty about lawyers’ ability
to influence clients’ judgments that are based on cognitive biases or affect,18
let’s examine more closely Korobkin and Guthrie’s experiments involving
the lawyer’s role as counselor.19 They gave laypersons and lawyers these bias-
inducing scenarios and asked whether they would settle or go to trial.
• Anchoring. A car buyer’s suit against the dealer for the purchase of a
“lemon,” in which the defendant’s initial offer of settlement acted either as
a high or low anchor for the plaintiff’s expectations. (See Section 10.1)
16. Russell Korobkin and Chris Guthrie, Psychology, Economics, and Settlement:
A New Look at the Role of the Lawyer, 76 Texas L. Rev. 77, 129–130 (1977).
17. Id. at 130.
18. The modesty implied by the first section of this chapter is supported by Korobkin
and Guthrie’s efforts to increase settlement rates in the scenarios described in the text that
follows. They employed five strategies: (1) informing the client about the psychological
factors that might inform her decision, (2) asking the client to consider an opposite or
alternative point of view, (3) recommending settlement, (4) giving reasons for the recom-
mendation, and (5) recommending settlement without giving any reasons. All five tended
to increase the settlement rate, but not dramatically.
19. The experiments were motivated by the higher rate of settlement of civil cases
than predicted by prospect theory. As we have seen, plaintiffs and defendants view litiga-
tion from different frames, with defendants tending to be more risk-taking than plaintiffs.
Moreover, litigants may have goals besides money, such as vindication, revenge, and jus-
tice. The high rate of settlement could nonetheless be explained if lawyers did not share
the cognitive biases of their clients and if they influenced their clients to settle.
conclusion 529
• Gain/loss framing. A car accident case in which the identical situation was
manipulated so that settlement appeared to the plaintiff either as a gain or
a loss. (See Section 14.1)
• Sympathy. A suit against a landlord who failed to provide the tenant with
heat during an entire winter, in which the landlord either had been will-
fully indifferent or had a sympathetic excuse (being out of the country on a
family emergency). The psychological issue here was not bias, but the par-
ticipants’ sense of justice.20
Lawyers tended to settle regardless of the frame. Clients’ settlement rates
depended on how the issue was framed.
In the lemon case, when the anchor was high (thus creating a high expecta-
tion), clients’ settlement rates were significantly lower than lawyers’. Since it is
not plausible that the client would accord any actual value to this anchor, the
lawyers’ tendency to ignore the anchor seems unequivocally utility maximizing.
In the accident case, the clients’ settlement rates were lower in the loss frame
than in the gains frame. This is predicted by prospect theory, as described in
Section 14.1. Whether the clients’ frame-dependent tendencies produce subopti-
mal decisions that affect their experienced utility in the long run depends on the
interactions of affective forecasting, adaptation, and regret.
And in the heater case, the clients’ settlement rates were lower when the land-
lord had an unsympathetic excuse for not fixing the heater, while the settlement
rates for lawyers were the same in both situations. Based on the experimental
results and follow-up interviews, Korobkin and Guthrie conclude that the law-
yers were single-mindedly focused on monetary expected value and indifferent
to the other variables, while some clients wished to maximize a value other than
wealth. (See Section 12.1) One certainly cannot characterize a tenant’s decision
to settle for less out of sympathy as irrational. Yet the lawyer may believe that the
client is overvaluing the sympathy discount in terms of his own utility func-
tion—say, because the excuse was conveyed in a personal, emotion-laden
manner, whose psychological effects will diminish in another week or so.
Korobkin and Guthrie’s distinction between a client’s errors and utilities
echoes David Luban’s suggestion in Paternalism and the Legal Profession21 that a
lawyer may permissibly compromise client autonomy where the intrusion inter-
feres only with a client’s “wants,” but not when it would interfere with the expres-
sion of a client’s “values.”22 Luban defines values as those reasons for choosing a
particular course of action “with which the agent most closely identifies—those
20. The subjects were informed that it would not affect the outcome of the small
claims court trial.
21. David Luban, Paternalism and the Legal Profession, 1981 Wisconsin L. Rev. 454
(1981).
22. Id. at 474.
530 problem solving, decision making, and professional judgment
that form the core of his personality, that make him who he is.”23 A cognitive
error approach to legal counseling is based on the assumption that biases or
errors can be distinguished from a client’s core preferences or utilities.
To put client-centered counseling in the terms used in this book, we start from
the premise that the lawyer’s role as counselor is to help a client maximize his or
her utility—taking into account the broad conception of utility considered in
Section 12.1. While the ideal, of course, is to maximize experienced utility, lawyers,
clients, and indeed all of us, can only forecast the expected utility of decisions.
In the simplest case, the lawyer provides the client with information about
substantive law and procedure, and makes predictions of the likely consequences
of particular courses of actions, based on which the client decides what course of
action to pursue. This inevitably requires that the lawyer understand the client’s
objectives, and sometimes requires working with him to clarify them. The lawyer
may also bring to the counseling process decision-making knowledge or tools
that not all clients necessarily possess—for example, how to structure a decision
process with multiple competing objectives or how to take risk into account
when calculating expected value.24
Even when only money is at stake, these can be complex matters; consider,
for example, quantifying risks, identifying uncertainties and understanding a
client’s risk tolerance in a particular context. But, as we have seen, subjective
utility can encompass a large variety of factors besides money—factors that the
client may not be able to specify, let alone predict how they will be affected by a
decision. Even relatively simple predictions of expected value are subject to cog-
nitive biases, and as one moves to a broader range of goals the possibilities of
bias grow immensely.
What role can the lawyer as counselor play in helping a client navigate these
manifold and often hidden shoals? Although not only laypersons but experts as
well are subject to biased judgment and decision making, there are several rea-
sons that a lawyer may be able to help a client address errors that the client
wouldn’t notice himself:
• the lawyer possesses expertise in decision making25—especially if he or she
has taken a course in problem solving and decision making;
• by virtue of her disinterest, the lawyer can provide the client with a differ-
ent and neutral perspective in the matter;
• an agent rather than a principal, the lawyer does not have the “endowment”
that a client may experience with respect to the status quo;
23. Id.
24. For example, the subjective linear model described in Section 4.5 and the deci-
sion tree described in Section 15.2.2.
25. See Russell Korobkin and Chris Guthrie, Psychology, Economics, and Settlement:
A New Look at the Role of the Lawyer, 76 Texas L. Rev. 77 (1997).
conclusion 531
• as a repeat player, the lawyer has the opportunity to get better feedback on
recurrent events, and has knowledge about likely outcomes and about cli-
ents’ satisfaction or dissatisfaction in vindicating certain kinds of interests.
For all of its promise, cognitive counseling cannot escape, and indeed may con-
tribute to, some fundamental dilemmas of the decision-making process described
in the preceding chapters. Most notably, if preferences are sometimes constructed
in the very process of making a decision, there is a danger that lawyer’s prefer-
ences may unwittingly influence the client’s decision. The very act of communicat-
ing information to clients subtly shapes preferences and influences choice.26 The
client may be led, either by the lawyer directly or by the “scripted” nature of the
lawyer-client interaction, to value certain aspects of utility at the moment differ-
ently than they will be valued in life outside the lawyer’s office. Indeed, the lawyer-
client interaction may temporarily alter the client’s preferences.27
Problem: The Case of Elizabeth Fletcher
We end the discussion of the lawyer as counselor with a hypothetical case that
illustrates the dilemmas of the counseling relationship in circumstances where
deeply held personal values are at stake in situations of risk or uncertainty.
Elizabeth Fletcher is a single woman in her early 40s, who consulted a lawyer
after she had been fired from her job as a customer service representative at a
regional electrical utility company. The reason given for her termination was
excessive absenteeism. Fletcher explained to her lawyer that a year and a half ago
she was diagnosed with an autoimmune disease that causes painful and poten-
tially life-threatening inflammation of the joints, muscles, and other organs,
including the lining around the lung and the heart. Periods during which the
disease is in remission alternate with periods of severe symptoms, requiring
Fletcher to miss more work than would an otherwise similarly situated but non-
disabled employee. She has also had to miss work because of frequent medical
appointments.
Fletcher asked to be allowed to work a flexible schedule as an accommodation
under the Americans with Disabilities Act (ADA). While the human resources
director offered to help her apply for and obtain social security disability benefits
(SSDI), he declined to make this accommodation. After more missed days, the
company terminated her employment.
Fletcher has now consulted a lawyer for help in selecting between two possi-
ble courses of action. She can either file for SSDI, with the company’s coopera-
tion, or she can sue the company under the ADA for failing to accommodate her
need for a flexible work schedule. She thought about doing both simultaneously,
26. Steven Ellman, Lawyers and Clients, 34 U.C.L.A. L. Rev. 717, 733–53 (1987).
27. William H. Simon, Lawyer Advice and Client Autonomy: Mrs. Jones’ Case, 50
Maryland L. Rev. 213, 216–17 (1991).
532 problem solving, decision making, and professional judgment
but her lawyer indicates that the law requires her to choose one option or the
other.
The lawyer, an expert in disability law, knows that a study by the Commission
on Mental and Physical Disability of the American Bar Association28 indicates
that defendants prevail in over 90 percent of litigated ADA cases. Recent amend-
ments to the ADA, providing a broader definition of “disability,” may change the
estimate somewhat, but the particular facts of Fletcher’s case do not seem to
distinguish it from others filed under the ADA, so the lawyer sees little reason to
adjust significantly the 10 percent base rate of anticipated success. Furthermore,
the costs associated with prosecuting the ADA case would be substantial, includ-
ing filing fees, deposition and expert witness costs, and a contingency fee of 30
percent of any ultimate recovery.
The costs associated with applying for social security disability benefits are
much lower and the likelihood of success greater. There is no application fee,
and Fletcher could apply without an attorney. Should she have to file an appeal,
that could be done for a fraction of the cost associated with civil litigation under
the ADA in federal district court. With the employer’s cooperation, the probabil-
ity of having her SSDI application approved are quite high, perhaps 85 percent.
The lawyer tries to provide the data to explain why filing for SSDI has much
greater expected value than suing the employer under the ADA. But Fletcher
becomes increasingly agitated, and eventually says that, while she accepts that
the success rate in ADA cases is low, she is confident that she has a better than
average chance of success. Her former employer is a big company. They could
easily have put her on a more flexible schedule without having to endure an
“undue hardship,” which would provide a defense under the ADA. Besides, if
she has to rely on Social Security, she feels certain that she will never get better.
The very thought of going on “the dole,” she explains, makes her feel depressed;
it’s like acknowledging that her life is over, that she is simply waiting to get
worse and die. “If I don’t keep fighting, I’ll lose the battle with this disease.
People with HIV go on SSDI. That’s when you know they’re on their way out.”
One could easily conclude that Elizabeth Fletcher’s judgment is distorted—in
this case by unrealistic optimism about her chances of winning an ADA suit, and
by the seemingly irrational belief that going on Social Security disability insur-
ance has the power to alter the course of her disease. Yet recall Shelley Taylor’s
research on “positive illusions,” in which cancer patients, crime victims, and
others who had experienced adverse life events adapted by finding meaning in
the adverse experience, by attempting to regain mastery over the negative event
28. See Study Finds Employers Win Most ADA Title I Judicial and Administrative
Complaints, 22 Mental and Physical Disability Law Reporter 403 (May/June 1998)
(nationwide rate of defense judgments in ADA Title I cases that went to judgment
between 1992 and 1997 was 92.11 percent; rate in Fifth Circuit (highest nationwide) was
98.1 percent, in the Ninth Circuit (lowest nationwide) 83.3 percent). Id. at 405.
conclusion 533
in particular and over their lives more generally, and by restoring self-esteem
through self-enhancing evaluations. (See Section 13.5)
One can understand Fletcher’s attraction to litigating the ADA case as a strat-
egy for gaining control and enhancing her self-esteem, even it if involves overop-
timism about her own chances of success. Even if her receipt of SSDI benefits
does not “objectively” affect her health, Fletcher’s belief in the causal connection
may affect her experienced utility of whatever option she eventually chooses to
pursue—and perhaps it will actually come to affect her health as well. On the
other hand, if she files the ADA suit, which she has a low chance of winning, she
will forego disability benefits at least for the duration of the suit, and reduce her
chances for obtaining SSDI for some time thereafter.
How would you counsel Elizabeth Fletcher under the circumstances?
Our own view is that in situations such as these, if the lawyer’s personal or
professional experience leads him to believe that the client will regret a decision
in the long run, it would be irresponsible not to raise his concerns in dialogue
with the client—but with sensitivity to the thin lines between dialogue, persua-
sion, and overbearing influence, and a realization that even raising the question
could affect Fletcher’s well-being.
This page intentionally left blank
part four
guiding and influencing
decisions
The chapters that follow are concerned mainly with changing people’s minds and
behavior—and sometimes, conversely, with preventing others from “messing
with your mind.” As a general introduction to the topic, consider this problem.
Suppose that you are concerned about the rise in obesity among American
youth, especially among disadvantaged children. You are concerned about both
the health consequences and the economic burdens that will ultimately be
imposed on society. Consider these various approaches for promoting healthy
behaviors by your target population.
You might try to change the behavior of children by persuading them to eat more
healthily. You could do this by informing them about the consequences of obe-
sity and how to follow a healthy diet. These efforts might be aided by labeling
foods according to their calories and fat content, though labeling may require
legislation. You might try to inform parents in order to affect their behavior with
respect to their kids’ eating and exercise habits; for example, reducing the time
spent in front of the television improves behavior in both dimensions.
536 problem solving, decision making, and professional judgment
incentives
psychology
1. http:// www.hotchalk.com/mydesk/index.php/the-buzz/418-pay-for-grades-a-
controversial-motivation http://www.nytimes.com/2007/06/19/nyregion/19schools.
html?_r=1.
guiding and influencing decisions 537
assist us in achieving our goals. Our ability to get ahead in life, indeed to survive,
depends on social interactions including our impulse to conform to social norms
and to obey authority.
The techniques of social influence are designed to change beliefs, attitudes,
and behavior through interpersonal interaction2 through such mechanisms as
conformity, obedience, and reciprocity. By contrast, the techniques of framing
and persuasion seek to influence beliefs, attitudes, and behavior through the way
an issue or choice is presented. These techniques build on the insights of JDM
research examined above, including how the number of choices and the way
they are presented—for example, as gains or losses—affect decisions. We include
in this category what Richard Thaler and Cass Sunstein call “choice architecture”—
policy makers’ use of framing devices to assist individuals in pursuing their own
or society’s best interests.3
Part 4 considers how people’s decisions and behavior can be influenced or
guided—for better or worse—by individuals or groups.
Chapter 17, Social Influence, draws on the realm of social psychology and con-
siders how individuals’ behavior is affected by their awareness of and relation-
ships with other people.
Chapter 18, Influencing Behavior through Cognition, draws on cognitive psy-
chology, the JDM literature, and behavioral economics. It considers how argu-
ments can be framed to make them more persuasive, how decisions can be
unconsciously affected by representations or associations placed in memory
shortly before they are made, and how policy makers can use knowledge of the
biases described in earlier parts of the book to guide citizens’ and consumers’
decision making.
Chapter 19, Improving Group Decision Making, considers how one can gain
the benefits and avoid the pathologies of group decision making, and concludes
with a discussion of the effects of accountability on individual and group
decision-making performance.
In the spring of 2004, photographs showing U.S. military police torturing and
humiliating prisoners at the Abu Ghraib prison in Iraq shocked the American
public. Competing explanatory theories quickly filled the news media, personal
conversations, Web sites, and blogs. Some attributed the events to a “few bad
apples” run amok. Others posited a vast military and CIA plan to use torture
systematically as an intelligence tactic. Through this frame, the guards at Abu
Ghraib were just scapegoats, taking a fall for following orders. Other people,
familiar with Philip Zimbardo’s famous “prison study,” conducted in the base-
ment of the psychology department at Stanford University in the early 1970s,1 or
with Stanley Milgram’s electrical shocks studies conducted a decade earlier,
viewed the atrocities at Abu Ghraib differently. To these observers, what hap-
pened at Abu Ghraib was a thoroughly predictable product of a particular align-
ment of social forces that caused ordinary people to do extraordinarily bad
things.2
The desire to understand the levers of social influence produced much of the
most important and interesting social psychological research in the latter half of
the twentieth century. Taken as a whole, this research shows that, just as deci-
sion making can be impaired through the application of cognitive heuristics that
work well most but not all of the time, so can it be impaired by social influence
processes that short circuit clear thinking about a particular action or choice. Of
course, social influences are not all bad. They conduce to harmony and social
order. They sometimes help us take advantage of others’ experience and exper-
tise, thereby avoiding costly mistakes. But processes of social influence have
downsides: they can induce us to make decisions we later regret.
1. In Zimbardo’s study, Stanford undergraduates were assigned the role of either pris-
oner or guard, in an experiment about role conformity that was expected to last two weeks.
Within days, the experiment spun out of control, as “guards” began sexually humiliating
and otherwise abusing “prisoners.” For Zimbardo’s reflections on the similarities between
the behavior of his subjects and the Abu Ghraib guards, see Philip G. Zimbardo, Power
Turns Good Soldiers into “Bad Apples,” Boston Globe, May 9, 2004, Op-Ed.
2. For a thorough treatment of the Stanford prison study and its findings’ power in
explaining Abu Ghraib, see Phillip G. Zimbardo, The Lucifer Effect: Understanding
How Good People Turn Evil (New York: Random House, 2008).
540 problem solving, decision making, and professional judgment
Recall Luis Trujillo, the lawyer introduced in Chapter 1. Project his legal career
back in time fifteen years, and imagine that he is a brand new lawyer at Cooper
& Lytton, a medium-sized Los Angeles law firm specializing in the defense of
white collar criminal cases and the civil lawsuits that often accompany them. It
is Trujillo’s first month in his new job, and he feels utterly at sea. He has to
figure out how to dress, how early to arrive at the office and how late to stay, how
to behave in meetings with partners, clients, and associates. In writing the briefs
he has been assigned to draft, he hasn’t quite figured out where to draw the line
between vigorous advocacy on the one hand and candor to the tribunal on the
other. Asked to respond to discovery requests, he still cannot quite discriminate
between a valid objections on the one hand and unethical dilatory tactics on the
other. Sitting in on a deposition preparation session with one of the firm’s most
important clients, Trujillo is deeply concerned; this looks an awful lot like the
“witness coaching” he found so distasteful in a Paul Newman movie, The Verdict,
that he had watched in his law school professional responsibility class. Every day
at work, Trujillo’s ethical intuitions are challenged, but he is also aware that
there is an awful lot about being a lawyer—a good, ethical, but also tough and
effective lawyer—that he just doesn’t yet understand.
How will the young Luis Trujillo learn to be an ethical but aggressive litiga-
tor? In all likelihood, he will observe what the firm’s more senior lawyers do, and
then he will do the same. Then, as time goes on, he will become a “Cooper &
Lytton litigator.”
Trujillo would not be alone in looking to the actions of others in his social
environment to help him make sense of an uncertain new identity. Much of the
time, particularly in novel situations, we do not know how to behave. We only
hope that we can figure it out before we make a bad impression or worse. In such
situations, we often use the behavior of others like us as a heuristic, a fast and
dirty shortcut for figuring out what to do and how to do it.
Social psychologists, such as Robert Cialdini, refer to this phenomenon as
“social proof,”3 others, like Philip Zimbardo and Susan Fiske, as “conformity.”4
Cialdini defines social proof as the tendency to use the actions of similar others
to decide on proper behavior for ourselves. In other words, people tend to see
behavior as appropriate in a given situation to the degree that they see others
performing it. The behavior of others thus serves as “proof” that it is appropriate.
A B
with the group and gave incorrect responses approximately 30 percent of the
time, despite the obviousness of the correct response.5
Asch’s initial studies generated numerous replications and extensions.6 In
133 studies, spanning 40 years, the average subject conformed 29 percent of the
time. Conformity rates were higher when the stimulus was ambiguous and
when confederates and participants were members of the same social reference
group. Conformity rates were relatively higher for female than for male subjects,
particularly where the object of conformity was gender counterstereotypic, and
were also higher in more collectivist cultures.7
Research on conformity raises interesting and as yet not fully answered ques-
tions. Foremost among these is, when subjects comply in these experiments, are
they actually misperceiving the stimulus, or are they simply going along with the
group, even though they know that the response they are giving is wrong? Or, is
something more complex and nuanced than either one of these two possibilities
going on? The best answer appears to be “all of the above.”
The earliest evidence that conformity effects can change perception itself
emerged from extensions of Muzafer Sherif’s conformity experiments in the
1930s, on which Asch’s research was based.8 In his studies, Sherif used a phe-
nomenon known as the “autokinetic effect,” the illusion of motion created when
one gazes at a stationary point of light. When looking at a point of light and
5. Solomon. E. Asch, Effects of Group Pressures upon the Modification and Distortion of
Judgment, in Groups, Leadership, and Men (Harold S. Guetzkow ed., Pittsburgh:
Carnegie Press, 1951).
6. For a thorough review, see Rod Bond and Peter B. Smith, Culture and Conformity:
A Meta-Analysis of Studies Using Asch’s Line Judgment Task, 119 Psychological Bulletin
111–37 (1996).
7. Id. For a focus on intercultural comparisons, see Heejung Kim and Hazel R.
Markus, Deviance or Uniqueness, Harmon or Conformity? A Cultural Analysis, 77 Journal
of Personality and Social Psychology 785–800 (1999).
8. Muzafer Sherif, The Psychology of Social Norms (New York: Harper & Row,
1936).
social influence 543
subjects resisted the impulse to conform. Another third resisted more often than
they succumbed. Moreover, in follow-up studies, Asch showed that when subjects
had at least one ally in dissent rates of conformity plummeted dramatically.14
Nonetheless, the conformity effect illustrated by Asch’s research has with-
stood the test of time, both in the lab and in the real world. For example, knowl-
edge about the behavior of peers influences peoples decisions about whether to
help in an emergency,15 pay their taxes,16 or properly dispose of their trash.17
14. Solomon E. Asch, Opinions and Social Pressure, Scientific American 31–35
(November 1955).
15. Bibb Latané and John M. Darley, The Unresponsive Bystander: Why Doesn’t
He Help? (Englewood Cliffs, NJ: Prentice-Hall, 1970).
16. John T. Scholz, Kathleen M. McGraw, and Marco R. Steenbergen, Will Taxpayers
Ever Like Taxes? Responses to the U.S. Tax Reform Act of 1986, 13 Journal of Economic
Psychology 625–56 (1992).
17. Robert B. Cialdini, Raymond R. Reno, and Carl A. Kallgren, A Focus Theory of
Normative Conduct: Recycling the Concept of Norms to Reduce Littering in Public Places, 58
Journal of Personality and Social Psychology 1015–26 (1990).
18. Bibb Latané and John M. Darley, The Unresponsive Bystander: Why Doesn’t
He Help?, supra note 15; John M. Darley and Bibb Latané, Bystander Intervention in
Emergencies: Diffusion of Responsibility, 8 Journal of Personality and Social Psychology
377–83 (1968).
social influence 545
19. Bibb Latané and Steve Nida, Ten Years of Research on Group Size and Helping, 89
Psychological Bulletin 308–24 (1981).
20. Kurt Eichenwald, Conspiracy of Fools: A True Story (New York: Broadway
Books, 2005).
546 problem solving, decision making, and professional judgment
21. Donald C. Langevoort, Where Were The Lawyers? A Behavioral Inquiry into Lawyers’
Responsibility for Clients’ Fraud, 46 Vanderbilt. L. Rev. 75, 105–106 (1993).
22. Jessica M. Nolan, Wesley Schultz, Robert B. Cialdini, Noah J. Goldstein, and
Vladas Griskevicius, “Normative Social Influence Is Underdetected” (unpublished manu-
script 2007).
social influence 547
affected by norms of people who are similar to them and groups with which they
identify.
This category of “similar” others can actually extend to complete strangers,
who happen to be (or to have been) in situations similar to our own. For exam-
ple, Noah Goldstein, Robert Cialdini, and their colleagues found that hotel guests
were most likely to participate in a towel reuse program when they were informed
that the majority of people who had stayed in the same room had participated in
the program.23 This message was even more effective than one providing the
same information about the behavior of people of one’s gender, an identity that
generally seems more salient than people who happened to have occupied the
same hotel room. Goldstein, Cialdini, et al. suggest that when we are deciding
how to behave in a particular situation, the social information that is most rele-
vant is how other people behave in that same situation.
Descriptive norm information has also been found to affect charitable giving.
Rachel Croson and her colleagues collaborated with public radio stations to con-
duct research during the stations’ “pledge drives.” In these studies, some listen-
ers who called in to make donations were randomly selected to receive different
information regarding donation amounts. For example, before being asked how
much money they would like to donate, some listeners were told that another
listener had just donated $200. Listeners receiving this information donated, on
average, $30 more than callers who were simply asked how much they would
like to donate. (The researchers controlled for the anchoring effect.)
In many situations, it is possible to make use of these sorts of techniques
without engaging in deception. In the absence of other data, information about
one individual can be taken as an indication of the norm. And it is highly likely
that radio stations and other organizations that depend on charitable donations
can truthfully provide information about at least one generous donor. Moreover,
in some cases, such as Cialdini’s study of neighborhood energy conservation,
many people actually may engage in a desired behavior.
23. Noah J. Goldstein, Robert B. Cialdini, and Vladas Griskevicius, A Room with a
Viewpoint: Using Normative Appeals to Motivate Environmental Conservation in a Hotel
Setting, 35 Journal of Consumer Research (2008) (electronically published without
page numbers).
548 problem solving, decision making, and professional judgment
Timur Kuran and Cass Sunstein describe three different types of cascades:
informational cascades, reputational cascades, and availability cascades, all three of
which can derive from conformity-related effects.24 An informational cascade
occurs when people with incomplete personal information about a subject base
their own beliefs about it on the apparent beliefs of others in their social environ-
ment. People’s words and actions often convey the impression that they hold a
particular belief. Cascade theory posits that in response to such communica-
tions, other individuals, who know that they personally lack reliable information
on the subject, accept a particular belief simply by virtue of its apparent accep-
tance by others. Under proper conditions, such a belief, no matter how fanciful,
can spread rapidly and widely through a population.
Reputational cascades play a significant role in this process. Even if they doubt
the veracity of the belief spreading by means of an informational cascade, people
may act and speak as if they believe it to be true. Originally, at least, they do this
to earn social approval and avoid social disapproval. Of course, once a person
adopts and repeats a belief, even if only to avoid censure, he or she may eventu-
ally come to believe it true, if only to reduce the cognitive dissonance that would
otherwise result.25 Reputational cascades can prove very powerful, as direct social
pressures to adopt a fast-propagating belief reinforce concerns that deviance
would have negative long-term effects on social networks.26
Informational and reputational cascades look similar, often co-occur, and
drive the same rapid propagation of beliefs. As theoretical constructs, they differ
only in their motivation. People participate in an informational cascade primar-
ily because they view adoption of an apparently sound belief held by others as
rational, given their own lack of information and the costs associated with inde-
pendent information search. For this reason, information cascades are also
sometimes referred to as “rational herding.”27 In contrast, people participate in
reputational cascades to avoid social disapproval, or to enhance their standing in
a community.
Informational and reputational cascades sometimes interact with the availabil-
ity heuristic (Section 9.6) in ways that create what Kuran and Sunstein call an
availability cascade. In these situations, informational and reputational cascades
24. Timur Kuran and Cass R. Sunstein, Availability Cascades and Risk Regulation, 51
Stanford Law Review 683 (1999).
25. Daryl J. Bem, Self-Perception Theory, in Advances in Experimental Social
Psychology 6, 60–61 (Leonard Berkowitz ed., New York: Academic Press, 1972).
26. Michael Klausner, Corporations, Contracts, and Networks of Contracts, 81 Virginia
Law Review 757 (1995); Marcel Kahan and Michael Klausner, Path Dependence in Corporate
Contracting: Increasing Returns, Herd Behavior and Cognitive Bias, 74 Washington Law
Quarterly 347 (1996).
27. Eric Talley, Precedential Cascades: An Appraisal, 73 Southern California Law
Review 87 (1999).
social influence 549
***
In summary, people look to the behavior of others in their social environ-
ments as a kind of heuristic, applied to resolve uncertainties over what they
should think or how they should behave. Other things being equal, tendencies
toward social conformity are stronger when people are in novel environments, or
where behavioral norms are particularly complex and ambiguous, or where the
negative sanctions attached to behavioral deviance are particularly harsh. This
tendency to conform our own behavior—and our sense of what is right, norma-
tive, or accurate—makes it possible for us to get along with other people, master
new roles and environments, and achieve our individual and collective goals. But
it can also cause us to behave in ways we later recognize as abhorrent.
Unfortunately, that recognition often comes too late.
550 problem solving, decision making, and professional judgment
At the end of World War II, descriptions of Nazi genocide shocked American
society and triggered, in academic communities as elsewhere, attempts to explain
why so many people collaborated with Hitler’s atrocities. In 1950, a group of
social scientists led by German exile, Theodor Adorno, provided one explanation
in The Authoritarian Personality.28 They posited that certain people possessed par-
ticularly rigid personalities, characterized by social conservatism, conformity,
authoritarianism, and a strong inclination to obey authority. This personality
type, Adorno concluded, was widespread in Germany in the 1930s and ‘40s and
made National Socialism possible.
Building on Adorno’s notion that nationality was correlated with conformity
and compliance with authority social psychologist Stanley Milgram later set out
to demonstrate that the more “contentious” French were less likely to conform
in an Asch paradigm than the more “retiring” Scandinavians.29 After failing to
find significant differences between French and Scandinavian subjects, Milgram
then designed an experiment to identify the “type of person” who would obey the
instructions of an apparent authority figure—even if the instruction was to do
something harmful to another person.
In Milgram’s now classic study,30 subjects were asked to play the role of
“teacher” by asking questions of a “learner” who was hooked to an instrument
that vaguely resembled an electric chair. Upon first entering the laboratory, the
subject, or “teacher,” assisted an experimenter, dressed in a white lab coat, in
strapping the “learner” into the chair and attaching a shocking device to his arm.
(See Figure 17.3.)
The “teacher” was then led into an adjacent room and seated in front of a
panel with a row of switches on it, each marked with a particular voltage level,
ranging from 15 to 450 volts and rising in 15 volt intervals. Under the switches
were descriptions of the levels of the shock each would trigger, ranging from
“slight shock” to “danger: severe shock.” The last two switches were ominously
labeled “XXX.”
The experimenter then explained to the subject that he should ask the learner
a question and, if the learner failed to respond correctly, the subject should
administer a shock. Each time the learner failed to respond correctly, the shock
to be administered was to increase by 15 volts.
Milgram’s subjects were not college students. They were adult men from
diverse occupational backgrounds who responded to a newspaper advertisement
for a study on “learning” to be conducted at Yale University.
Unbeknownst to subjects, the learner was actually a confederate of the exper-
imenter. Prerecorded sounds (e.g., screams, protests) reflecting varying degrees
of pain were played, depending on the level of voltage the subject was “applying”
to the learner. When the subject indicated a desire to stop either questioning the
learner or applying the shock, he was prompted with one of four cues, depend-
ing on the nature of the subject’s hesitation: “Please continue”; “The experiment
requires that you continue”; “It is absolutely essential that you continue”; or
“You have no other choice, you must go on.”
Surprisingly, even though they were not coerced or threatened into obeying
the experimenter’s instructions, nearly 65 percent of the subjects were willing to
continue until they had applied what appeared to be life-threatening levels of
electricity to those who failed to provide correct answers. Milgram reported that
“subjects were observed to sweat, tremble, stutter, bit[e] their lips, groan, and dig
their fingernails into their own flesh.” Some subjects cried; others burst out into
hysterical laughing fits. In post-experiment debriefings, nearly all the subjects
reported that they felt that it was wrong to apply the shocks.
Neither Milgram nor any of the social psychologists, psychiatrists, or others
he consulted anticipated that people would comply at such high rates, or to so
great an extent. And yet, in subsequent replications the results were similar at
different locations with different subjects. By the end, Milgram and his associ-
ates had run literally thousands of people from a wide cross-section of American
society, including postal workers, college students, high school students and
teachers, construction workers, college professors, and salespeople. Obedience
was the norm. As Milgram concluded, “It is the extreme willingness of adults to
go to almost any lengths on the command of an authority that constitutes the
chief finding of the study.”31
Obedience effects are robust across many situations and groups of subjects.
However, as is described in greater detail below, subtle changes in the situation
can have profound effects on obedience behavior. Individual differences play
some role as well.32 Not surprisingly, authoritarianism33 sometimes predicts
higher levels of obedience. Trust, moral development, and conventionality also
show some (albeit weak or inconsistent) effects on obedience levels.34
All but one of Milgram’s trials included only male subjects, but most of the
experimental evidence shows no gender effect on obedience to apparent
authority,35 In the one trial including females, Milgram found no sex differences
in obedience rates.36 Ten of eleven replications found the same result.37 The one
31. Ibid.
32. For a review of the literature, see Thomas Blass, Understanding Behavior in the
Milgram Obedience Experiment: The Role of Personality, Situations, and Their Interactions, 60
Journal of Personality and Social Psychology 398–413 (1991).
33. Theodor W. Adorno, Else Frenkel-Brunswik, Daniel J. Levinson, and
R. Nevitt Sanford, The Authoritarian Personality (New York: Harper, 1950).
34. See Thomas Blass, The Migram Paradigm after 35 Years: Some Things We Now Know
about Obedience to Authority, 29 Journal of Applied Social Psychology 955–78 (1991).
35. Ibid. 968–69.
36. Stanley Milgram, Obedience to Authority: An Experimental View (New York,
NY: Harper & Row, 1974).
37. Thomas Blass, supra note 34, 968.
social influence 553
38. Wesley Kilham and Leon Mann, “Level of Destructive Obedience as a Function of
Transmitter and Executant Roles in the Milgram Obedience Paradigm,” Journal of
Personality and Social Psychology, 29 (1974): 696–702.
39. Charles K. Hofling et al., An Experimental Study in Nurse-Physician Relationships,
143 Journal of Nervous and Mental Disease 171–80 (1966).
40. Neil M. Davis and Michael R. Cohen, Medication Errors: Causes and
Prevention (Philadelphia: G. F. Stickley, 1981).
41. Adrian J. Barrio, Rethinking Schneckloth v. Bustamonte: Incorporating Obedience
Theory into the Supreme Court’s Conception of Voluntary Consent, University of Illinois
Law Review 215 (1997).
42. C. A. Elizabeth Luus and Gary L. Wells, Eyewitness Identification Confidence, in
Adult Eyewitness Testimony: Current Trends and Developments (David F. Ross, J.
Don Read, and Michael P. Toglia eds., New York: Cambridge University Press, 1994).
554 problem solving, decision making, and professional judgment
troubling evidence of the extent to which obedience can lead us astray. One of
the military police who served in Abu Ghraib reported that they were told by
superiors, “Use your imagination. Break them. We want them broke by the time
we come back.”43 The deadly environment, the pressure for intelligence, and the
ambiguity of how they were supposed to “break” the prisoners led to the abuse,
torture, and even death of detainees.
43. Phillip C. Zimbardo, The Lucifer Effect: Understanding How Good People
Turn Evil, supra note 3, 352.
44. Id. at 330.
45. Id. at 474.
social influence 555
In Section 14.3 we discussed various factors that lead people to take “sunk costs”
into account when making decisions about the future. We continue that discus-
sion here, focusing more on social determinants.
Once we have taken a position, whether through an activity or communica-
tion, we tend to behave in a manner consistent with that position and to accede
to requests that are consistent with it. Why people might act this way is no mys-
tery: the social costs of changing one’s mind or of otherwise appearing inconsis-
tent can be steep.
The key to this foot-in-the-door technique is that compliance with the first
request causes the target to think of himself in a way that makes him more likely
to accede to the second through a process of self-attribution.51 That is, by freely
agreeing to perform and then performing the first action, the foot-in-the-door
target comes to view himself as “the kind of person” who would do things of that
kind or support the position expressed by the first action.
Fundraisers use the foot-in-the-door technique all the time, which is why your
graduate school will try to get you to make a donation, no matter how small,
before you have earned your first dollar as a lawyer or policy professional—
perhaps even while you are still paying tuition. Lawyers and investigators use the
technique, too, for example, when they are trying to persuade reluctant witnesses
to discuss a case with them. Later, perhaps even in the same case, a mediator
may use the technique in an attempt to help the parties reach a settlement.
Police officers use the foot-in-the-door technique when dealing with suspects.
For instance, a police officer may ask permission to enter a residence “just to
talk” (first request). Once inside, and after some friendly conversation, she may
then ask for consent to conduct a search (bigger second request). One can readily
see from this example how various social influence factors, in this case both
obedience to authority and the consistency and commitment tendency, could
work together to amplify compliance.
51. Donald R. Gorassini and James M. Olson, Does Self-Perception Change Explain the
Foot-in-the-Door Effect?, 69 Journal of Personality and Social Psychology 91 (1995).
52. Robert B. Cialdini, John T. Cacioppo, Rodney Bassett, and Geoffrey A. Miller, Low-
Ball Procedure or Producing Compliance: Commitment then Cost, 36 Journal of Personality
and Social Psychology 463–76 (1978).
558 problem solving, decision making, and professional judgment
As Richard Birke and Craig Fox have observed, foot-in-the-door, low-ball, and
other aspects of the consistency and commitment tendency often converge in
affecting the parties’ behavior in mediations.53 First, mediators often begin by
inducing the parties to commit to a general statement of principles, or an aspira-
tion. Once the mediator has this foot in the door, she is more likely to induce
party compliance with later, more substantial requests. The low-ball technique
also often comes into play, as Birke and Fox note. Once the parties have invested
significant time and energy into crafting a tentative agreement, they can often be
induced to agree to last-minute modifications, even if they make the deal less
beneficial to them.
Although the vast majority of civil cases settle, they often settle very late, after
enormous amounts of money have been spent on pleadings, discovery, and
motion practice.58 There are, of course, many reasons for this, including the asym-
metrical valuation of losses and gains, the financial incentives motivating lawyers
who bill by the hour, the persistence of information asymmetries between the
parties, and the various cognitive distortions that comprise what was referred to
in Section 14.3 as the sunk costs fallacy. But the escalation of commitment likely
also plays a role in keeping cases from settling optimally.
56. See Max Bazerman, Judgment in Managerial Decision Making (5th ed., New
York: Wiley, 2002) for a review.
57. Barry M. Staw and Jerry Ross, Behavior in Escalation Situations: Antecedents,
Prototypes, and Solutions, in Research in Organizational Behavior (Barry M. Staw and
Larry L. Cummings eds., Greenwill, CT: JAI Press, 1987); Joel Brockner and Jerry Z.
Rubin, Entrapment in Escalating Conflicts (New York: Springer-Verlag, 1985).
58. Samuel R. Gross and Kent D. Syverud, Getting to No: A Study of Settlement
Negotiations and the Selection of Cases for Trial, 90 Michigan Law Review 319–93 (1991).
560 problem solving, decision making, and professional judgment
59. Barry M. Staw and Jerry Ross, Behavior in Escalation Situations: Antecedents,
Prototypes, and Solutions, supra note 57; John Platt, Social Traps, 28 American Psychologist
642–43 (1973).
60. Gregory B. Northcroft and Gerrit Wolf, Dollars, Sense, and Sunk Costs: A Life Cycle
Model of Resource Allocation Decisions, 9 Academy of Management Review 225–34
(1984).
61. Barry M. Staw and Jerry Ross, Behavior in Escalation Situations: Antecedents,
Prototypes, and Solutions, supra note 57; John Platt, Social Traps, supra note 59.
62. Barry M. Staw, Knee-Deep in the Big Muddy: A Study of Escalating Commitment to a
Chosen Course of Action, 16 Organizational Behavior and Human Performance 27–44
(1976); Max Bazerman, R.I. Beekum, and F. D. Schoorman, Performance Evaluation in a
Dynamic Context: A Laboratory Study of the Impact of Prior Commitment to the Ratee, 67
Journal of Applied Psychology 873–876 (1982).
63. Richard E. Nisbett and Lee Ross, Human Inference: Strategies and
Shortcomings of Social Judgment (Englewood Hills, NJ: Prentice-Hall, 1980).
64. Barry M. Staw and Jerry Ross, Behavior in Escalation Situations: Antecedents,
Prototypes, and Solutions, supra note 57.
65. Ibid.
66. Lynn G. Zucker, Organizations as Institutions, in Research in The Sociology
of Organizations 2, 53–111 (Samuel B. Bacharach ed., Greenwich, CT: JAI Press, 1981);
Paul S. Goodman, Max Bazerman, and Edward Conlon, Institutionalization of Planned
Organizational Change, in Research in Organizational Behavior: Vol. 2, 215–46
(Barry M. Staw and Lynn L. Cummings eds., Greenwich, CT: JAI Press, 1980).
social influence 561
67. Barry M. Staw and Jerry Ross, Behavior in Escalation Situations: Antecedents,
Prototypes, and Solutions, supra note 57; Gerald R. Salancik, Commitment and the Control of
Organizational Behavior and Belief, in New Directions in Organizational Behavior
1–54 (Barry M. Staw and Gerald R. Salancik eds., 1977).
562 problem solving, decision making, and professional judgment
• Calculate the costs and make a plan for closing at the outset of the
project.
17.4 liking
We tend to like those who like us. We tend to keep track of “credits and debits”
with people we are not close to, but are less likely to do so for those we like a lot.
We tend to credit views expressed by people we like and discredit views expressed
by people we dislike. We prefer to say “yes” to those we like. For these and other
reasons, “liking” can impair our judgments and make us more susceptible to
influence.68 Several factors promote liking, including: familiarity, similarity,
cooperation, compliments, and physical attractiveness.69
17.4.1 Familiarity
All else being equal, we tend to like people and things that are more, rather
than less, familiar. As the late Stanford psychologist Robert Zajonc demon-
strated, mere exposure to an initially neutral or favorable stimulus enhances its
evaluation.70 So, for example, people express greater liking of, and express higher
levels of agreement with, people whose faces they have encountered frequently,
even in subliminal exposures.71 These sorts of exposure effects appear to be even
stronger in the real world than they are in the laboratory.72
68. A study described in Section 17.5 leaves open the question of how effective liking
is in gaining compliance. Dennis T. Regan, Effects of a Favor and Liking on Compliance, 7
Journal of Experimental Pscyhology 627–39 (1971). Liking for a male confederate was
manipulated (i.e., the subject overheard him behave in either a “pleasant” or “nasty” way
to someone with whom he was speaking on the phone), and the participants received a soft
drink (a favor) from either the confederate, or from the experimenter, or received no favor
at all. The confederate then asked the participants to purchase some raffle tickets. Regan
found that the favor strongly increased compliance with the request, but he did not find
a similar effect for liking. However, he cautioned that “[i]t would not be warranted to con-
clude, on the basis of the manipulation used in this study, that liking does not generally
affect compliance.” Id. at 635.
69. Different social psychologists proffer slightly different taxonomies. Ours repre-
sents a combination of Robert Cialdini’s and Susan Fiskes’. See Robert Cialdini,
Influence: Science and Practice 136–53 (3rd ed. New York: HarperCollins, 1993) and
Susan T. Fiske, Social Beings, supra note 4, 257–276.
70. Robert. B. Zajonc, Attitudinal Effects of Mere Exposure, 9 Journal of Personality
and Social Psychology 1–27 (1968).
71. Robert F. Bornstein, Dean R. Leone, and Donna J. Galley, The Generalizability of
Subliminal Mere Exposure Effects: Influence of Stimuli Perceived without Awareness on Social
Behavior, 53 Journal of Personality and Social Psychology 1070–79 (1987).
72. Robert F. Bornstein, Exposure and Affect: Overview and Meta-Analysis of Research
1968–1987, 106 Psychological Bulletin 265–89 (1989).
social influence 563
17.4.2 Similarity
Not surprisingly, we tend to like those who share our cultures, values, and beliefs
more than we like people who don’t. Common notions that “opposites attract”
notwithstanding, the vast weight of social psychological evidence supports the
proposition that people are preferentially attracted to and are more satisfied with
their interactions with people whose attitudes, interests, and personalities are
similar to their own.73 With respect to the effect of attitude similarity on liking,
proportions are critical. “If a person only agrees with us on twelve out of twenty-
four topics, he or she is not liked as much as another who agrees with us on four
out of six topics.”74
The similarity principle has obvious legal policy implications, as various empir-
ical researchers have demonstrated. In one such study, mock jurors who per-
ceived a defendant as sharing many of their beliefs and attitudes were less inclined
to find him guilty and were more inclined toward leniency in sentencing.75
Results like these have obvious implications for jury selection, judicial qualifica-
tions, and legal advocacy.76
73. Ellen Berscheid and Harry T. Reis, Attraction and Close Relationships, in Handbook
of Social Psychology 193–281 (Daniel. T. Gilbert, Susan T. Fiske, and Gardner Lindzey
eds., 4th ed. New York: McGraw-Hill, 1988).
74. Curt Bartol and Anne Bartol, Psychology and Law: Theory, Research and
Application, 3d Edition (Florence, KY: Wadsworth Publishing, 2004)
75. William Griffitt and Thomas Jackson, Simulated Jury Decisions: The Influence of
Jury-Defendant Attitude Similarity-Dissimilarity, 1 Social Behavior and Personality
(1973).
76. Curt Bartol and Anne Bartol, Psychology and Law: Research and
Application, supra note 75.
77. Susan T. Fiske, supra note 4, 270.
78. See Richard Birke and Craig R. Fox, Psychological Principles in Negotiating Civil
Settlements, 1 Harvard Negotiation and Law Review 1–57 (1999); Max H. Bazerman
and Margaret A. Neale, The Role of Fairness Considerations and Relationships in a Judgmental
Perspective of Negotiation, in Barriers to Conflict Resolution (Robert H. Mnookin,
564 problem solving, decision making, and professional judgment
shown that those who are cooperative are perceived by others as more effective,
on average, than lawyers who are not.79
17.4.4 Compliments
In the psychological literature, the use of compliments is classified under “ingra-
tiation tactics.” Ingratiation tactics are a class of strategic behaviors that are
designed to elicit liking from the target.80 The use of compliments to engender
liking is an artful skill, for if the tactic “becomes obvious to a target (entirely
transparent) [it] is likely to fail miserably at increasing a target’s liking for the
ingratiator,” and may even be counterproductive. To avoid transparency, for
example, a low-status person who can obviously gain from successful ingratia-
tion may do best to compliment a high-status person in an indirect manner,
such as by telling a third person what they like about the high-status person in a
way that allows the high-status person to “overhear” the compliment.81
Lee Ross, Kenneth J. Arrow, and Amos Tversky, eds., Stanford, CA: Stanford Center on
Conflict and Negotiation, 1995), 86–107.
79. See Gerald R. Williams, Legal Negotiation and Settlement 19 (St. Paul,
Minnesota: West Publishing, 1983).
80. Other ingratiation tactics include opinion conformity, rendering favors, self-
deprecation, and modesty. See Randall A. Gordon, Impact of Ingratiation on Judgments and
Evaluations: A Meta-Analytic Investigation, 54 Journal of Personality and Social
Psychology 71 (1996).
81. However: “Developing a relationship takes time and must progress at its own pace.
In fact, pursuing friendliness as an influence tactic usually requires that the relationship
between the agent and the target already be in place before the request is made so the
relationship can be used effectively. If the agent tries to cultivate a relationship very quickly
and to use it simply as a vehicle in which to lodge the influence request, it is likely that the
target will see the friendliness gestures as superficial and insincere, a perception that will
raise the target’s defensiveness rather than lower it.” Roy J. Lewicki, Joseph A. Litterer,
John W. Minton and David M. Saunders, Negotiation 316 (McGraw Hill, 1994).
82. Alice H. Eagly, Richard D. Ashmore, Mona G. Makhijani, and Laura C. Longo,
What Is Beautiful Is Good: A Meta-Analytic Review of Research on the Physical Attractiveness
Stereotype, 110 Psychological Bulletin 109–28 (1990).
83. See Michael G. Efran, The Effect of Physical Appearance on the Judgment of Guilt,
Interpersonal Attraction, and Severity of Recommended Punishment in a Simulated Jury
Task, 8 Journal of Research in Personality 45–53 (1974) (finding that attractive female
social influence 565
that, when other variables are controlled for, less physically attractive defendants
generally receive more severe sentences.84 In the civil litigation context as well,
the evidence that physically attractive individuals experience better legal out-
comes is both compelling and disturbing.85
Similar effects have been found in the context of hiring, voting, and helping
behavior. In all three contexts, decision makers seem unaware of the effect of
physical attractiveness on their judgment.86
Some qualifications to the case for physical attraction effects are in order here.
First, although physical attractiveness powerfully influences people’s judgments
of others, notions of attractiveness—and the extent of their impact on person
judgment—vary cross-culturally. For example, Chinese immigrants to Canada
did not show the effect in a study by Karen K. Dion, A. Won-Ping Pak, and
Kenneth L. Dion.87 In a Taiwanese study, the effect of attractiveness was moder-
ated by Western value orientation (perhaps showing the independent effect of
familiarity).88
Second, people who score higher in measures of self-consciousness show
stronger effects of physical attractiveness on their judgments than do people
defendants were less likely than unattractive female defendants to be found guilty by,
and receive lighter sentences from, male jurors); Martin F. Kaplan and Gwen D.
Kemmerick, Juror Judgment as Information Integration: Combining Evidential and Non-
evidential Information, 30 Journal of Personality and Social Psychology 493 (1974);
Gloria Leventhal and Ronald Krate, Physical Attractiveness and Severity of Sentencing, 40
Psychol. Rep. 315, 315–17 (1977) (finding that mock jurors recommended shorter sen-
tences for physically attractive defendants than for physically unattractive defendants,
regardless of juror’s gender, defendant’s gender, or seriousness of offense); John E.
Stewart, II, Defendant’s Attractiveness as a Factor in the Outcome of Criminal Trials: An
Observational Study, 10 J. Applied Soc. Psychol. 348 (1980); Wayne Weiten, The Attraction-
Leniency Effect in Jury Research: An Examination of External Validity, 10 J. Applied Soc.
Psychol. 340 (1980).
84. See, for example, J. E. Stewart II, Defendant’s Attractiveness as a Factor in the Outcome
of Criminal Trials: An Observational Study, 10 Journal of Applied Social Psychology
348–61 (1980).
85. For a literature review, see A. Chris Downs and Phillip M. Lyons, Natural
Observations of the Links Between Attractiveness and Initial Legal Judgments, 17 Personality
and Social Psychology Bulletin 541–47 (1990).
86. The studies are collected in Robert B. Cialdini, supra note 3, 141–42.
87. Karen K. Dion, A. Won-Ping Pak, and Kenneth L. Dion, Stereotypic Physical
Attractiveness: A Sociocultural Perspective, 21 Journal of Cross-Cultural Psychology
158–79 (1990).
88. David R. Shaffer, Nicole Crepaz, and Chien-Ru Sun, Physical Attractiveness
Stereotyping in Cross-Cultural Perspective: Similarities and Differences Between Americans and
Taiwanese, 31 Journal of Cross-Cultural Psychology 557–82 (2000).
566 problem solving, decision making, and professional judgment
who score lower.89 This suggests that liking based on physical attractiveness is
mediated by the salience of motives toward self-enhancement.
These caveats aside, existing experimental evidence—in both the laboratory
and the field—suggests that physical attractiveness is a significant determinant
of liking, which in turn has a powerful effect on judgment and choice.
17.5 reciprocity
89. Mark Snyder, Ellen Berscheid, and Peter Glick, Focusing on the Exterior and the
Interior: Two Investigations of the Initiation of Personal Relationships, 48 Journal of
Personality and Social Psychology 1427–39 (1985).
90. Dennis T. Regan, Effects of a Favor and Liking on Compliance, 7 Journal of
Experimental Social Psychology 627–39 (1971).
social influence 567
In one condition, subjects were led to believe that the confederate would know
whether they made a pledge (the public compliance condition). In another (the
private compliance condition), subjects were led to believe that the confederate
would not. The presence of the favor increased compliance in both public and
private conditions. That is, people receiving an unsolicited favor were more likely
to comply with subsequent requests from the person doing the initial favor,
regardless of whether the compliance was public or private. But subjects in the
public compliance condition donated significantly more money than did those
in the private compliance condition.91
We reciprocate concessions as well as favors, as demonstrated by research on
what Robert Cialdini calls the door-in-the-face technique. The basic insight here is
that a person is more likely to induce compliance with a request for a small favor
if she asks for a more extreme favor first. The technique works as follows: the
persuader presents the target with a significant request, something she knows
the target will turn down (the door in the face). Then, after the target refuses, the
persuader makes the smaller request, the one she was actually interested in all
along. Viewing the second request as a concession on the persuader’s part, the
target feels compelled to accede.
In the classic study of this door-in-the-face effect, Cialdini and his associates
asked college students who were walking across campus if they would be willing
to accompany a group of juvenile offenders on a day trip to the zoo.92 Only sev-
enteen percent said “yes.” In a second condition, students were first presented
with a more extreme request: would they be willing to spend two hours a week
for two years as unpaid counselors to juvenile offenders? After refusing this
extreme request, subjects in this second condition were asked to take part in the
zoo trip. Fifty percent agreed so to do.
The door-in-the-face technique prompts action—not merely a verbal agree-
ment to act. For example, researchers investigated whether door-in-the-face vic-
tims who had agreed to work for two unpaid hours in a community mental-health
agency actually showed up to perform their duties as promised.93 The tactic of
starting with a larger request (to volunteer for two hours of work per week in the
agency for at least two years) produced more verbal agreement to the smaller
retreat request (76 percent) as compared to the tactic of asking for the smaller
request alone (29 percent). Even more remarkable was the rate of follow-through.
91. Mark A. Whately, J. Matthew Webster, Richard H. Smith, and Adele Rhodes, The
Effect of a Favor on Public and Private Compliance: How Internalized Is the Norm of
Reciprocity?, 21 Basic and Applied Social Psychology 251 (1999).
92. Robert B. Cialdini et al., Reciprocal Concessions Procedure for Inducing Compliance:
The Door-in-the-Face Technique, 31 Journal of Personality and Social Psychology 206
(1999).
93. Ibid.
568 problem solving, decision making, and professional judgment
17.6 conclusion
Humans are social animals—we care a great deal about what other people think
of us. There is nothing irrational in this. More often than not, our personal out-
comes depend on the regard in which others hold us, and on their willingness to
cooperate, promote our interests, and assist us in achieving our goals. It should
come as no surprise, then, that processes of social influence play a major role in
shaping behavior.
Social influence processes affect the work of lawyers and policy makers in
significant ways. For this reason, it is important to understand them. Lawyers
can affirmatively use influence mechanisms to enhance their persuasiveness
when advocating on behalf of their clients. On the defensive front, lawyers can
use their knowledge of social influence to guard against situations in which third
parties may subtly attempt to alter their attitudes, or explain to clients ways in
which others might attempt to influence them to act against their own best inter-
ests. The ability to recognize and react intelligently to various sources of social
influence makes the difference between being at their mercy and resisting
their appeal.
social influence 569
Policy makers, too, must understand and be able to use the levers of social
influence. Policy initiatives, no matter how well intended or otherwise sound,
may founder if they fail to take such forces as social proof into account. Conversely,
the achievement of policy goals may be dramatically enhanced though the skilled
deployment of the tendency toward consistency and commitment, or liking, or
the triggering of norm cascades. Social influence is integral to persuasion, and
persuasion—both to effecting it and resisting it.
The various cognitive processes mediating message processing, persuasion,
and behavioral choice can be either systematic or heuristic in nature.94 Systematic
processing is a System II function—motivated, deliberate, conscious, and
demanding of cognitive resources. Heuristic processing is a System I phenom-
enon, which can function automatically, with little awareness, and requiring
only minimal cognitive effort. In some situations, a person might apply a quick
persuasion rule, such as “believe the authority,” or “if everyone else thinks it’s
right, it must be,” instead of carefully analyzing a message’s merit. Various fac-
tors, such as the listener’s level of distraction95 or the message’s relevance to the
listener96 play a role in determining which type of processing, systematic or heu-
ristic, occurs. Similar processes—either systematic or heuristic, mediate behav-
ioral choice.
As this discussion suggests, social influence factors function in much the
same way as other mental heuristics. In a message and choice-saturated world,
shortcuts are essential. And given the social worlds in which we live, cooperation
with and at times deference to others is both rational and effective. But, like all
heuristics, social influences can lead us into errors, so we should strive to under-
stand and seek to control their effects.
94. Shelly Chaiken, The Heuristic Model of Persuasion, in Social Influence: The
Ontario Symposium Vol. 5, 3–39 (Mark P. Zanna, James M. Olson, and C. P. Herman
eds., Hillsdale, N.J.: Erlbaum, 1987); John T. Cacioppo, Richard E. Petty, Chuan Feng Kao,
and Regina Rodriguez, Central and Peripheral Routes to Persuasion: An Individual Difference
Perspective, 51 Journal of Personality and Social Psychology 1032–43 (1986).
95. Richard. E. Petty, Gary L. Wells, and Timothy C. Brock, Distraction Can Enhance or
Reduce Yielding to Propaganda: Thought Disruption versus Effort Justification, 34 Journal of
Personality and Social Psychology 874–84 (1976).
96. Richard E. Petty, John T. Cacioppo, and Rachel Goldman, Personal Involvement as
a Determinant of Argument Based Persuasion, 41 Journal of Personality and Social
Psychology 847–55 (1981).
This page intentionally left blank
18. influencing behavior through cognition
This chapter asks how lawyers and policy makers can influence decision making
by citizens, consumers, legislators, judges, and other officials through approaches
that are mediated by people’s judgment and perception rather than social influ-
ence. We begin with argument, the paradigmatic form of persuasion, and then
turn to priming, and end by considering how policy makers can influence citi-
zens’ and consumers’ behavior by framing the context for their choices.
The chances are that readers of this book—whether students, lawyers, or policy
makers—spend much of their time making and listening to arguments based on
reason and evidence. But arguments can be more or less persuasive indepen-
dent of their merits. Paradoxically, to ensure that meritorious arguments prevail,
it often helps to use communication techniques that have nothing to do with the
merits.
The study of persuasive argument dates back at least to Aristotle’s Rhetoric.
Our focus here is on the psychology of persuasion, in which argument is some-
times modeled as a multistage process:1
1. Exposure to the message.
2. Attention to the message.
3. Comprehension of the message.
4. Acceptance of the message’s conclusion.
5. Memory of the message or its conclusion.
6. Action based on the message.
Much of the psychological research on rhetoric focuses on how communicators
can increase recipients’ comprehension, acceptance, and memory of the message.
For example, in their popular book, Made to Stick: Why Some Ideas Survive and
Others Die, Chip and Dan Heath provide a practical guide to making arguments
memorable and influential.2
Simplicity. Simplicity is not about dumbing down ideas or using sound bites;
rather it is about prioritizing. Because people have limited attention, it is often
best to focus on a single, core message. Adding secondary points and unneces-
sary detail can detract from the impact of a message. Heath and Heath suggest
that the ideal of simplicity is a message that gets at the core of the idea you are
trying to convey and includes nothing else. For example, Bill Clinton’s campaign
advisor, James Carville, developed the communication strategy captured by the
phrase: “It’s the economy, stupid.” This meant that Clinton’s campaign speeches
would focus almost exclusively on the economy. As Carville put it, “There has to
be message triage. If you say three things, you don’t say anything.”
Unexpectedness. Hearing things that are unexpected often leads to curiosity,
a state that increases our interest in and the attention we pay to a topic or ques-
tion.3 Heath and Heath argue that this often leads to extra thinking, which helps
make ideas stick in our memories. Describing your idea in terms that are coun-
terintuitive or surprising can be a good way to make a message stick.
Heath and Heath describe the example of Nordstrom, a department store
known for its excellent customer service. In order to maintain this reputation,
managers must impress on their employees how much they value good service.
Rather than just giving speeches about “the importance of customer service,”
managers give examples of past instances of outstanding service by Nordstrom
employees. Because they were surprising, examples such as ironing a shirt for a
customer who needed it for a meeting that afternoon and cheerfully gift-wrapping
clothing a customer bought at Macy’s, got the message across effectively.
Concreteness. There is a tendency, especially when we have a lot of knowledge
about a topic, to communicate ideas in abstract terms. But abstract ideas are dif-
ficult for nonexperts to understand and are easily misconstrued even by experts.
The Nordstrom example described above is also an example of the power of con-
creteness. If managers just told employees that they should “go above and beyond
what it expected,” that would leave a lot of room for interpretation and probably
wouldn’t leave a lasting impression. Providing concrete examples of the good
service makes clear exactly what is expected.
Credibility. The traditional way to establish credibility is to have credentials—for
example, advanced training or a degree. But people can sometimes establish cred-
ibility by alternative means. Heath and Heath refer to “testable credentials,” in
which listeners are asked to evaluate the credibility of a contention for themselves.
During a presidential debate with Jimmy Carter, Ronald Regan asked the rhetorical
2. Chip Heath and Dan Heath, Made to Stick: Why Some Ideas Survive and
Others Die (New York: Random House, 2007).
3. George Loewenstein, The Psychology of Curiosity: A Review and Reinterpretation, 116
Psychological Bulletin 75–98 (1994).
influencing behavior through cognition 573
question: “Are you better off now than you were four years ago?” Rather than argue
directly that the economy had worsened during Carter’s tenure, he asked the audi-
ence to test the proposition for themselves.
Emotionality. Research on charitable giving has shown that people donate
more money in response to appeals that focus on a single, identifiable victim
than they do when appeals cite statistics reflecting mass-scale suffering.4
Researchers who study this phenomenon proffer the following explanation for
it: learning about a single victim evokes emotion, whereas learning about statis-
tics shifts people into a more analytical frame of mind that suppresses emotion.
Emotion is the motivational engine that drives people to respond to messages
involving appeals for charitable donations. Focusing on analytical themes under-
mines messages intended to promote charitable giving.5
In his book, The Political Brain, psychologist Drew Westen makes a similar
contention about political appeals. He argues that political messages based on
rational, emotion-free arguments fail to have much effect on voters. Because
people’s decisions are often guided primarily by emotion and rationalized after
the fact, attempts to persuade are more likely to resonate with voters if they focus
on deeply held values and principles that tend to evoke emotion.6
Stories. Stories help ideas stick. They give coherence to a message and,
because we find them entertaining and engrossing, they capture our attention
better than a simple list of facts. Also, unlike more direct persuasive messages,
stories tend not to put the audience in an evaluative frame of mind. Rather,
people tend to empathize with the story’s protagonist and to be less critical of the
underlying message.
A study by Nancy Pennington and Reid Hastie found that when participants
judged evidence in a hypothetical criminal trial, they were more likely to render
a verdict for the side that presented evidence in an order that made it easy to
construct a story of the events in question.7 Recall our discussion, in Section 8.3,
of the power of narrative. While arguments can establish formal or abstract
truths, stories are better at convincing people of the verisimilitude of a point
because stories make a point seem representative of reality. Recall that people use
the representativeness heuristic to make judgments about the likelihood that
4. Deborah A. Small, George Loewenstein, and Paul Slovic, Sympathy and Callousness:
Affect and Deliberations in Donation Decisions, 102 Organizational Behavior and
Human Decision Processes 143–53 (2007).
5. George Loewenstein and Deborah A. Small, The Scarecrow and the Tin Man: The
Vicissitudes of Human Sympathy and Caring, 11 Review of General Psychology 112
(2007).
6. Drew Westen, The Emotional Brain: The Role of Emotion in Deciding the
Fate of the Nation (Public Affairs, New York, 2007).
7. Nancy Pennington and Reid Hastie, Explanation-Based Decision Making: Effects of
Memory Structure on Judgment;14 Journal of Experimental Psychology: Learning,
Memory, & Cognition 521–33 (1988).
574 problem solving, decision making, and professional judgment
Recall the now familiar S-shaped curve of Prospect Theory of Figure 14.1, and
recall that losses from a reference point loom larger than gains. Because the
reference point is essentially subjective, even arbitrary, a strategic communicator
may try to get her audience to adopt a reference point that will influence attitudes
in the direction she wants. Consider these experiments.
George Quattrone and Amos Tversky conducted an experiment asked partici-
pants to imagine that a hypothetical country had dedicated $100 million to reduc-
ing crime by unemployed immigrant youths. They were asked to allocate the
$100 million between two groups of equal size, Alphans and Betans.
Half the participants were told that, by the age of twenty-five, 3.7 percent of
Alphans and 1.2 percent of Betans had criminal records, while the other half
were told that, by the age of twenty-five, 96.3 percent of Alphans and 98.8 percent
of Betans had no criminal record. The participants had to determine whether to
give a little bit more to the Alphans, or a lot more to the Alphans—the group with
the higher rate of criminality. Objectively, the two groups were given identical
information, but the information in the second condition was framed in terms
of a lack of criminality instead of criminality.
8. Jerome Bruner, Making Stories: Law, Literature, Life (New York: Farrar,
Straus, & Giroux, 2002).
9. Shelly Chaiken, The Heuristic Model of Persuasion, in Social Influence: The
Ontario Symposium Vol. 5, 3–39 (Mark P. Zanna, James M. Olson, and C. Peter Herman
eds., Hillsdale, NJ, 1987). Lawrence Erlbaum et al., Central And Peripheral Routes to
Persuasion: An Individual Difference Perspective, 51 Journal of Personality and Social
Psychology 1032–43 (1986).
influencing behavior through cognition 575
In the “criminality” frame (3.7 percent Alphans versus 1.2 percent Betans)
participants allocated more of the money to the Alphans than in the noncriminality
framing (96.3 percent versus 98.8 percent). Quattrone and Tversky attributed
this to what they call the “ratio-difference principle.” In the criminality frame,
Alphans are perceived as three times more criminal as Betans; but they are
seen as only slightly less noncriminal than the Betans under the noncriminality
frame. This is a result of diminishing sensitivity, shown in the S-shaped prospect
theory curve as one moves away from the reference point. According to the ratio-
difference principle, “the impact of any fixed positive difference between two
amounts increases with their ratio.”10 In other words, the difference between 3.7
percent and 1.2 percent seems bigger than a difference between 98.8 percent and
96.3 percent because the ratio of the first pair is much bigger than that of the
second pair. Our sensitivity to differences in objective numbers like crime rates
is an approximate function of the ratio associated with that difference.
In another experiment, Quattrone and Tversky found that whether an eco-
nomic indicator was described as the unemployment rate or the employment rate
had a significant impact on how concerned people seemed about a 5 percent
worsening of it. People weighed the difference between 5 percent and 10 percent
unemployment much more heavily than the difference between 95 percent and
90 percent employment. Indeed, descriptive statistics of the sort used in these
experiments are common in discussions about public policy, and this research
suggests that subtle differences in framing can affect how much people approve
(or disapprove) of a policy, irrespective of its actual impact.
Heath and Heath describe how messages can be shaped at a macro level to make
them more persuasive. Recent research by other scholars suggests that subtle
manipulations of language can affect the meaning that recipients take from a
message.
10. George Quattrone and Amos Tversky, Contrasting Rational and Psychological
Analyses of Political Choice, 82 APSR 719–36 (1988).
11. Caitlin M. Fausey, Neil, Snider, Lera Boroditsky (in prep). Speaking of Accidents:
“He did it” invites more punishment than “It happened.”
576 problem solving, decision making, and professional judgment
Both are common and sound natural in English, and they are generally consid-
ered to be equivalent ways to say the same thing. Consider, for example, the fol-
lowing story:
Mrs. Smith and her friends were finishing a lovely dinner at their favorite
restaurant. After they settled the bill, they decided to head to a nearby café
for coffee and dessert. Mrs. Smith followed her friends and as she stood up,
she flopped her napkin on the centerpiece candle. She had ignited the napkin!
As Mrs. Smith reached to grab the napkin, she toppled the candle and ignited
the whole tablecloth, too! As she jumped back, she overturned the table
and ignited the carpet as well. Hearing her desperate cries, the restaurant
staff hurried over and heroically managed to put the fire out before anyone
got hurt.
Contrast this story:
Mrs. Smith and her friends were finishing a lovely dinner at their favorite
restaurant. After they settled the bill, they decided to head to a nearby café
for coffee and dessert. Mrs. Smith followed her friends and as she stood up,
her napkin flopped on the centerpiece candle. The napkin had ignited! As
Mrs. Smith reached to grab the napkin, the candle toppled and the whole table-
cloth ignited, too! As she jumped back, the table overturned and the carpet ignited
as well. Hearing her desperate cries, the restaurant staff hurried over and
heroically managed to put the fire out before anyone got hurt.
Fausey and Boroditsky found that people who read the former, agentive ver-
sion of the story assigned significantly more blame for the fire to Mrs. Smith.
When asked how much of the $1500 in total damages from the fire Mrs. Smith
should have to pay, participants who read the agentive version, on average, said
she should have to pay $935, while those who read the nonagentive version said,
on average, $689 (i.e., less than half the total amount).
The implications of this research for the courtroom are apparent. Outside the
courtroom, policy makers and other actors sometimes try to blunt criticism by
using the passive, “mistakes were made.” In an interview on CBS’s 60 Minutes,
Marine Staff Sgt. Frank Wuterich, who was charged with eighteen counts of
murder after the squad he was leading killed twenty-four civilians in apparent
response to the death of one of their fellow marines, said “I am completely sorry
that that happened.” Note that the research described above deals with cases in
which the passive voice is used by a third party. A perpetrator’s use of the
nonagentive form may trigger a cynical response both in courtrooms and in the
court of public opinion.
important effect on the answers you receive.12 They distinguish between “verbs
of action” (such as to help, to push, to cheat, etc.) and “verbs of state” (such as to
respect, to dislike, to love). Semin and De Poot’s research shows that questions
about action verbs tend to yield answers that focus on the sentence subject, while
the answers to questions about state verbs tend to focus on the sentence object.
So, if I ask you “Why do you think John helped David,” you will tend to answer
by telling me something about John that explains his helping behavior. On other
hand, if I ask you “Why do you think John likes David,” your answers will likely
refer to something about David that makes him likeable.
At first this might not seem to be the kind of thing that is subject to easy
manipulation: action verbs and state verbs tend not to be interchangeable. It is
hard to imagine a defense lawyer asking her client to explain to the jury why he
“disliked” the person he assaulted instead of why he “attacked” the victim. As
Semin and De Poot point out, though, once we know the effect of these verbs,
other elements of a sentence are susceptible to customization. For example, we
sometimes have a choice about whom to use as the subject and whom to use as
the object of a sentence. Imagine, for example, a rape case in which it is undis-
puted that the victim and the defendant danced together before the alleged crime.
Knowing that to dance is an action verb and that answers to action verbs tend to
focus causal explanations on the object of a sentence, a defense attorney might
prefer to ask the victim whether she danced with the defendant rather than
whether the defendant danced with her.
12. Gün R. Semin and Christianne J. De Poot, The Question-Answer Paradigm: You
Might Regret Not Noticing How a Question Is Worded, 54 Journal of Personality and
Social Psychology 558–68 (1997).
13. Gregory M. Walton and Mahzarin R. Banaji, Being What You Say: The Effect of
Essentialist Linguistic Labels on Preferences, 22 Social Cognition 193–213 (2004). This
paper was inspired by the following paper examining a similar difference in children:
Susan A. Gelman and Gail D. Heyman, Carrot-Eaters and Creature-Believers: The Effects of
Lexicalization on Children’s Inferences About Social Categories, 10 Psychological Science
489–93 (1999).
578 problem solving, decision making, and professional judgment
Banaji and Walton then wondered if the same thing would happen when
people were being asked to think about their own behavior. Under the guise that
they were conducting a handwriting study, they had participants write out a
behavioral tendency or preference of their own, using either the noun of the verb
form. If a person said he liked Coke, he was either asked to write “I am a Coke-
drinker” or “I drink a lot of Coke” three times in his “natural handwriting.” They
found similar effects. People who had written about their own behavioral prefer-
ences using the noun form tended to say that those preferences were stronger
than if they had used the verb form.
In subsequent research, Christopher Bryan, Gregory Walton, and Carol Dweck
investigated whether a similar manipulation might affect not only people’s atti-
tudes about products but important behavioral decisions such as whether or not
to vote in a presidential election.14 The day before the 2008 presidential election,
they asked people who were eligible to vote but not yet registered to complete a
questionnaire about their “attitudes about voting and the electoral process.” In
that questionnaire, people were either asked questions like “How important is it
to you to vote in the upcoming election” or “How important is it to you to be a
voter in the upcoming election.” After the election, the researchers used official
records to determine whether participants had turned up at the polls.
It turned out that participants who had completed the questionnaire about
their attitudes about “being a voter” were more likely to have voted than partici-
pants who had answered questions about “voting.” The fact that such a seem-
ingly minuscule manipulation can have an effect on the decision to vote in a
presidential election—behavior that one would expect to be determined by
motives such as a sense of civic duty or interest in the outcome—suggests that
this subtle difference may have significant implications.
Where do these nouns get their power? The answer is not yet completely
clear, but researchers agree that the use of a noun to refer to a preference or
behavior invokes the idea of essentialism. This is the folk notion that a person
(or object) has some fundamental essence that makes her what she is and deter-
mines her attitudes, thoughts, and behaviors.15 It may be that, when a character-
istic is described in the noun form, the implication is that this is a central aspect
of that person’s self. It is conceivable that this affects even our perceptions of our
own characteristics and behavior.
14. Christopher J. Bryan, Mark R. Lepper, and Carol S. Dweck, “Voting versus Being
a Voter: The Effect of a Subtle Linguistic Manipulation on Voting Behavior” (unpublished
data, Stanford University 2005).
15. Dale Miller and Deborah Prentice, Some Consequences of a Belief in Group Essence:
The Category Divide Hypothesis, in Deborah Prentice and Dale Miller eds., Cultural
Divides: Understanding And Overcoming Group Conflict 213–238 (New York:
Russell Sage Foundation, 1999).
influencing behavior through cognition 579
16. George Lakoff and Mark Johnson, Metaphors We Live By (Chicago: University
of Chicago Press, 1980).
580 problem solving, decision making, and professional judgment
17. George Lakoff, Don’t Think of an Elephant! Know Your Values and Frame
the Debate (White River Junction, VT: Chelsea Green, 2004).
18. George Lakoff, Moral Politics: How Liberals and Conservatives Think
(Chicago: University of Chicago Press, 1996).
influencing behavior through cognition 581
both models to some extent, it may be possible to influence people’s policy opin-
ions by invoking the relevant family model in political appeals.
Even when people understand and agree with communications that call for
action, they do not necessarily take the next step and act on them. People may
have an inclination to act, but if their intended behavior has no available outlet,
then they may be less likely to do so. For example, researchers found that the
best predictor of whether people would show up at a health clinic was not their
attitudes about health and medicine, but how far they were from the clinic.19
Channel factors are simply ways of structuring a situation to increase the likeli-
hood that the recipient of a communication will act on it.
Consider this example:20 Robert Leventhal and his colleagues were studying
the effects of fear on the likelihood that people would get a tetanus shot. They
found that when conveying the risk of tetanus, a fear-inducing appeal made
people change their attitudes in favor of getting the shot significantly more than
a neutral informational appeal. Nevertheless, even with this attitude, only 3 per-
cent of participants followed up and got inoculated against tetanus. Strikingly, the
simple act of giving participants a map and instructing them to select a particular
time when they would go to get inoculated raised this percentage to 28 percent.
By the same token, sales of war bonds improved during World War II when
the government arranged for the bonds to be sold at workplaces rather than at
banks and post offices.21 In sum, people may have every intention to act in a
certain way, but if they don’t have a “channel” for their behavior, their intention
may never be carried out.
18.5 priming
19. Berenice E. Van Dort and Rudolf H. Moos, Distance and the Utilization of a Student
Health Center, 24 Journal of American College Health Association 159–62 (1976).
20. Howard Leventhal, Robert Singer, and Susan Jones, Effects of Fear and Specificity of
Recommendation upon Attitudes and Behavior, 2 Journal of Personality and Social
Psychology 20–29 (1965).
21. Dorwin Cartwright, Some Principles of Mass Persuasion: Selected Findings of Research
on the Sale of U.S. War Bonds, 2 Human Relations 253 (1949).
582 problem solving, decision making, and professional judgment
which people are therefore more susceptible to suggestion about what the rele-
vant norms are). In Section 11.3, we described how the priming of particular
schemas can influence people’s perceptions of and behavior toward members of
stereotyped groups. Here we broaden the discussion to other forms of judgment
and decision making.
Recall the “Ultimatum Game,” described in Section 12.1.2, in which one par-
ticipant is asked to propose a division of a fixed sum of money that a second
participant can either accept or reject.22 Would priming participants with cues
that invoked a business environment lead them to behave in a more competitive,
cut-throat manner, consistent with a business stereotype? Researchers asked
some participants to play the ultimatum game seated in front of a large desk on
which a briefcase, an executive-style silver pen, and a black-leather portfolio had
been placed. For other participants, a backpack replaced the briefcase, a card-
board box replaced the executive portfolio, and the participants were given a
standard wooden pencil instead of an executive pen. It turned out that partici-
pants in the business-objects condition proposed significantly less generous
divisions of the money than participants in the second condition.
The result of the research on the ultimatum game is important because it
suggests that random and seemingly irrelevant features of our environments
can serve as primes and have a significant influence on our behavior. Perhaps
the most striking demonstration of the significance of implicit priming effects is
a study by Jonah Berger and his colleagues that suggests that the polling place in
which people vote (e.g., church, school, or firehouse) may influence how they
vote. Berger et al. analyzed data from Arizona’s 2000 general election and found
that people who cast their votes in schoolhouses were more likely than others to
vote to raise taxes to support education.23 This study suggests that subtle environ-
mental cues can influence decisions on consequential issues even in complex
real-world environments.
In Section 1.5.2 we introduced the concept of social scripts that define appro-
priate behavior for ourselves and others, including public officials. People often
have recourse to alternative scripts, and priming can affect which alternative
someone applies to a given situation. For example, Thomas Gilovich presented
participants with a hypothetical foreign conflict and asked whether the United
22. Aaron C. Kay, S. Christian Wheeler, John A. Bargh, and Lee Ross, Material Priming:
The Influence of Mundane Physical Objects on Situational Construal and Competitive
Behavioral Choice, 95 Organizational Behavior and Human Decision Processes
83–96 (2004).
23. Jonah Berger, Marc Meredith, and S. Christian Wheeler, “Can Where People Vote
Influence How They Vote? The Influence of Polling Location Type on Voting Behavior”
(2006) Stanford Graduate School of Business Research Paper, No. 1926. The researchers
controlled for the voters’ political views, demographics, and whether or not they lived near
schools.
influencing behavior through cognition 583
24. Thomas Gilovich, Seeing the Past in the Present: The Effects of Associations to Familiar
Events on Judgments and Decisions, 40 Journal of Personality and Social Psychology
797–808 (1981).
25. Christopher J. Bryan, Carol S. Dweck, Lee Ross, Aaron C. Kay, and Natalia
Mislavsky, “Ideological Mindsets: A Demonstration of Political Malleability” (2006) (man-
uscript submitted for publication).
26. James R. Kluegel and Eliot R. Smith, Beliefs About Inequality: Americans’
Views of What is and What Ought to Be (New York: Aldine deGruyter, 1986); Irwin
Katz and R. Glen Hass, Racial Ambivalence and American Value Conflict: Correlation and
Priming Studies of Dual Cognitive Structures, 55 J. Pers. Soc. Psychol. 893–905 (1988);
Robert E. Lane, The Fear of Equality, 53 Am. Polit. Sci. Rev. 35–51 (1959).
584 problem solving, decision making, and professional judgment
The question we just asked implies that policy makers can use their knowledge
about the techniques of social and cognitive influence to help citizens act in their
own best interests or in the interests of society. Of course, policy makers’ use of
such techniques raises core normative questions about the obligations of liberal
democratic governments to their citizens, who have heterogeneous values and
goals and diverse views about how to achieve them. But although liberal democ-
racy implies a presumption favoring autonomy, most political theorists hold that
citizens’ autonomy may sometimes be limited for the benefit of the greater soci-
ety and many hold that their choices may be (paternalistically) limited for the
benefit of individuals themselves.
29. Kip Viscusi and Welsey A. Magat, Learning About Risk: Consumer and Worker
Responses to Hazard Information (Cambridge: Harvard University Press, 1987).
30. Id. at 128.
31. W. Kip Viscusi, Alarmist Decisions with Divergent Risk Information, 107 The
Economic Journal 1657 (1997).
586 problem solving, decision making, and professional judgment
If people neglect probability, they may fix, or fixate, on the bad outcome, in a way
that will cause anxiety and distress but without altering behavior or even improv-
ing understanding. Sunstein suggests that it might be possible to tell people not
only about the risk but also about the meaning of the probability information—
for example, by comparing a risk to others encountered in ordinary life.” But, he
asks, “if the risk is low, and of the sort that usually do not trouble sensible human
beings, is it really important to force disclosure of facts that will predictably cause
high levels of alarm?”32
The problem is that the very discussion of a low-probability risk leads people
to imagine or visualize the worst case—even if the discussion consists mostly of
trustworthy assurances that the likelihood of harm really is infinitesimal.
Emphasizing the low probability that the risk will eventuate is a lame strategy
against affect-laded imagery. Of course, there’s another potential problem as
well: “People may attempt to deal with their fear by refusing to think about the
risk at all.”33
Uri Gneezy and Aldo Rustichini examined the effect of introducing a small
fine to discourage a nuisance behavior.35 They conducted a study in ten day-care
centers in Haifa, Israel. The day-care centers all operate from seven-thirty in the
morning until four in the afternoon, but teachers regularly have to stay past four
because some parents were late picking up their children. The researchers
arranged for some of the day-care centers to introduce a small fine (equivalent to
approximately one eighth the amount of a parking ticket) each time a parent was
more than ten minutes late. Surprisingly, the effect of the fine was to increase the
number of late-arriving parents in those day-care centers. After several weeks of
fines and increased rates of lateness, those centers that had introduced fines
eliminated them, but to no avail. The rate of lateness in those centers remained
at the same high level they were at under the fine system.
The researchers hypothesize that before the implementation of the fines, par-
ents had been operating with the understanding that teachers who stayed late to
watch over their children were doing so out of generosity, and therefore that the
parents should avoid imposing on them. The introduction of a fine, however,
was interpreted as a price for the service of staying late. Once this occurred, par-
ents thought of the service as a commodity that they could buy as often as was
convenient. The researchers further theorize that rates of lateness stayed high
when the fines were removed because the service continued to be seen as a com-
modity—only now it was a commodity with a price of zero.
Other research shows that even large incentives can sometimes be counter-
productive.36 Researchers went door-to-door in Switzerland asking residents if
they would be willing to accept a nuclear waste dump in their town. Although
residents were obviously apprehensive about the idea, approximately half of the
people polled said they would accept the dump. They presumably did this out of
a sense of civic responsibility; the dumps have to go somewhere after all. Other
residents were asked the same question but told that if they accepted the dump
in their neighborhood, they would be compensated with the equivalent of
approximately six weeks’ pay each year. The rate of acceptance dropped to half
the rate when no compensation was offered. Residents may have been willing to
accept a certain level of risk out of a sense of civic responsibility, but being paid
to risk their health was a losing proposition.
The common theme in these examples is that the introduction of economic
incentives into transactions that are governed by social norms changes the way
those transactions are understood by the people involved. In short, relying on
economic incentives to influence behavior that is governed by nonmaterial
motives sometimes can backfire.
35. Uri Gneezy and Aldo Rustichini, A Fine Is a Price, 24 Journal of Legal Studies
1–17 (2000).
36. Bruno S. Frey and Felix Oberholzer-Gee, The Cost of Price Incentives: An Empirical
Analysis of Motivation Crowding Out, 87 American Economic Review 746–55 (1997).
588 problem solving, decision making, and professional judgment
37. Much of this section is based on Richard Thaler and Cass Sunstein, Nudge:
Improving Decisions About Health, Wealth, and Happiness (New Haven: Yale
University Press, 2008).
38. Katherine L. Milkman et al., How Can Decision Making Be Improved?, 4 Perspectives
on Psychol. Sci. 379 (2009).
39. For an earlier version of this argument, see Richard H. Thaler and Cass R. Sunstein,
Libertarian Paternalism, 93 Am. Econ. Rev. 175 (2003); Christine Jolls and Cass Sunstein,
Debiasing Through Law (2005); Christine Jolls, Cass Sunstein, and Richard Thaler, A
Behavioral Approach to Law and Economics, 50 Stanford Law Review 1471 (1998).
influencing behavior through cognition 589
40. Sara M. Banks et al., The Effects of Message Framing on Mammography Utilization,
14 Health Psychology 178–84 (1995).
590 problem solving, decision making, and professional judgment
appeals that emphasize the potential health benefits of early detection are likely
to be less effective. Because people tend to be risk-averse in the domain of gains,
such frames are likely to make people want to avoid the risk of scary news associ-
ated with detection.
In contrast, prevention behaviors such as wearing sunscreen or using con-
doms are perceived as safe. In fact, not performing such behaviors is generally
seen as the risky option. Because people tend to be risk-seeking in the domain of
losses, loss-framed appeals, like warning people of the potential costs of unpro-
tected sun exposure, might actually make people more likely to take the risk of
not wearing sunscreen. Instead, it is preferable to use gain-framed appeals, like
emphasizing the benefits of wearing sunscreen, to encourage this type of non-
risky behavior.41
Similar to prevention behaviors, recuperative behaviors (i.e., obtaining treat-
ment for an existing condition) are usually seen as nonrisky. Therefore, as with
prevention behavior, Rothman and Salovey find that loss-framed appeals—such
as those that emphasize the potential costs of failing to obtain treatment—tend to
be counterproductive because they are more likely encourage risk-seeking atti-
tudes. Gain-framed appeals that focus on the health benefits of treatment are
preferable.
18.6.3.b Providing Feedback and “Mapping” Choices to Experienced
Utility Simply making people more aware of the relationship between an action
and its consequences can shape behavior. For example, the environmental costs
of people’s actions are rarely salient. But the display in the Toyota Prius, which
shows drivers their momentary gas mileage, may affect their acceleration and
speed. Thaler and Sunstein describe a device called the Ambient Orb, which
changes color when home energy use rises above a certain level; people who
began using this device decreased their energy usage by 40 percent.
The information provided consumers as a basis for evaluating and choosing
among options is often obscure, with different options often being described
with different metrics. The problem will be familiar to readers who have shopped
for cell phone providers or cable TV packages. Thaler and Sunstein suggest that
a good choice architecture makes information about relevant dimensions of each
option salient.
Consider the advantage of displays that show supermarket shoppers the cost
per unit of different brands of the same item, or the advantage of giving purchas-
ers of home mortgages a comparison of lenders’ fee structures and interest rates.
Thaler and Sunstein suggest that consumers would make better choices if credit
card lenders were required to report a summary of fees and interest rates on a
41. Alexander J. Rothman, Peter Salovey, Carol Antone, Kelli Keough, and Christina
D. Martin, The Influence of Message Framing on Intentions to Perform Health Behaviors, 29
Journal of Experimental Social Psychology 408–33 (1993).
influencing behavior through cognition 591
42. Going from 34 to 50 MPG saves 94 gallons; but going from 18 to 28 MPG saves 198
gallons.
43. 20 Science 1593–94 (June 2008). DOI: 10.1126/science.1154983.
592 problem solving, decision making, and professional judgment
in the choice set. Thaler and Sunstein suggest the middle ground of simplifying
and structuring choice sets.
Consider the challenge of sorting through all the options for Medicare Plan D
prescription drug plans. Each plan contains information about all the different
drugs it covers at what rate and with what copayments. The challenge of sorting
through all the plans to find the right coverage would confound even a well-
informed person. But a computer program could recommend a smaller set of
options based on the prescriptions a particular individual uses most.
Large choice sets can be made easier to navigate by grouping options. For
example, some investment funds, rather than provide customers with a list of
funds in alphabetical order, group them by risk level (low, moderate, high) and
portfolio composition (primarily stocks, primarily bonds). Without limiting the
number of options available, this approach can limit the number of options that
people actually consider.
People could also be given the option of delegating their decision to an expert,
for example in choosing among 401(k) retirement plans. Delegation to an expert
could even be made the default choice.
18.6.3.e Making Incentives Salient For incentives to be effective in shaping
people’s behavior, they must be noticeable and salient. But the effects of deci-
sions often are delayed or difficult to imagine. Consider a family choosing
whether to purchase a car or rely on alternative means of transportation. The
costs of owning and operating a car may far exceed the costs of the alternatives.
But the financial shock of purchasing a car occurs only once, while one feels the
pain of losing money every time one pays a bus fare, or rents a car, or watches
the meter on a taxi. Recall from our discussion of prospect theory that losses are
more upsetting when segregated rather than integrated. (See Section 14.4) By
the same token, anticipating myriad cab fares and rental car fees may feel more
salient than the single loss of the price of purchasing a car. Perhaps making the
installment payments and operating costs of the car equally salient would help
level the playing field.
Prospect theory also suggests that several small incentives are likely to
affect behavior more than would one large incentive. Receiving a year-long
$100-per-month subsidy after buying a hybrid car might increase hybrid
purchases more than providing a single payment of $1200—and would almost
surely lead to more of an increase than discounting the car purchase by $1200.
Hefty fines for speeding that are only occasionally enforced are probably
less effective at reducing speeding than moderate fines that are frequently
enforced.
18.6.3.f Criticisms of Libertarian Paternalism The effectiveness of choice
architecture is an empirical question that is highly dependent on the contexts in
which its techniques are applied. We expect that, over time, social scientists will
develop useful generalizations about what works, what doesn’t, and why. But as
policy makers continue to explore the techniques of choice architecture, it is
influencing behavior through cognition 593
worth noting some recurring concerns about its underlying political premise,
libertarian paternalism.
• Libertarian paternalism is an insidious use of government power because, by
not regulating conduct directly, it does not mobilize the same sort of focused
opposition as hard regulatory policies. But if there is no “neutral” way to
structure a choice, policy makers might as well choose frames and defaults
that are most likely to achieve citizens’ interests or, where there is no appar-
ent conflict, the public interest. In any event, a democratic government
should be transparent to its citizens about its use of choice architecture.
• Libertarian paternalism encourages citizens’ passivity and manipulates
their choices rather than seeking to educate them to be better decision
makers. But the literature on debiasing is not optimistic about the possi-
bilities for countering biases, and the pervasiveness of bounded rationality
means that even very thoughtful individuals will make many decisions by
default.
• Choice architects are subject to their own biases and to corruption, partial-
ity, and narrow-mindedness. Choice architecture offers the same opportu-
nities for influence by special interests as any regulation. But the deliberative
decision making that underlies good policies provide safeguards against
biases. (Though, to refer back to the first criticism, choice architecture may
embody tacit assumptions that would be more readily exposed in the con-
test over non-libertarian regulations.44)
• Policy makers’ choice of defaults, or even about what information to pro-
vide individuals, may assume an unjustified homogeneity of interests.
Maybe there are individuals who would genuinely prefer to have more cash
today than to save for retirement. When Thaler and Sunstein argue that
governments should inform parents’ choices of schools, they focus on
information about educational achievement, but some parents may be
more concerned about social or religious values. But as long as it isn’t oner-
ous to override the defaults, why shouldn’t policy makers choose defaults
that they believe to be in individuals’ or society’s best interests? And if gov-
ernments do not provide all the information people need to make a deci-
sion based on their own interests, they have access other sources of
information.
• Although choice architects endeavor to protect citizens’ liberty in pursuing
their diverse values and goals, the fact that goals are often constructed in
the very process of decision making undercuts its implicit distinction
between people’s ends and their means for achieving them. Mark Kelman
argues that efforts to assist citizens in pursuing their own values tend to
44. For example, there is an asymmetry between the defaults involving organ donation,
with some individuals believing that the practice violates religious imperatives.
594 problem solving, decision making, and professional judgment
embody policy makers’ implicit theory of the “good society.”45 But there are
plenty of instances where people’s bounded rationality and biases pretty
clearly result in bad decisions in terms of interests that were not con-
structed on the spot.
There are no simple solutions to these problems of choice architecture. But
Thaler and Sunstein offer an alternative to defaults that may obviate some of
them and improve people’s decision-making processes: mandated choice. Suppose
that you can’t obtain a driver’s license without specifying whether or not you
wish to be an organ donor if you are in a fatal accident. Suppose that an employer
cannot hire you until you have specified what, if any, retirement savings plan you
want to enroll in. Mandated choice doesn’t work in all circumstances, and it
doesn’t obviate the inevitability of framing—consider Carolyn’s cafeteria. But,
when it is feasible, it nudges citizens to deliberate about important choices rather
than nudging the choice in a particular direction—perhaps the ideal sort of
libertarian paternalism in a deliberative democracy.
45. Mark G. Kelman, The Heuristics Debate: Its Nature and Its Implications
for Law and Policy (Oxford University Press, forthcoming 2010); Mark Kelman, Hedonic
Psychology and the Ambiguities of “Welfare,” 33 Phil. & Pub. Aff. 391, 395–97 (2005).
19. group decision making and the effects
of accountability on decision quality
Many if not most important decisions by policy makers and lawyers are made in
group settings. For example, as Luis Trujillo and Christine Lamm deal with the
problem at Terra Nueva, they will involve teams of colleagues and experts and
work with other groups of stakeholders. Sometimes they will just seek advice;
sometimes the group will have the authority to make decisions. This is charac-
teristic of decision making in the private and nonprofit sectors as well as govern-
ments, and in private as well as public institutions. In situations like these, social
dynamics within the decision-making group can exert a powerful effect on deci-
sion quality.
This chapter explores a variety of questions relating to the performance of
small decision-making groups. Do groups outperform their individual members
and, if so, when? What are the most common sources of dysfunction in decision-
making groups, and how can their effects be minimized? Are there ways in
which a group can be constituted, or its work structured, to set the stage for opti-
mal group decision making? We conclude by considering how a decision mak-
er’s accountability to others—for individual as well as group decisions—can
affect decisions for better or worse.
1. Harold J. Leavitt, Suppose We Took Groups Seriously, in Man and Work in Society
67–77 (Eugene L. Cass and Frederick G. Zimmer eds., New York: Van Nostrand Reinhold,
1975).
596 problem solving, decision making, and professional judgment
In any event, Leavitt observed, small groups will form within organizations as
a result of natural patterns of human social interaction. Given the inevitability of
groups, it is best to manage them in ways that will maximize their effectiveness.
Leavitt’s faith in the power of groups spread rapidly and became the received
wisdom in popular and academic thinking about business and government
administration.
At about the same time, the social psychologist Irving L. Janis published a
book on the perils of group decision making. In an expanded 1982 edition enti-
tled Groupthink,2 Janis sought to explain government decision-making disasters,
such as the failure to anticipate the Japanese attack on Pearl Harbor, the deci-
sions to escalate the war in Vietnam, the Bay of Pigs invasion, and the Watergate
cover-up.
The basic idea of Janis’s “groupthink” hypothesis was that highly cohesive
decision-making groups have a strong tendency to suppress their members’
independent thinking and to converge too quickly on poor-quality decisions. The
drive to avoid intragroup conflict prevents rigorous evaluation of competing
alternatives. Divergent perspectives and theory-disconfirming information are
suppressed in favor of supportive opinions and theory-confirming views.
Buckling under the overt and covert social pressure, even those group members
who harbored serious private reservations about the group’s preferred choice do
not express them and acquiesce in the decision. Janis also described a few suc-
cessful policy decisions, including President John F. Kennedy’s response to the
Cuban Missile Crisis and the development of the Marshall Plan after World War
II, in which the decision-making process did not appear to be infected with these
dynamics.
Three decades of research tell a more nuanced and complex story than either
Leavitt’s rosy or Janis’s pessimistic account. Groups can add value to a decision-
making process, but they can also run it off the rails. Small groups often make
better decisions than their average member would make alone, but they seldom
outperform their best members.3 Small groups fall prey to virtually all of the
cognitive biases that plague individual decision makers and sometimes actually
amplify the effects of those biases. Moreover, deliberation, which commentators
like Leavitt credit with improving the effectiveness of group decision making,
can actually introduce additional sources of error into the process.
Most of this chapter will focus on decision making by groups whose mem-
bers interact with each other. As a prelude, however, it is useful to contrast the
4. James Surowiecki, The Wisdom of Crowds: Why the Many Are Smarter than
the Few and How Collective Wisdom Shapes Business, Economies, Societies, and
Nations (New York: Doubleday, 2004).
5. Id. xii.
598 problem solving, decision making, and professional judgment
dressed weight. However, the average of the estimates, 1,197 pounds, turned out
to be only one pound less than the weight of the ox when slaughtered and
dressed.
The power of statistical averaging illustrated by Galton’s experience derives
from a law known as the Condorcet Jury Theorem.6 Here’s how the theorem
works: Suppose that a number of people are independently answering a yes or no
question, and that the probability that each individual will answer the question
correctly is at least somewhat greater than 50 percent. Then the probability that a
majority of group members will answer the question correctly increases as the
size of the group increases. If the answers are not binary, but are spread over a
continuous range, Condorcet’s theorem predicts that the average answer will
more closely approximate the correct answer as the size of the group increases.
(If the probability that each individual member will decide correctly is less than 50
percent, the probability that a majority will answer correctly approaches zero.)
Over half a century ago, Friedrich Hayek observed that aggregated knowledge
is crucial for providing information to the market.7 Yet the mechanisms for sys-
tematically accumulating and organizing information from dispersed individu-
als (other than through revealed preferences exhibited by their participation in
the market) were largely unexplored until the late 1990s, when Caltech and
Hewlett-Packard Laboratories undertook an innovative experiment to imple-
ment an Information Aggregation Mechanism within the corporation.
Kay-Yut Chen, a scientist at Hewlett-Packard Laboratories (HP) in Palo Alto,
California, and Charles R. Plott, a professor of economics and political science at
Caltech, experimented with internal “securities markets” to improve product
sales forecasting.8 They selected a small number of employees from different
parts of HP’s business operations and different geographical regions. Some
employees had information about clients, others about pricing strategies, others
about competitors’ strategies and products. Participants were given an initial
portfolio of securities (created for the experiment) with each security associated
with one of ten different levels of HP sales, and were allowed to buy and sell
securities associated with these levels. To make money, participants had to accu-
rately predict the level of sales HP would experience three months in the future.
If participants accurately did so (by holding securities within that range), they
would receive a payoff, but they would receive nothing for securities held at any
other level.9
Chen’s and Plott’s futures market consistently yielded better marketing fore-
casts than did HP’s official predictions, arrived at through deliberation by the
company’s sales forecasting experts.10 Why? Because it provided an effective
mechanism for aggregating the specialized knowledge of an independent and
diverse group of people, and because the participants had no reason to strategi-
cally distort their predictions.
It is this efficient aggregation of decentralized, privately held information that
makes projects like the Iowa Electronic Markets (IEM) so interesting. IEM is a
small-scale futures market where interested individuals make contracts based on
their own predictions of economic and political events, such as elections. In
2008, for example, individuals could make contracts based on whether Barack
Obama or John McCain would win the U.S. Presidential election. The market is
operated by faculty at the University of Iowa’s Henry B. Tippie College of
Business; the results routinely outperform polls and surveys in predicting elec-
toral outcomes.11 A similar idea is the Hollywood Stock Exchange (HSX), which
allows people to place wagers on movie box office returns, Oscar results, and
other film-related events.12 HSX produces surprisingly accurate predictions,
demonstrating the power of markets to aggregate the information dispersed over
a large population of independent actors.13
***
Decision makers must often forecast future events in order to assess the
advantages and disadvantages of alternative courses of action. These examples
suggest that information markets can sometimes do a better job than experts.
But it is often not feasible to use statistical groups, and the responsibility falls on
19.3 groupthink
14. Irving L. Janis, Groupthink, 5 Psychology Today 43–46, 74–76 (1971); Irving L. Janis,
Victims of Groupthink: A Psychological Study of Foreign Policy Decisions and
Fiascos (Boston: Houghton Mifflin, 1972); Irving L. Janis, Groupthink: Psychological
Studies of Policy Decisions and Fiascos (Boston: Houghton Mifflin, 1982); Irving L.
Janis, Crucial Decisions: Leadership in Policymaking and Crisis Management (New
York: Free Press, 1989); Irving L. Janis and Leon Mann, Decision Making: A Psychological
Analysis of Conflict, Choice and Commitment (New York: Free Press, 1977).
15. Robert S. Baron, So Right It’s Wrong: Groupthink and the Ubiquitous Nature of Polarized
Group Decision Making, in Advances in Experimental Social Psychology 37, 219 (Mark
P. Zanna, ed., San Diego, CA: Elsevier Academic Press, 2005); Mark Schafer and Scott
Crichlow, Antecedents of Groupthink, 40 Journal of Conflict Resolution 415–35 (1996).
group decision making and the effects of accountability 601
Divergent
Thinking Frame the decision to be made
Identify objectives
Identify alternative courses of action
Evaluate alternative course of action vs.
objectives
Make a decision
Convergent Implement and monitor the decision
Thinking
Make a decision
Implement and monitor the decision
From his examination of selected policy fiascos and successes, Janis identi-
fied four sets of attributes characterizing groupthink.
1. groupthink symptoms;
2. decision-making characteristics;
3. antecedents; and
4. remedial interventions.
We review the first three here and will return to the last in Section 19.8.
21. Id.; James Esser, Alive and Well After 25 years: A Review of Groupthink Research, 73
Organizational Behavior and Human Decision Processes 116–41 (1998).
22. Matie L. Flowers, A Laboratory Test of Some Implications of Janis’s Groupthink
Hypothesis, 35 Journal of Personality and Social Psychology 888–96 (1977); Eugene
M. Fodor and Timothy Smith, The Power Motive as an Influence on Group Decision Making,
42 Journal of Personality and Social Psychology 178–85 (1982); Michael R. Callaway
and James K. Esser, Groupthink: Effects of Cohesiveness and Problem-Solving Procedures on
Group Decision Making, 12 Social Behavior and Personality 157–64 (1984).
group decision making and the effects of accountability 605
23. Amiram Vinokur and Eugene Burnetein, The Effects of Partially Shared Persuasive
Arguments on Group Induced Shifts: A Group Problem Solving Approach, 29 Journal of
Personality and Social Psychology 305–15 (1974).
24. James R. Larson et al., Discussion of Shared and Unshared Information in Decision
Making Groups, 67 Journal of Personality and Social Psychology 446–61 (1994);
Garold Stasser and Dennis Stewart, Discovery of Hidden Profiles by Decision Making Groups:
Solving a Problem versus Making a Judgment, 63 Journal of Personality and Social
Psychology 426–34 (1992); Garold Stasser and William Titus, Pooling of Unshared
Information in Group Decision Making: Biased Information Sampling During Discussion, 48
Journal of Personality and Social Psychology 1467 (1985).
25. Daniel Gigone and Reid Hastie, The Common Knowledge Effect: Information Sharing
and Group Judgment, 65 Journal of Personality and Social Psychology 959–74 (1993).
The phenomenon is also called the “common information sampling bias.” Gwen M.
Wittenbaum and Garold Stasser, Management of Information in Small Groups, in What’s
Social About Social Cognition? Research on Socially Shared Cognition in Small
Groups 3–28 (Judith L. Nye and Aaron M. Brower eds., Thousand Oaks, CA: Sage, 1996).
26. Gigone and Hastie, supra.
606 problem solving, decision making, and professional judgment
27. Garold Stasser and William Titus, Pooling of Unshared Information in Group
Decision Making: Biased Information Sampling During Discussion, 48 Journal of
Personality and Social Psychology 1467 (1985); Garold Stasser and William Titus,
Hidden Profiles: A Brief History, 14 Psychological Inquiry 304 (2003).
28. See Garold Stasser and William Titus, Hidden Profiles: A Brief History, 14
Psychological Inquiry 304 (2003).
29. Gwen M. Wittenbaum and Jonathan M. Bowman, A Social Validation Explanation
for Mutual Enhancement, 40 Journal of Experimental Social Psychology 169–84
(2004); Gwen M. Wittenbaum, Anne P. Hubbell, and Connie Zuckerman, Mutual
Enhancement: Toward an Understanding of the Collective Preference for Shared Information,
77 Journal of Personality and Social Psychology 967–78 (1999); Janice R. Kelly and
Steven J. Karau, Group Decision Making: The Effects of Initial Preferences and Time Pressure,
25 Personality and Social Psychology Bulletin 1342–54 (1999); Garold Stasser,
Sandra Vaughn, and Dennis Stewart, Pooling Unshared Information: The Benefits of Knowing
How Access to Information Is Distributed Among Group Members, 82 Organizational
Behavior and Human Decision Processes 102–16 (2000); Jennifer R. Winquist and
James R. Larson, Information Pooling: When It Impacts Group Decision Making, 74 Journal
of Personality and Social Psychology 371–77 (1998); James R. Larson, Pennie G.
Foster-Fishma, and Timothy M. Franz, Leadership Style and the Discussion of Shared and
Unshared Information in Decision Making Groups, 24 Personality and Social Psychology
Bulletin 482–95 (1998); James R. Larson et al., Diagnosing Groups: The Pooling,
Management, and Impact of Shared and Unshared Case Information in Team-based Medical
Decision Making, 75 Journal of Personality and Social Psychology 93–108 (1998);
Gwen M. Wittenbaum, Information Sampling in Decision Making Groups: The Impact of
group decision making and the effects of accountability 607
Members Task-Relevant Status, 29 Small Group Research 57–84 (1998); Garold Stasser
and Gwen M. Wittenbaum, The Reevaluation of Information During Group Discussion, 1
Group Processes & Intergroup Relations 21–34 (1998); Dennis Stewart and Garold
Stasser, Expert Role Assignment and Information Sampling During Collective Recall and
Decision Making, 69 Journal of Personality and Social Psychology 619–28 (1995);
James R. Larson, Pennie G. Foster-Fishman, and Christopher B. Keys, Discussion of Shared
and Unshared Information in Decision Making Groups, 67 Journal of Personality and
Social Psychology 446–61 (1994); Daniel Gigone and Reid Hastie, The Common
Knowledge Effect: Information Sharing and Group Judgment, 65 Journal of Personality
and Social Psychology 959–74 (1993); Daniel Gigone and Reid Hastie, The Impact of
Information on Group Judgment: A Model and Computer Simulation, in Understanding
Group Behavior: Vol. 1. Consensual Action by Small Groups (Erich H. White and
James H. Davis eds., Hillsdale, NJ: Erlbaum, 1996); Garold Stasser and Dennis Stewart,
Discovery of Hidden Profiles by Decision Making Groups: Solving a Problem versus Making a
Judgment, 63 Journal of Personality and Social Psychology 426–34 (1992); Garold
Stasser, Laurie A. Taylor, and Coleen Hanna, Information Sampling in Structured and
Unstructured Discussions of Three- and Six-Person Groups, 57 Journal of Personality and
Social Psychology 67–78 (1989); Garold Stasser and William Titus, Effects of Information
Load and Percentage of Common Information on the Dissemination of Unique Information
During Group Discussion, 53 Journal of Personality and Social Psychology 81–93
(1987).
30. Craig D. Parks and Rebecca Cowlin, Acceptance of Uncommon Information into
Group Discussion When That Information Is or Is Not Demonstable, 66 Organizational
Behavior and Human Decision Processes 307–15 (1996).
31. Id.
32. Id.
608 problem solving, decision making, and professional judgment
33. Steven J. Karau and Janie R. Kelly, The Effects of Time Scarcity and Time Abundance
on Group Performance Quality and Interaction Process, 28 Journal of Experimental
Social Psychology 542–71 (1992).
34. James R. Larson, Pennie G. Foster-Fishman, and Christopher B. Keys, Discussion
of Shared and Unshared Information in Decision Making Groups, 67 Journal of Personality
and Social Psychology 446–61 (1994); Dennis Stewart, Robert S. Billings, and Garold
Stasser, Accountability and the Discussion of Unshared, Critical Information in Decision
Making Groups, 2 Group Dynamics 18–23 (1998).
35. Garold Stasser and William Titus, Effects of Information Load and Percentage of
Common Information on the Dissemination of Unique Information During Group Discussion,
53 Journal of Personality and Social Psychology 81–93 (1987); James R Larson,
Pennie G. Foster-Fishman, and Christopher B. Keys, Discussion of Shared and Unshared
Information in Decision Making Groups, 67 Journal of Personality and Social
Psychology 446–61 (1994); Garold Stasser, Information Salience and the Discovery of
Hidden Profiles by Decision-Making Groups: A Thought Experiment, 52 Organizational
Behavior and Human Decision Processes 5156–81 (1992).
group decision making and the effects of accountability 609
Wittenbaum and her collaborators indicate that the discussion of shared infor-
mation leads people both to feel more like a “group” and to feel better about that
group.39
Conversational norms. As every good lawyer and politician knows, effective
communicators tailor their messages to what they believe to be the knowledge,
background, and expectations of their audience. By creating a “cognitive common
ground,” one makes the message both more comprehensible and more persua-
sive to the communicator.
This phenomenon may play a role in the common knowledge effect. If I am a
member of a group, and want to make a point, I may begin by mentioning infor-
mation that is believed by others in the group. If group members follow the
rhetorical axiom, “begin your remarks by mentioning facts you know others
believe true,” they will mention shared information more frequently than
unshared information.40 In this way, the common knowledge effect can result
from the operation of a simple conversational norm.
19.4.3 Reducing the Bias in Favor of Shared Information: What Works, What
Doesn’t, What Makes Things Worse?
Empirical research on the conditions that exacerbate or ameliorate the common
knowledge effect has yielded two generalizations.
First, forewarning people that they may have relevant information that other
group members may not possess has no effect.41 Nor does instructing group
members not to come to judgments about the best course of action before all
information possessed by group members is surfaced.42
43. Steven J. Karau and Janice R. Kelly, The Effects of Time Scarcity and Time Abundance
on Group Performance Quality and Interaction Process, 28 Journal of Experimental Social
Psychology 542–71 (1992); James R. Larson, Pennie G. Foster-Fishman, and Christopher
B. Keys, Discussion of Shared and Unshared Information in Decision Making Groups, 67
Journal of Personality and Social Psychology 446–61 (1994); Dennis D. Stewart,
Robert S. Billings, and Garold Stasser, Accountability and the Discussion of Unshared,
Critical Information in Decision Making Groups, 2 Group Dynamics 18–23 (1998).
44. Garold Stasser and Dennis Stewart, Discovery of Hidden Profiles by Decision Making
Groups: Solving a Problem versus Making a Judgment, 63 Journal of Personality and
Social Psychology 426–34 (1992).
45. Dennis Stewart and Garold Stasser, Expert Role Assignment and Information
Sampling During Collective Recall and Decision Making, 69 Journal of Personality and
Social Psychology 619–28 (1995); Garold Stasser, Sandra I. Vaughan, and Dennis D.
Stewart, Pooling Unshared Information: The Benefits of Knowing How Access to Information
Is Distributed Among Group Members, 82 Organizational Behavior and Human
Decision Processes 102–16 (2000).
46. James R. Larson, Pennie G. Foster-Fishman, and Christopher B. Keys, Discussion
of Shared and Unshared Information in Decision Making Groups, 67 Journal of Personality
and Social Psychology 446–61 (1994); Craig D. Parks and Rebecca Cowlin, Group
Discussion as Affected by Number of Alternatives and by a Time Limit, 62 Organizational
Behavior and Human Decision Processes 267–75 (1995).
47. Craig D. Parks and Rebecca Cowlin, Group Discussion as Affected by Number of
Alternatives and by a Time Limit, 62 Organizational Behavior and Human Decision
Processes 267–75 (1995).
48. Craig D. Parks and Rebecca Cowlin, Acceptance of Uncommon Information into
Group Discussion When That Information Is or Is Not Demonstrable, 66 Organizational
Behavior and Human Decision Processes 307–15 (1996).
612 problem solving, decision making, and professional judgment
Consider again the Terra Nueva lawsuit, and the efforts of community and con-
sumer activists to use it to mobilize other community groups to take actions
relating to toxic exposure-related harms. Imagine that going into the meeting
described above, representatives from East Los Angeles Tenants United (ELATU),
L.A. Consumers’ Action (LACA), and the Terra Nueva Tenants’ Union are enthu-
siastic about filing similar suits elsewhere. But the organizer from the United
Electrical Workers (UEW) is more cautious. She wants to see how the Terra
Nueva suit unfolds before allocating more union resources to similar initia-
tives—in part, because she thinks that toxic chemicals, rather than the foam
insulation, might be causing the Terra Nueva tenants’ health problems. Suppose
that the plaintiffs’ lawyers have learned about this possibility as well, and that
they share her concern about taking action in other communities, which at this
point seems a risky strategy.
Assuming that the UEW representative and the lawyers make their knowl-
edge and views known to the rest of the group, what do you think will happen
during this meeting? As different strategy options are discussed, will the UEW
representative’s and lawyers’ concerns moderate the enthusiasm for expanding
the litigation strategy with which the other organizations came into the meeting?
Or will deliberation simply embolden participants to take broader actions than
they might have decided on alone?
Your intuition might be that deliberation would tend to temper the judgment
of a group’s most extreme members on either side and draw the group toward a
middle ground.49 But it turns out that rather than moderating the views of the
community organizers, the group may be pulled toward their more extreme
views.50 Group discussion tends to magnify differences within the group and
enhance the position of those whose views initially predominate.
In an early experiment involving what has come to be called group polarization,51
James Stoner presented participants with hypothetical problems involving a
49. See Rupert Brown, Group Processes (2d ed. Malden, MA: Blackwell Publishing,
2000).
50. Cass R. Sunstein, The Law of Group Polarization, 10 Journal of Political
Philosophy 175–95 (2002).
51. Serge Moscovici and Marisa Zavalloni, The Group as a Polarizer of Attitudes, 12
Journal of Personality and Social Psychology 125–35 (1969); David G. Myers and
group decision making and the effects of accountability 613
choice between cautious and risky courses of action. After making tentative deci-
sions on their own, the participants were placed in groups and asked to reach a
unanimous decision. When participants’ median position before discussion was
risk seeking, group deliberation led to a “risky shift,”52 in which decisions were
riskier than their individual prediscussion choices. Subsequent studies have
shown that “cautious” shifts occur when the predominant view of the individual
participants is risk averse.53 (In addition to the group’s tending to shift to one
extreme or another, individual members internalize the group’s position and
tend to hold it after the discussion has ended.54)
Group polarization is not restricted to shifts in risk preference. For example,
David Myers demonstrated that groups of moderately pro-feminist women sub-
jects became more strongly pro-feminist following discussion of feminist ideas.55
Jurors are affected by group polarization in awarding damages.56 And Cass
Sunstein and his colleagues recently demonstrated that circuit court judges’
ideological tendencies are amplified when they sit on panels consisting of judges
appointed by presidents from the same political party. Republican appointees
show an increased tendency to vote in a stereotypically conservative fashion
when they are accompanied by two other Republican appointees, while
Democratic appointees show the same effect in the opposite direction when
accompanied by their co-partisans.57
57. Cass R. Sunstein, David Schkade, and Lisa M. Ellman, Ideological Voting on Federal
Courts of Appeals: A Preliminary Investigation, 90 Virginia Law Review 301–55 (2004).
Similar polarization effects in panel and individual judicial decision making was found
three decades earlier by Walker and Main. Thomas G. Walker and Eleanor C. Main, Choice
Shifts in Political Decisionmaking: Federal Judges and Civil Liberties Cases, 3 Journal of
Applied Social Psychology 39–48 (1973).
58. Robert S. Baron and Gard Roper, Reaffirmation of Social Comparison Views of Choice
Shifts: Averaging and Extremity Effects in an Autokinetic Situation, 33 Journal of Personality
and Social Psychology 521–30 (1976); Daniel J. Isenberg, Group Polarization: A Critical
Review and Meta-Analysis, 50 Journal of Personality and Social Psychology 1141–51
(1986); Joel Cooper, Kimberly A. Kelly, and Kimberlee Weaver, Attitudes, Norms, and Social
Groups, in Blackwell Handbook of Social psychology: Group Processes 259–82
(Michael A. Hogg and R. Scott Tindale eds., Malden, MA: Blackwell Publishers, 2001).
59. John C. Turner, Social Influence (Pacific Grove, CA: Brooks/Cole, 1991).
60. Robert S. Baron et al., Social Corroboration and Opinion Extremity, 32 Journal of
Experimental Social Psychology 537–60 (1996).
group decision making and the effects of accountability 615
group from others, members attribute more extremity to their own group than
is objectively the case; they distinguish their group from others by accentuating
the perceived distance between their group norm and the norms they attribute to
out-groups.61 Group members then conform their own behaviors, norms, and
self-conceptions to their group’s.62 Polarization thus occurs through a process of
contrast with other groups within a specific social context.63
The social accounts of group polarization are not mutually exclusive. Indeed,
each incorporates theoretical elements of the other and experimental research
supports all of them.64
19.5.1.b Cognitive/Informational Accounts of Group Polarization A second set
of theories focuses on cognitive rather than social explanations for group polar-
ization. The persuasive argument theory builds on the observation that, before
group discussion of an issue begins, a member will usually have formulated an
initial position and formulated arguments favoring his position. The group dis-
cussion exposes the participant to new arguments that support or detract from
his initial position. The phenomenon of confirmation bias suggests that the par-
ticipant will find new information and arguments that support his initial posi-
tion more persuasive than information and arguments that oppose it. As a purely
statistical phenomenon, this predicts group polarization, in much the same way
65. Eugene Burnstein and Amiram Vinokur, What a Person Thinks upon Learning He
Has Chosen Differently from Others: Nice Evidence for the Persuasive Arguments Explanation
of Choice Shifts, 11 Journal of Experimental Social Psychology 412–26 (1975); Eugene
Burnstein and Amiram Vinokur, Persuasive Argumentation and Social Persuasion as
Determinants of Attitude Polarization, 13 Journal of Experimental Social Psychology
315–32 (1977); Markus Brauer, Charles M. Jud, and Melissa D. Gliner, The Effects of
Repeated Expressions on Attitude Polarization During Group Discussions, 68 Journal of
Personality and Social Psychology 1014–29 (1995).
66. Chip Heath and Richard Gonzalez, Interaction with Others Increases Decision
Confidence but Not Decision Quality: Evidence Against Information Collection Views of
Interactive Decisionmaking, 61 Organizational Behavior and Human Decision
Processes 305 (1995).
67. Markus Brauer, Charles M. Jud, and Melissa D. Gliner, The Effects of Repeated
Expressions on Attitude Polarization During Group Discussions, 68 Journal of Personality
and Social Psychology 1014–29 (1995).
68. Harris Bernstein, Persuasion as Argument Processing, in Group Decisionmaking
(Hermann Brandstatter, James H. Davis, and Gisela Stocker-Kreichgauer eds., London:
Academic Press, 1982); Rupert Brown, Group Processes (2d ed. Malden, MA: Blackwell
Publishing, 2000); Cass Sunstein, Deliberative Trouble? Why Groups Go to Extremes, 110
Yale Law Journal 72–119 (2000).
group decision making and the effects of accountability 617
• Groups that work together over a sustained period of time, making numer-
ous similar decisions, may get feedback on the quality of those decisions,
which may moderate overconfidence and lead to more effective use of
information in subsequent decision making.
• Polarization is moderated where a group is subject to external shocks (e.g.,
losing an election, being sued, experiencing a severe policy failure) or
where the group must temper its preferences to preserve its prestige with
external constituencies.
• Polarization can be reduced by increasing decision-making independence
through mechanisms like the Delphi technique (Section 19.8.4), and
through norms and facilitation that encourage open-mindedness, the
inclusion of minority perspectives, and evidence-based decision making.69
Group members’ concerns about their relations with others in the group play a
significant role in the pathologies of group decision making. Individual group
members may have interests that differ from those of their colleagues. So, for
example, in a meeting to help a company’s managers decide whether to settle or
litigate an employment discrimination case, the lawyer may be concerned to
maintain the impression that she is “on their side” even as she attempts to
impress upon them the nature of the exposure to risk that they face. She may
also want to convey a gender counter-stereotypical “toughness” that she hopes
will instill confidence in her competence as a vigorous advocate.
Other participants in the decision-making process may have their own sets of
competing processing goals. The manager who made the now-challenged deci-
sion may seek to save face, preventing his colleagues from concluding that he
made a mistake or otherwise acted imprudently. Yet another group member may
simply want to create a good impression and be viewed as a “team player.” And
all (or almost all) of the group members will be attempting to maintain a certain
level of social harmony, both during and after the decision making process.
In other words, the goal of making a “good” decision is only one of the items
on members’ individual and collective agendas. The susceptibility of a group to
decision-making errors will depend in substantial measure on these often hidden
competing motivations.
69. Tom Postmes, Russell Spears, and Sezgin Cihangir, Quality of Decisionmaking and
Group Norms, 80 Journal of Personality and Social Psychology 918–30 (2001).
618 problem solving, decision making, and professional judgment
Earlier in the book, we described a variety of biases and heuristics that system-
atically distort individuals’ judgment and decision making. Despite some hopes
to the contrary,70 groups are not less susceptible to these phenomena than indi-
viduals, and in some circumstances are more so.
Overconfidence. As we saw in Section 10.4, individuals often are overconfi-
dent in the accuracy of their own judgment. It turns out that groups don’t do
better—even when even they appoint a “devil’s advocate” or explicitly consider
reasons why their estimates might be wrong—and may actually be more over-
confident.71 Group overconfidence may result from the group polarization
effect.72 Persuasive argument theory73 suggests that if (as often happens) group
members enter the group discussion with an overconfident orientation, they will
make more confidence-inducing than caution-supporting arguments, thus skew-
ing overall group opinion in an optimistic direction. Second, social comparison
processes may come into play, as group members offer overconfident outlooks
on the subject of group discussion so as to demonstrate loyalty and commitment
to the project, and to assure that they compare favorably in this regard with other
group members.74
Confirmation bias. Small deliberating groups, no less than individuals,
fall prey to the tendency to overweight information that confirms a prior
position, theory, or attitude and to underweight information that would tend to
disconfirm it. Groups manifest a preference for information that confirms indi-
vidual members’ prediscussion preferences. The more social support for their
views participants anticipated from an upcoming group discussion, the more
preference they showed for confirming information.75 For both laypeople and
70. George P. Huber, Managerial Decision Making (Glenview, IL: Scott Foresman,
1980).
71. Scott Plous, A Comparison of Strategies for Reducing Interval Overconfidence in Group
Judgments, 80 Journal of Applied Psychology 443–54 (1995).
72. Roger Buehler, Deanna Messervey, and Dale Griffin, Collaborative Planning and
Prediction: Does Group Discussion Affect Optimistic Biases in Time Estimation?, 91
Organizational Behavior and Human Decision Processes 47–63 (2005).
73. Amiram Vinokur and Eugene Burnetein, The Effects of Partially Shared Persuasive
Arguments on Group Induced Shifts: A Group Problem Solving Approach, 29 Journal of
Personality and Social Psychology 305–15 (1974).
74. Daniel Kahneman and Dan Lovallo, Timid Choices and Bold Forecasts: A Cognitive
Perspective on Risk Taking, 39 Management Science 17–31 (1993).
75. Bozena Zdaniuk and John M. Levine, Anticipated Interaction and Thought
Generation: The Role of Faction Size, 36 British Journal of Social Psychology 201–18
(1996). On a related matter, group discussion may attenuate or eliminate the theory-per-
severance effect, in which people retain a belief in the functional relationship between two
variables even after evidence for any such relationship has been discredited. Edward F.
group decision making and the effects of accountability 619
groups of experts making decisions within their areas of expertise (say, business
managers making economic decisions), the more homogeneous the members’
preferences, the stronger the effect.76 Only when the deliberating group is
reasonably balanced on the issue in question does confirmation bias diminish.
Various judgmental errors. The little research that exists provides no clear pic-
ture about the relative susceptibility of individuals and groups to the availability
and representativeness heuristics.77 Group discussion seems to amplify conjunc-
tion errors when the individual members’ susceptibility to conjunction bias is
high, and to attenuate the errors when it is low.78 Group interaction does not
appear to attenuate base rate neglect and may, indeed, increase it.79 The evidence
on groups’ proneness to hindsight bias is equivocal.80 Deliberating groups appear
to be even more prone to unrealistic optimism than individuals;81 this may result
from the same informational, social comparison, and social identity manage-
ment processes mentioned in connection with group polarization.82
Wright and Scott D. Christie, “The Impact of Group Discussion on the Theory-Perseverance
Effect” (unpublished manuscript, St. Francis Xavier University, Department of Psychology,
Antigonish, NS). Craig A. Anderson, Inoculation and Counterexplanation: Debiasing
Techniques in the Perseverance of Social Theories, 1 Social Cognition 126–39 (1982).
76. Stefan Schulz-Hardt et al., Biased Information Search in Group Decisionmaking, 78
Journal of Personality and Social Psychology 655–69 (2000).
77. Mark F. Stasson et al., Group Consensus Processes on Cognitive Bias Tasks: A Social
Decision Scheme Approach, 30 Japanese Psychological Research 68–77 (1988).
78. R. Scott Tindale, Susan Sheffey, and Joseph Filkins, “Conjunction Errors by
Individuals and Groups” (paper presented at the meetings of the Society for Judgment
and Decisionmaking, New Orleans, LA, 1990); R. Scott et al., “An Attempt to Reduce
Conjunction Errors in Decision-Making Groups” (paper presented at the annual meeting
of the Society for Judgment and Decisionmaking, Washington, DC, 1993). Other research-
ers obtained mixed results in studies testing susceptibility to other types of representative-
ness biases, finding bias amplification in some contexts, but not others. Mark F. Stasson
et al., Group Consensus Processes on Cognitive Bias Tasks: A Social Decision Scheme Approach,
30 Japanese Psychological Research 68–77 (1988).
79. Linda Argote, Mark A. Seabright, and Linda Dyer, Individual versus Group Use of
Base-Rate and Individuating Information, 38 Organizational Behavior and Human
Decision Processes 65–75 (1986); Linda Argote, Rukmini Devadas, and Nancy Melone,
The Base-Rate Fallacy: Contrasting Processes and Outcomes of Group and Individual Judgment,
48 Organizational Behavior and Human Decision Processes 296–311 (1990); R.
Scott Tindale, Decision Errors Made by Individuals and Groups, in N. John Castellan Jr.
(ed.), Individual and Group Decision Making: Current Issues 109–24 (Hillsdale, NJ,
England: Lawrence Erlbaum Associates, Inc., 1993).
80. Dagmar Stahlberg et al., We Knew It All Along: Hindsight Bias in Groups, 63
Organizational Behavior and Human Decision Processes 46–58 (1995).
81. Roger Buehler, Deanna Messervey, and Dale Griffin, Collaborative Planning and
Prediction: Does Group Discussion Affect Optimistic Biases in Time Estimation?, 91
Organizational Behavior and Human Decision Processes 47–63 (2005).
82. Id.
620 problem solving, decision making, and professional judgment
Gain/loss framing effects and other departures from expected utility theory.
The research on group susceptibility to framing effects shows mixed results.
One study found that group discussion amplified the asymmetrical treatment of
potential losses and gains,83 while another found that group discussion attenu-
ated framing bias.84 Yet another found bias when the outcome had been framed
by individual group members as a loss or a gain before group discussion began;
but framing bias was attenuated if the outcome was not framed until after group
discussion began, or if it was reframed in the course of the discussion.85 The
research does not suggest any robust generalizations about gain/loss framing or,
consistency with expected utility theory more generally.86
Escalation of commitment. Some research suggests that small groups are, if
anything, more prone than individuals to escalate commitment to a failing
course of action.87 This should not be surprising, since the escalation of commit-
ment is caused, at least in part, by the desire to appear consistent to others,88 and
since group polarization could readily contribute to escalation. However, other
research found no difference between individuals and groups in the amounts of
resources they were willing to invest in a failing endeavor.89
The fundamental attribution error. Curiously, group discussion appears to
attenuate susceptibility to the fundamental attribution error—the tendency to
83. Timothy W. McGuire, Sara Kiesler, and Judith Siegel, Group and Computer-
Mediated Discussion Effects in Risk Decisionmaking, 52 Journal of Personality and
Social Psychology 917–30 (1987).
84. Margaret A. Neale et al., “Choice Shift” Effects in Group Decisions: A Decision Bias
Perspective, 2 International Journal of Small Group Research 33–42 (1986).
85. Paul W. Paese, Mary Bieser, and Mark E. Tubbs, Framing Effects and Choice Shifts
in Group Decisionmaking, 56 Organizational Behavior and Human Decision
Processes 149–65 (1993).
86. Norbert L. Kerr, Robert J. MacCoun, and Geoffrey P. Kramer, Bias in Judgment:
Comparing Individuals and Groups, 103 Psychological Review 687–719 (1996); John C.
Mowen and James W. Gentry, Investigation of the Preference-Reversal Phenomenon in a New
Product Introduction Task, 65 Journal of Applied Psychology 715–22 (1980); John Bone,
John Hey, and John Suckling, Are Groups More (or Less) Consistent than Individuals?, 8
Journal of Risk and Uncertainty 63–81 (1999).
87. Glen Whyte, Escalating Commitment in Individual and Group Decisionmaking, 54
Organizational Behavior and Human Decision Processes 430 (1993).
88. Barry Staw and Jerry Ross, Behavior in Escalation Situations: Antecedents, Prototypes,
and Solutions, in Research in Organizational Behavior (Larry L. Cummings and Barry
Staw eds., Greenwich: CT: JAI Press, 1987).
89. Max H. Bazerman, Toni Giuliano and Allan Appelman, Escalation of Commitment
in Individual and Group Decisionmaking, 33 Organizational Behavior and Human
Performance 141–52 (1984).
group decision making and the effects of accountability 621
This section considers several approaches to reducing the pathologies that can
infect group decision making. These include measures suggested by Irving Janis
and his collaborators to prevent groupthink and other forms of premature con-
vergence, and two mechanisms—the “stepladder” and Delphi techniques—that
can be used to maximize the constructive contribution of each member to a
decision-making group.
90. Edward F. Wright and Gary L. Wells, Does Group Discussion Attenuate the
Dispositional Bias?, 15 Journal of Applied Psychology 531–46 (1985).
91. Irving L. Janis, Groupthink: Psychological Studies of Policy Decisions
and Fiascos (Boston, MA: Houghton Mifflin, 1982); Irving L. Janis and Leon Mann,
Cognitive Complexity in International Decisionmaking, in Psychology and Social
Policy 33–49 (Peter Suedfeld and Philip Tetlock eds., Washington, DC: Hemisphere
Publishing Corp., 1992).
92. Alexander S. Haslam, Psychology in Organizations: The Social Identity
Approach (London, UK & Thousand Oaks, CA: Sage Publications, 2001).
622 problem solving, decision making, and professional judgment
is obtained from experts outside the group, and even appoint formal devil’s
advocates to aggressively question any emerging consensus. As Clark McCauley
has described the leader’s task, “If groupthink is a disease of insufficient
search for information, alternatives, and modes of failure, the cure is better
search procedures: impartial leadership that encourages airing of doubts and
objections in the group and encourages search for information and evaluation
from sources outside the group, procedural norms supporting systematic con-
sideration of alternatives, ‘second-chance’ meetings after preliminary consen-
sus, and special attention to the motives and communications of threatening
competitors.”93
Janis’ specific suggestions for enhancing vigilant decision making include
the following:
1. The leader should be impartial and should not state or even imply prefer-
ences or expectations about the decision at the outset.
2. The leader should assign the role of critical evaluator to each member,
encouraging all participants to vigorously air objections and doubts.
3. Periodically, the group should be broken up into smaller subgroups,
under different leaders, to work on the same question. Then the groups
should be brought back together to compare perspectives and discuss
differences.
4. Periodically, each group member should discuss the subject under delib-
eration with trusted associates who are not in the decision-making group,
and report back to the group their ideas and reactions.
5. Experts and qualified colleagues who are not members of the decision-
making group should periodically be invited to attend group meetings and
challenge the views of the group’s members.
6. At each meeting devoted to the discussion of policy alternatives, at least
one member should be assigned the role of devil’s advocate.
7. Whenever the issue being discussed involves relations with a rival organi-
zation, nation, or other collectivity, ample time should be allocated to
constructing varying conceptions of the rival’s intentions.
8. After the group reaches a preliminary consensus about the best decision,
the leader should conduct a “second chance” meeting at which group
members are instructed to express, as forcefully and persuasively as they
can, all of their doubts about the preliminary consensus, and to reconsider
the entire problem before making a final choice.
93. Clark McCauley, Group Dynamics in Janis’s Theory of Groupthink: Backward and
Forward, 73 Organizational Behavior and Human Decision Processes 142–62
(1998).
group decision making and the effects of accountability 623
98. Joseph S. Valacich and Charles Schwenk, Devil’s Advocacy and Dialectical Inquiry
Effects on Face-to-Face and Computer-Mediated Group Decisionmaking, 63 Organizational
Behavior and Human Decision Processes 158–73 (1995).
99. Leonard M. Jessup, Terry Connolly, and Jolene Galegher, The Effects of Anonymity
on Group Process in an Idea-Generating Task, 14 MIS Quarterly 313–21 (1990).
100. Joseph S. Valacich and Charles Schwenk, Devil’s Advocacy and Dialectical Inquiry
Effects on Face-to-Face and Computer-Mediated Group Decisionmaking, 63 Organizational
Behavior and Human Decision Processes 158–73 (1995).
101. Dennis Falk and David W. Johnson, The Effects of Perspective-Taking and Egocentrism
on Problem Solving in Heterogeneous and Homogeneous Groups, 102 Journal of Social
Psychology 63–72 (1977).
102. Bibb Latane, Kipling Williams, and Stephen Harkins, Many Hands Make Light the
Work: The Causes and Consequences of Social Loafing, 37 Journal of Personality and
Social Psychology 822–32 (1979).
103. David Johnson and Frank Johnson, Joining Together: Group Theory and
Group Skills (Engelwood Cliffs, NJ: Prentice-Hall, 1987).
104. Steven G. Rogelberg, Janet L. Barnes-Farrell, and Charles A. Lowe, The Stepladder
Technique: An Alternative Group Structure Facilitating Effective Group Decisionmaking, 77
Journal of Applied Technology 730–37 (1992).
group decision making and the effects of accountability 625
made until all group members have been incorporated into the group in this
manner.
The stepladder technique tends to improve a group’s decision performance.
Stepladder groups outperformed both their average and best member more
often than conventional deliberating groups.
The discussion so far has focused on groups coming to conclusions about issues
recognized as requiring a decision. But organizations may have impediments to
even recognizing the need for a decision. They can become trapped in a particu-
lar mindset, and be unable or unwilling to adjust to a changing reality.
105. Norbert L. Kerr and R. Scott Tindale, Group Performance and Decisionmaking, 55
Annual Review of Psychology 623–55 (2004); Brian Mullen, Craig Johnson, and
Eduardo Salas, Productivity Loss in Brainstorming Groups: A Meta-Analytic Integration, 115
Psychological Bulletin 210–27 (1991).
106. Gene Rowe and George Wright, Expert Opinions in Forecasting: The Role of the
Delphi Technique, in Principles of Forecasting 125–44 (J. Scott Armstrong ed., Norwell,
MA: Kluwer Academic Publishers, 2001).
626 problem solving, decision making, and professional judgment
For all of the hazards described in the preceding pages, group decision making
is inevitable. In an increasingly complex world, many problems cannot be ade-
quately addressed by a single individual with limited knowledge and cognitive
capacity. Groups can pool knowledge, information, and experience. They can
divide a complex task into multiple components. And groups have the potential
to correct members’ errors, to reduce decision variability, and moderate indi-
vidual idiosyncrasies. Moreover, participation in making decisions can increase
people’s motivation and sense of commitment, and can enhance the perceived
legitimacy of whatever course of action is chosen.
Rupert Brown noted, “[T]he question is not . . . how we can dispense with
group decision making, but how we can capitalize on its advantages while cir-
cumventing its defects.”108 The key to answering his question lies in assembling
a decision-making group whose members bring diverse knowledge sets, skills,
and perspectives to the task, and then utilizing the potential unique contribution
of each member. The independence of group members must be preserved, so
that they can, in fact, make their diverse contributions. This requires addressing
pressures toward social conformity, power imbalances, and members’ needs to
maintain their social position. The decision-making process must be designed to
ensure that the group considers a broad range of alternative options before con-
verging on a decision. And, where the group is to persist over time and make
repeated decisions, mechanisms must be designed to aggregate and re-present
information to enable the group and its members to receive feedback and learn
from experience.
Lawyers and policy makers regularly serve as the leaders of decision-making
groups. The leader’s role is key in harnessing the power of groups and in pre-
venting the pitfalls that attend group decision making. Understanding the
108. Rupert Brown, Group Processes (2d ed. Malden, MA: Blackwell Publishing,
2000).
628 problem solving, decision making, and professional judgment
The preceding discussion has focused on the interaction of decision makers with
other people. But knowledge that one will be required to justify one’s decision to
others may influence one’s decision, even without any actual interaction with
one’s audience. Mere anticipation of others’ reactions may cause decision makers
to change their approach to the problem.
Accountability has become a popular concept in business, politics, and else-
where, but its effects are not always clearly understood. Received wisdom tells us
that that a decision maker’s accountability to others improves the quality of the
decision. Though this sometimes is true, accountability can also lead to poorer
decision making. This section summarizes the literature on the effects of
accountability.109
All other things being equal, accountability is likely to reduce error and bias
in contexts in which, for whatever reasons, people tend to make mistakes that
they could prevent with extra attention or effort. For example, the expectation
that they will have to justify their decision leads people to think more carefully
and logically, and not to be satisfied with using unarticulated criteria or unsub-
stantiated empirical judgments to arrive at answers. Moreover, accountability
makes people more likely to identify their own sources of bias, because of the
need to justify themselves to others who do not necessarily view the decision
with the same biases.
But, like other social phenomenon, the benefits of accountability depend on a
number of factors, and holding a decision maker accountable can sometimes be
counterproductive. Like other social phenomena, the benefits of accountability
depend on a number of factors, such as:
• Whether the views of the audience are known or unknown. Decision makers
who know the views of the audience to whom they are accountable tend to
conform decisions to gain the audience’s favor. For example, officials
awarding financial aid who did not need to justify their decisions to
students tended to award aid based on recipients’ needs; accountable
officials tried to please all applicants by giving something to everyone.
When the audience’s views are unknown, decision makers tend to engage
109. This section is based largely on the excellent review essay by Jennifer S. Lerner
and Philip E. Tetlock, Accounting for the Effects of Accountability, 125 Psychological
Bulletin 225 (1999).
group decision making and the effects of accountability 629
We have also devoted much attention to errors that attend intuitive decision
making—the so called “heuristics and biases” research agenda in the field of
judgment and decision making (JDM) and the insights of its offspring, behav-
ioral economics. And we have described methods for helping avoid these errors,
to the limited extent possible. Beyond the core domain of JDM, we have also
looked at the dynamics of social influence in the belief that lawyers and policy
makers are called upon both to influence others and to protect clients and citi-
zens from pernicious influence.1
How does the academic study of these matters bear on your development
of expertise as lawyers and policy makers? Expertise depends as much on intu-
ition as analysis. Yet, as the JDM and social psychology literature demonstrate,
intuition is extremely vulnerable to biases, framing, and influence. How—as the
JDM scholar Robin Hogarth asked—does one educate intuition?
The scientific method and statistics play important roles in the process. They
suggest that even when one-off intuitive judgments must be made rapidly on the
spot, they should be tested in retrospect—in aggregate, where possible—so you
can see how well you fared and learn how to improve. The scientific method and
statistics help you avoid some of the recurrent errors that we have examined.
They press us to seek disconfirming evidence: a contingency table requires you
to fill in the cells where there is no result. Statistics counter the availability and
representativeness heuristics by treating sample size as an essential determinant
of statistical significance. It counters overconfidence by requiring that statistical
hypothesis testing state the acceptable margin of error.
We believe that developing the systematic habits of thought inherent in delib-
erative decision making improves subsequent problem solving done at the intu-
itive end of the spectrum, or at least facilitates reflective monitoring of intuitive
judgments. We find a compelling parallel in Constantin Stanislavski’s descrip-
tion of an actor’s preparation:2
One cannot always create subconsciously and with inspiration. No such genius
exists in the world. Therefore our art teaches us first of all to create consciously and
rightly, because that will best prepare the way for the blossoming of the subcon-
scious, which is inspiration. The more you have of conscious creative moments in
your role, the more chance you will have of a flow of inspiration.
Gary Blasi, from whom we borrowed our protagonist, Luis Trujillo, says that he
“has acquired a significant body of knowledge—about opposing lawyers, about
trial judges, about the likely consequences of certain actions—from his many
previous interactions with other lawyers, other judges.” It would be more accu-
rate to say that Trujillo, like any other professional, has had the opportunity to
gain this knowledge. For example, while Trujillo’s observation (in Chapter 1)
about the judge’s reception to motions for summary judgment may well be cor-
rect, it may also be based on a few vivid personal experiences or on settled
wisdom with little empirical foundation.
While experience is inevitable, learning from experience is not. The last item
in our deliberative checklist in Section 1.4.1 included monitoring the outcome of
the decision in order to learn from it. As Gary Klein writes, experts learn from
experience in several ways:
• They engage in reflective practice, so that each opportunity for practice has
a goal and evaluation criterion.
• They compile an extensive experience bank.
• They obtain feedback that is accurate, diagnostic, and reasonably timely.
• They review experiences to derive new insights and learn from mistakes.
Personal characteristics play a role as well. Mark Kelman writes:
People who might be described as “open-minded” rather than dogmatic, especially
in the sense that they accept the possibility that their own thinking is fallible and
that they feel obliged to evaluate the quality of arguments without much regard to
their predispositions about how an issue ought to be resolved, are less prone to
make many of the errors that . . . researchers have identified as “biases.” If one
wants to get a quick, intuitive feel for the sorts of dispositions one might be trying
to measure here, think about how two subjects might answer the questions: “People
should always take into consideration evidence that goes against their own beliefs”
and “No one can talk me out of something that I know is right.”3
The ability to learn from experience depends not only on the expert’s personal
qualities and stance toward learning, but on the learning environment in which
he or she works. Learning requires reliable feedback and requires using that
feedback advertently. Robin Hogarth distinguishes kind learning structures, in
which people receive good feedback, from wicked environments, in which feed-
back can be misleading.4 Determinants of the quality of the learning environ-
ment include the ratio of relevant to irrelevant or random feedback, and how
exacting or lenient the system is with respect to incorrect judgments.5 With
respect to the speed and systematicity of feedback, consider the difference
3. Mark G. Kelman, The Heuristics Debate: Its Nature and Its Implications for
Law and Policy (Oxford University Press, forthcoming 2010), citing Keith Stanovich,
Who Is Rational? (Mahwah, NJ: Lawrence Erlbaum 1999).
4. Hogarth, supra at 98.
5. Id. 88–89.
634 problem solving, decision making, and professional judgment
6. Mark Twain, Life on the Mississippi 57–64 (New York: Penguin Classics, 1883:
1962).
conclusion 635
you mustn’t get the shoal soundings and marks of one trip mixed up with the
shoal soundings and marks of another, either, for they’re not often twice
alike. You must keep them separate.” . . .
[Twain sees what he believes to be a dangerous reef, and Bixby orders him
to run over it.]
As it disappeared under our bows, I held my breath; but we slid over it like oil.
“Now don’t you see the difference? It wasn’t anything but a wind reef. The
wind does that.”
“So I see. But it is exactly like a bluff reef. How am I ever going to tell them
apart?”
“I can’t tell you. It is an instinct. By and by you will just naturally know one
from the other, but you never will be able to explain why or how you know
them apart.”
It turned out to be true. The fact of the water, in time, became a wonderful
book—a book that was a dead language to the uneducated passenger, but
which told its mind to me without reserve, delivering its most cherished
secrets as clear as if it uttered them with a voice.
7. Emily Pronin, Thomas Gilovich, and Lee Ross, Objectivity in the Eye of the Beholder:
Divergent Perceptions of Bias in Self Versus Others, 111 Psychological Review 781–99 (2004).
636 problem solving, decision making, and professional judgment
In concluding the book, let us review some common ways in which biases
hinder learning from experience.
Confirmation biases. As we saw in Section 10.3, people tend to seek out evi-
dence that supports their existing (or preferred) hypothesis and avoid looking for
evidence that contradicts it. Closely related is the biased assimilation of evidence—
our tendency to view evidence that favors our beliefs more favorably and less
critically than evidence that challenges them. Belief perseverance is a cousin of
these phenomena. Once we have formed an initial belief about any fact or value,
we tend to process new information in light of our belief. If we have accumulated
a network of supporting beliefs, it can become extremely difficult to reverse
course, even when new evidence suggests that it would be reasonable to do so.
What Jerome Bruner and his colleagues have termed the “thirst for confirm-
ing redundancy”8 has motivational as well as cognitive roots. We get invested—-
sometimes literally—in our own ideas and practice. The possibility of their being
wrong can be a blow to our pocketbooks as well as our egos and reputations.
Poor feedback mechanisms. Learning from mistakes depends on feedback.
Sometimes it is difficult to get feedback just because the data are not available,
or not available quickly or in sufficient quantity or in a form conducive to learn-
ing. We are also prone to distortions in acquiring data—because of stress, sche-
mas, and expectations—and to distortions in retention and retrieval—because of
intervening events or just plain forgetting. We tend not to notice omissions (the
dog that didn’t bark in the nighttime). If these are largely cognitive phenomena,
there’s a strong motivational one as well: avoiding feedback to minimize regret.
Under- and over-generalization. Data are often presented in ways that make it
difficult to generalize, to connect the dots. On the flip side, we readily overgener-
alize, seeing relationships where none exist. This results from basic statistical
errors—misunderstandings of probability, randomness, independence, sample
size, regression to the mean—exacerbated by the illusion of control and the
availability heuristic.
The flipside of the illusion of control is the self-fulfilling prophecy, where our
empirical judgment actually affects the outcome. An example is the “Pygmalion”
effect, where students who are predicted to perform better actually do so. There’s
a story about a nineteenth-century physician who didn’t wash his hands between
examining patients for a particular infectious disease and was therefore very
good in predicting that the next patient he examined would have the disease.
Hindsight bias—the erroneous belief that we “knew it all along”— is also a
barrier to learning from experience. Because many decisions are made in condi-
tions of uncertainty about the outcomes, and because improving our
The overarching hypothesis of this book is that academic study can lay the
foundation for developing expertise in problem solving, decision making, and
professional judgment on the job. If many aspects of this hypothesis remain to be
tested, it nonetheless seems like a pretty good bet. At very least we hope that you
have learned something interesting about human behavior, including your own.
9. Lee Ross, The Intuitive Psychologist and His Shortcomings: Distortions in the Attribution
Process, in L. Berkowitz (ed.), Advances in Experimental Social Psychology (Vol. 10)
173–220. (New York: Academic Press, 1977).
This page intentionally left blank
index
Figures, notes, and tables are indicated by f, n, and t following the page numbers.
A defined, 19n26
Abdul Latif Jameel Poverty Action dread, discounting for, 521
Lab (J-PAL), 198, 210 evaluability and framing, 511–12
Above-average effect, 408 exchange principle, 508–11
Absolute vs. relative wealth and familiarity, 499
happiness, 360 feelings, 491–94, 491f
Abu Ghraib prison, 539, 553–54 high anxiety, 498–99
Accompanied groups, 444 insurance as protection against
Accountability, effects of, 628–29 worry, 515–16
Ackerman, Frank, 383 judgments of risk and benefit, 502–3
Action biases, 500–501 maximin principle, 508–11
Action vs. inaction, 430–33 mitigating factors, 497t
Actor-observer effect, 332 natural origins, 500
Adams, James, 87–88 outrage, 498–99
Adaptation and “ordinization” perceptions, cultural
neglect, 400–401 determinants of, 503–5
Ad hoc construction of precautionary principles, 521
preferences, 386–88 probability neglect, 514–19
Adjustment, 216. See also Anchoring, rational decision making, 367–68
adjustment, and activation “robust satisficing,” 508–11
Adorno, Theodor, 550 single action bias, 515
Affect, 491–522 uncertainty vs. risk, 505–11
acceptability of risks, 498–99 valuation by calculation or feeling, 513
action biases, 500–501 vividness, 495–97
affect heuristic, 367–68 voluntariness, 499
aggravating factors, 497t worry, 514–19
ambiguity vs. risk, 505–11 Affect heuristic, 367–68
anchoring and risk, 519 Affection, 347–48
anxiety, 493–94 Affective forecasting, 398
assessing risks, 520 Agentive language, 575–76
availability heuristic, 495–97 Aggravating factors, 497t
betrayal aversion, 502–3 Aggregation of views without group
catastrophic harm, 521–22 process, 597–600
certainty equivalent of lottery, 508f Agreement effect, 283
cognitive and affective processing Ainslie, George, 415
of probabilities, 513–14 Alhakami, Ali, 503
continuity principle, 508–11 Allais, Maurice, 489
controllability, 500 Allais’ paradox, 489, 490t
cost-benefit analysis (CBA), 520–22 Allhoff, Fritz, 379
cultural cognition, 504 Alter, Adam, 293
640 index
Lawyers (cont.) M
as experts, 29–31 Made to Stick: Why Some Ideas
framing problems, 40–42 Survive and Others Die
problem-solving (Heath & Heath), 571–72
processes by, 4–7, 29–31 Magat, Wesley, 585
“Leaky” decisions, 349–54 Maheswaran, Ravi, 149, 150
Leavitt, Harold J., 595–96 Malfeasance, social proof and, 545–46
Lee, Leonard, 352 Malmfors, Torbjörn, 517
Legal frames, 40n9 Malouff, John, 269
Leibold, Jill, 326, 328 Mandated choice, 594
Lepper, Mark, 246–48, 282, 287, Marginal probabilities, 129, 129t, 211
289, 446–48 Marginal utility, declining, 360–61
Lerner, Jennifer, 629 Margolis, Howard, 503
Leventhal, Robert, 581 Marketplace of ideas, 72
Levitt, Steven, 181 Markus, Hazel, 366, 447
Levy, Becca, 330 Maturation, 403
Liberman, Varda, 363 MAUT (Multi-Attribute
Libertarian paternalism, 588, 592–94 Utility Theory), 102
Lieppe, Michael, 556 Maximin principle, 508–11
Life on the Mississippi (Twain), 26, 634–35 McCaffrey, Edward, 425
Likelihood and likelihood ratio, 226, 230 McCauley, Clark, 622
Liking, 255–56, 562–66 McConnell, Allen, 326, 328
Line of best fit, 166–69 McCord, Joan, 197–98
Linguistic form, 575–81 McGarity, Thomas O., 112
Litigation Risk Analysis, xxvii McMakin, John, 394
Litigation vs. settlement, 477–80 Mean, 153, 170–72
Loewenstein, George Measurement reliability, 189
decoupling decisions and Measurement validity, 190
consequences, 439 Median, 154
memory and anticipation, Medvec, Victoria, 429, 432–33
utility from, 354–56 Meehl, Paul, 296
projection bias, 402–4 Memory. See Perception and
risks as feelings, 491–92, 494 memory, biases in
self-control, habits, and routines, 416 Menkel-Meadow, Carrie, 67–68, 73
utility, sources of, 346 Mental accounting, 437–42
visceral factors, 414 Mentorship, 634–35
Loftus, Elizabeth, 242, 244, 245, 250–51 Mere exposure effect, 255
Logic models, 58f, 59 Meta-analysis, 205–6
Long-run preferences, 403 Metacognition, 252–60
Loomis, Graham, 519 fluency as source of truth and liking
Lopez, David, 280–81 judgments, 255–56
Lopez, Gerald, 8 judgment biases, correction of, 256–57
Lord, Charles, 282, 289 mere exposure effect, 255–56
Loss aversion, 419, 422, 476 processing fluency, 255
Losses, integration of, 438 recall, 254
Lott, John R., 181 truth and liking judgments, 255–56
Lottery, certainty equivalent of, 508f Metaphors, 579–81
Low-ball technique, 557–58 Meyerowitz, Judith, 408
Luban, David, 530 Milgram, Stanley, 543, 550–56
index 655
Present value. See Net present value affect, role of, 19–20
Preston, Elizabeth, 289 recognition-primed decision
Prevention behaviors, 589 making, 16–17
Primary emotions, 354 schemas and scripts, role of, 17–19
Priming, 315–16, 581–84 by lawyers, 4–7, 29–31
The Principles of Morals and overview, 3–4
Legislation (Bentham), 346 path constraints, 10
Prior decisions and finality, 437 by policy makers, 7–8, 29–31
Prior expectancies and memory problem space, 9–10, 10f, 11f
retrieval, 322–23. See also Expectancies, schemas, defined, 14, 18n20
effects of scripts, defined, 18
Prior probability, 226, 277. See also sensemaking, 19
Base rate Procedural rationality, 368–69
Probabilistic sampling, 608–9 Processing and judging information,
Probability. See Statistics and probability biases in, 267–302. See also Bias
Probability axiom, 460 anchoring, adjustment, and
Probability neglect, 514–19 activation, 267–72
Probability tables, 211, 212t, 228t biases in, 267–302. See also Bias
Problem-solving, 3–31. See also calibration, 291
Decision-making processes confirmation, 277–89
affect, defined, 19n26 hindsight, 272–77
creativity in, 61–90. See also intuition, analysis, and
Creativity in problem solving experts’ role, 293–301
defined, 8–11 intuition, overcoming, 293
deliberative, 11–13 overconfidence, 289–93
alternatives, 13 prior expectations and theories,
causes, diagnosis of, 12–13 resistance to modification of, 277–89
course of action, choice of, 13 Procrastination, 414
framing problems, 12. See also Project determinants, 559
Frames and framing Projection bias, 401–4
interests and objectives, Pronin, Emily, 635
identification of, 12. See also Propulsion theory, 78
Interests and objectives, Prosecutor’s fallacy, 235–38. See also
identification of Confusion of the inverse
thinking, divergent and Prospect theory, 208, 419–28, 420f, 476–87
convergent, 13–14, 15f Prototypes. See Schemas; Stereotypes and
emotional intelligence, 20 stereotyping
expertise, nature of, 25–29, 31 Pseudocertainty effect, 485t, 486
experts, 29–31 Psychological determinants, 560
intuition and deliberation, Psychological health,
interaction of, 21–25 qualifying effects, 381
bounded rationality, 22–24 Psychological immune
deliberation, limits of, 22–24 system, 359, 364, 400
necessity and limits of intuition, 25 Psychological ownership, 423
quasi-rational thought, 23 Psychophysical numbing, 361–62
satisficing, 22, 385 Public policy and multiple
two-systems model of regression, 181–82
cognition, 21–22, 21t p-value, 146, 164
intuitive, 14–20 Pygmalion Effect, 190
658 index
Q Reason-based decision
Quality-adjusted life years (QALY), 379–81 strategies, 99–102, 390–95
Quantitative decision strategies, 102 Reb, Jochen, 501
Quantitative variables, 153–84. See also Reciprocity, 563–64, 566–68
Variables Recognition-primed
Quasi-rational thought, 23 decision making, 16–17, 298–99
Quattronel, George, 574–75 Recollected experience, 356–59
Quinn, Diane, 330 Recuperative behaviors, 589
Redelmeier, Donald E., 446
R Reference points, 419, 476
Rachlinski, Jeffrey Refinement, 403
anchoring in negotiations, 269 Reflexive controls, 193–94
hindsight bias, 272, 274, 276–77 Regan, Dennis T., 566
litigation vs. settlement, 478 Regression
settlement offers, 527 analysis of variance equation, 176
targets and risk-seeking, 481 chance and, 172
uncertainty, 505 coefficient of determination, 176
Racial bias, 304–30 correlation, 169–70, 171f, 177f
behavioral confirmation, 326–27 employment discrimination, 179–80
juvenile probation, 324 equations, 168
weapon, perception of possession, 319 hypothesis testing, 178–79
Randomized controlled studies, 191–92 independence, 165
Randomness, 123–24, 136–41. See also intercepts, 168
Statistics and probability intuitions about, 172–75
Random samples, 132 linear, 179
Random variables, 123–24 line of best fit, 166–69
Ranked groups, 444 to mean, 170–72
Rational decision making, 366–73 multiple, 179–81
affect heuristic, 367–68 nonlinear relationships, 167
completeness, dominance, and predictors, explanations by, 175–77
trade-offs, 369–72 public policy and multiple
consequential rationality, 368 regression, 181–82
constitutive incommensurability, 370 representativeness heuristic, 173
corporate cost-benefit analysis involving scatterplots, 166, 166f
human safety, 371 simple linear, 179
decomposability, 372 statistical significance, 178–79
ecological rationality, 373 variables, 165–82
invariance, 372 variance around line of best fit, 175
irrelevant alternatives, variance around mean, 175
independence from, 372 winner’s curse, 174, 174f
parsimoniousness, 372 Regret
procedural rationality, 368–69 aversion, 428–34
subjective expected avoidance, 427, 436
utility theory, 368–73 phenomenon, 354
transitivity, 372–73 suboptimal learning processes and, 434
Rationale construction, 616 Reich, Charles, 71–72
Rationalization, 364 Reichenbach, Hans, 119
Reactive egoism, 262–63 Relative wealth and happiness, 360
Reality checks, 109–10 Reliability, 300
index 659