Math With Bad Drawings
Math With Bad Drawings
Hachette Book Group supports the right to free expression and the value of
copyright. The purpose of copyright is to encourage writers and artists to
produce the creative works that enrich our culture.
The publisher is not responsible for websites (or their content) that are not
owned by the publisher.
The Hachette Speakers Bureau provides a wide range of authors for speaking
events. To find out more, go to www.HachetteSpeakersBureau.com or call (866)
376-6591.
COVER
TITLE PAGE
COPYRIGHT
DEDICATION
INTRODUCTION
I
HOW TO THINK LIKE A MATHEMATICIAN
1. Ultimate Tic-Tac-Toe
2. What Does Math Look Like to Students?
3. What Does Math Look Like to Mathematicians?
4. How Science and Math See Each Other
5. Good Mathematician vs. Great Mathematician
II
DESIGN: THE GEOMETRY OF STUFF THAT WORKS
III
PROBABILITY: THE MATHEMATICS OF MAYBE
IV
STATISTICS: THE FINE ART OF HONEST LYING
V
ON THE CUSP: THE POWER OF A STEP
ACKNOWLEDGMENTS
ENDNOTES
NEWSLETTERS
For Taryn
INTRODUCTION
Lots of days in the classroom, I strike out. Math feels to my students like a
musty basement where meaningless symbols shuffle back and forth. The kids
shrug, learn the choreography, and dance the tuneless dance.
But on livelier days, they see distant points of light and realize that the
basement is a secret tunnel, linking everything they know to just about
everything else. The students grapple and innovate, drawing connections, taking
leaps, building that elusive virtue called “understanding.”
Unlike in the classroom, this book will sidestep the technical details.You’ll
find few equations on these pages, and the spookiest ones are decorative
anyway. (The hard core can seek elaboration in the endnotes.) Instead, I want to
focus on what I see as the true heart of mathematics: the concepts. Each section
of this book will tour a variety of landscapes, all sharing the underground
network of a single big idea: How the rules of geometry constrain our design
choices. How the methods of probability tap the liquor of eternity. How tiny
increments yield quantum jumps. How statistics make legible the mad sprawl of
reality.
Writing this book has brought me to places I didn’t expect. I hope that
reading it does the same for you.
—BEN ORLIN, October 2017
I
HOW TO THINK LIKE A MATHEMATICIAN
WHAT IS MATHEMATICS?
You can make the case that all creative endeavors are about pushing against
constraints. In the words of physicist Richard Feynman, “Creativity is
imagination in a straitjacket.” Take the sonnet, whose tight formal restrictions—
Follow this rhythm! Adhere to this length! Make sure these words rhyme!
Okay… now express your love, lil’ Shakespeare!—don’t undercut the artistry but
heighten it. Or look at sports. Humans strain to achieve goals (kick the ball in the
net) while obeying rigid limitations (don’t use your hands). In the process, they
create bicycle kicks and diving headers. If you ditch the rulebook, you lose the
grace. Even the wacky, avant-garde, convention-defying arts—experimental
film, expressionist painting, professional wrestling—draw their power from
playing against the limitations of the chosen medium.
Creativity is what happens when a mind encounters an obstacle. It’s the
human process of finding a way through, over, around, or beneath. No obstacle,
no creativity.
But mathematics takes this concept one step further. In math, we don’t just
follow rules. We invent them. We tweak them. We propose a possible constraint,
play out its logical consequences, and then, if that way leads to oblivion—or
worse, to boredom—we seek a new and more fruitful path.
For example, what happens if I challenge one little assumption about parallel
lines?
Euclid laid out this rule regarding parallel lines in 300 BCE; he took it for
granted, calling it a fundamental assumption (a “postulate”). This struck his
successors as a bit funny. Do you really have to assume it? Shouldn’t it be
provable? For two millennia, scholars poked and prodded at the rule, like a piece
of food caught between their teeth. At last they realized: Oh! It is an assumption.
You can assume otherwise. And if you do, traditional geometry collapses,
revealing strange alternative geometries, where the words “parallel” and “line”
come to mean something else entirely.
New rule, new game.
As it turns out, the same holds true of Ultimate Tic-Tac-Toe. Soon after I
began sharing the game around, I learned that there’s a single technicality upon
which everything hinges. It comes down to a question I confined to a
parenthetical earlier: What happens if my opponent sends me to a mini-board
that’s already been won?
These days, my answer is the one I gave above: Since that mini-board is
already “closed,” you can go wherever you want.
But originally, I had a different answer: As long as there’s an empty space on
that mini-board, you have to go there—even though it’s a wasted move.
This sounds minor—just a single thread in the tapestry of the game. But
watch how it all unravels when we give this thread a pull.
I illustrated the nature of the original rule with an opening strategy that I
dubbed (in a triumph of modesty) “The Orlin Gambit.” It goes like this:
When all is said and done, X has sacrificed the center mini-board in exchange
for superior position on the other eight. I considered this stratagem pretty clever
until readers pointed out its profound uncleverness. The Orlin Gambit wasn’t
just a minor advantage. It could be extended into a guaranteed winning strategy.
Rather than sacrifice one mini-board, you can sacrifice two, giving you two-in-a-
rows on each of the other seven. From there, you can ensure victory in just a
handful of moves.
Embarrassed, I updated my explanation to give the rule in its current version
—a small but crucial tweak that restored Ultimate Tic-Tac-Toe to health.
New rule, new game.
This is exactly how mathematics proceeds. You throw down some rules and
begin to play. When the game grows stale, you change it. You pose new
constraints. You relax old ones. Each tweak creates new puzzles, fresh
challenges. Most mathematics is less about solving someone else’s riddles than
about devising your own, exploring which constraints produce interesting games
and which produce drab ones. Eventually, this process of rule tweaking, of
moving from game to game, comes to feel like a grand never-ending game in
itself.
Mathematics is the logic game of inventing logic games.
The history of mathematics is the unfolding of this story, time and time
again. Logic games are invented, solved, and reinvented. For example: what
happens if I tweak this simple equation, changing the exponents from 2 into
another number, like 3, or 5, or 797?
Alas, this will be a short, bleak chapter. I’d apologize for that, but I’ll be too
busy apologizing for other things, like the often soul-pureeing experience of
math education.
You know what I’m talking about. To many students, “doing math” means
executing the prescribed sequence of pencil maneuvers. Mathematical symbols
don’t symbolize; they just dance across the page in baffling choreography. Math
is a tale told by an abacus, full of “sin” and “θ,” signifying nothing.
Allow me to offer two brief apologies:
First, I apologize to my own students for any times I’ve made you feel this
way about math. I’ve tried not to; then again, I’ve also tried to stay on top of
email, cut back on ice cream, and avoid four-month lapses between haircuts.
Please enter my plea as “merely human.”
Second, I apologize to mathematics, for any violence you have suffered at my
hands. In my defense, you’re an intangible tower of quantitative concepts bound
together by abstract logic, so I doubt I’ve left any lasting scars. But I’m not too
proud to say I’m sorry.
That’s it for this chapter. I promise the next one will have more explosions,
like any good sequel should.
Chapter 3
WHAT DOES MATH LOOK LIKE TO MATHEMATICIANS?
A few patterns are already peeking out. But our vision is only as powerful as
our technology, and tables are not a fancy gadget. Of the infinite x-y pairs that
satisfy this equation, the table can show only a handful at a time, like the
scrolling text of a stock ticker. We need a superior visualization tool: the
mathematical equivalent of a television screen.
Enter the graph.
By treating x and y as a sort of longitude and latitude, we transform each
intangible pair of numbers into something satisfyingly geometric: a point. The
infinite set of points becomes a unified curve. And from this, a story emerges, a
tale of motion and change.
1. TWINS NO MORE
Back in 9th grade, my friend John and I looked oddly similar: pensive round-
faced brown-haired boys who spoke about as often as furniture. Teachers called
us by the wrong names; older students thought we were one person; and in the
yearbook, we’re mislabeled as each other. For all I remember, we sparked a
friendship just to mess with people.
Then, with time, we thinned out and grew up. John is now a broad-chested
six foot two and looks like a Disney prince. I’m five nine and have been
described as “a cross between Harry Potter and Daniel Radcliffe.” The
“practically twins” stage of our friendship is long gone.
And as it is with me and John, so it is with mathematics and science.
Back when science and math still had their baby fat, they didn’t just look the
same. They were the same. Isaac Newton didn’t worry whether history would
classify him as a scientist or a mathematician: he was both, inextricably. The
same went for his intellectual elder siblings—Galileo, Kepler, Copernicus. For
them, science and math were interwoven, inseparable. Their core insight was
that the physical cosmos followed mathematical recipes. Objects obeyed
equations. You couldn’t study one without the other, any more than you could
eat a single ingredient out of a baked cake.
Science and math have diverged since then. Just look at how they’re taught:
separate classrooms, separate teachers, separate (if equally mind-numbing)
textbooks. They’ve thinned out, added muscle, aged beyond their wide-eyed 9th-
grade innocence.
But still, people get them confused. Any fool can see that I’m not John, and
John isn’t me, but the citizen on the street has more trouble telling the
differences between math and science—especially at a layman’s distance.
Perhaps the easiest way to tell them apart is to answer this question: What do
science and math look like, not to the layperson, but to each other?
Here, then, is a brief biography of knot theory. It was born from a practical
need. Soon, it grew into something deliberately impractical, a logic game for
poets and philosophers. And yet somehow this mature creature, which had
barely spared a thought for reality over the years, became profoundly useful in a
field far removed from the one of its birth.
This is no isolated case. It’s a basic pattern of mathematical history.
Remember the strange alternate geometries from Chapter 1? For centuries,
scholars saw them as mere novelties, a poet’s fancy. They bore no
correspondence to reality, which was understood to follow Euclid’s assumption
about parallel lines.
Then, along came a young patent clerk named Einstein. He realized that these
wacko alternatives aren’t just thought experiments; they underlie the structure of
the cosmos. From our tiny perspective, the universe appears Euclidean, much in
the way that the curved Earth appears flat. But zoom out, shedding the
prejudices of the surface dweller, and you’ll see a different picture entirely, a
shifting landscape of strange curvatures.
Those “useless” geometries turned out to be pretty darn useful.
Perhaps my favorite example concerns logic itself. Early philosophers like
Aristotle developed symbolic logic (“If p, then q”) as a guide for scientific
thinking. Then mathematical theorists got their hands on it, and turned logic into
something bizarre and abstract. Reality fell away. By the 20th century, you had
folks like Bertrand Russell writing Latin-titled tomes that aimed to “prove” from
elementary assumptions that 1 + 1 = 2. What could be more useless, more
irredeemable?
One logician’s mother nagged him: C’mon, honey, what’s the point of all this
abstract mathematics? Why not do something useful?
That mother was named Ethel Turing. And as it turns out, her son Alan was
kind of on to something: a logic machine we now call “the computer.”
I can’t blame her for doubting. Who would have guessed that her son’s
abstract research into logical systems would help to define the next century? No
matter how many examples I encounter, this historical cycle of useful to useless
to useful again remains a wonder and a mystery to me.
My favorite description of this phenomenon is a phrase coined by physicist
Eugene Wigner: “the unreasonable effectiveness of mathematics.” After all,
bacteria don’t know any knot theory, so why should they follow its rules? The
space-time continuum hasn’t studied hyperbolic geometry, so why does it
execute its theorems so perfectly? I’ve read some philosophers’ answers to this
question, but I find them tentative and conflicting, and none have worked to
blunt my astonishment.
So, how best to understand the relationship between the poet we call Math
and the adventurer known as Science? Perhaps we ought to see them as a
symbiotic pair of very different creatures, like an insect-eating bird perched on
the back of a rhino. The rhino gets its itchy problems solved. The bird gets
nourished. Both emerge happy.
When you visualize math, picture something dainty and elegant astride the
wrinkled gray mass of reality below.
Chapter 5
GOOD MATHEMATICIAN VS. GREAT MATHEMATICIAN
It’s a lot of fun to bust myths. Just look at the carefree explosions and ear-to-ear
smiles of television’s MythBusters, and you can see it’s a career with high job
satisfaction.
What’s trickier is tweaking myths. Lots of the culture’s prevailing views
about mathematics aren’t flat-out wrong—just crooked, or incomplete, or
overemphasized. Is computation important? Sure, but it’s not all-important. Does
math require attention to detail? Yes, but so do knitting and parkour. Was Carl
Gauss a natural genius? Well, yeah, but most beautiful math comes not from
depressive German perfectionists but from ordinary folks like you and me.
Before we close this section, this chapter offers one final exploration of how
to think like a mathematician, a chance to revise and annotate some popular
myths. Like most myths, they’ve got a basis in truth. And, like most myths, they
miss the flux, uncertainty, and struggle toward understanding that make us
human—and that make us mathematicians.
A few years ago, when I lived in England, I taught a boy named Corey. He
reminded me of a soft-spoken 12-year-old Benjamin Franklin: quiet and
insightful, with long ginger hair and round spectacles. I can totally picture him
inventing bifocals.
Corey poured his heart into every homework assignment, drew lucid
connections across topics, and packed up his papers at the period’s end with such
care and patience that I always fretted he’d be late for his next lesson. So it’s no
surprise that on the first big test in November, Corey nailed every question.
Well… every question that he’d had time to answer.
The bell rang with the last quarter of his test still blank. He scored in the low
70s and came to me the next day with a furrowed brow. “Sir,” he said, because
England is an amazing land where clumsy 29-year-old teachers get fancy
honorifics, “why are tests timed?”
I figure honesty is the best policy. “It’s not because speed is so important. We
just want to see what students can do by themselves, without anyone’s help.”
“So why not let us keep working?”
“Well, if I held the class hostage for a whole day, it might annoy your other
teachers. They want you to know about science and geography, because of their
nostalgic attachment to reality.”
I realized that I had never seen Corey like this: jaw clenched, eyes dark. He
was radiating frustration. “I could have answered more,” he said. “I just ran out
of time.”
I nodded. “I know.”
There wasn’t much else to say.
Intentionally or not, school mathematics sends a loud, clear message: Speed
is everything. Tests are timed. Early finishers get to start their homework. Just
look how periods end: with a ringing bell, as if you’ve just finished a round in a
perverse, compulsory, logarithm-themed game show. Math comes to feel like a
race, and success becomes synonymous with quickness.
All of which is supremely silly.
Speed has one fabulous advantage: It saves time. Beyond that, mathematics is
about deep insight, real understanding, and elegant approaches, none of which
you’re likely to find when moving at 600 miles per hour. You learn more
mathematics by thinking carefully than by thinking fast, just as you learn more
botany by studying a blade of grass than by sprinting like the dickens through a
wheat field.
Corey understands this. I only hope that teachers like me don’t manage,
against our own best intentions, to persuade him otherwise.
My wife, who is a research mathematician, once pointed me toward a funny
pattern in mathematical life.
• Step #1: There’s a tricky and exciting question on the loose, an important
conjecture in need of proof. Many try to tame the beast, without success.
• Step #2: Someone finally proves it, via a long and convoluted argument
that’s full of insight but very difficult to follow.
• Step #3: Over time, new proofs are published, growing shorter and
simpler, shorter and simpler, until eventually the original proof is relegated
to “historical” status: an inefficient Edison-era lightbulb made obsolete by
sleeker, more modern designs.
I doubt you’ve heard of Ngô B u Châu, unless (a) You got bored one day and
decided to memorize every winner of the Fields Medal, math’s most prestigious
prize, or (b) You’re from Vietnam, where Châu is a national celebrity and
household name.
(By the way, huge kudos to Vietnam for that. America’s closest
approximation to a celebrity mathematician is Will Hunting: not even our most
famous Matt Damon character.)
Châu is a fierce competitor, who craved as a youngster to be the best of the
best. For a time, he was. Winning back-to-back gold medals at the International
Mathematical Olympiad, he became the pride of his school and the envy of his
peers, the Simone Biles of Vietnamese math.
But his undergraduate years brought a slow quicksand agony, as he gradually
realized that he didn’t understand the math he was learning. “My professors
thought I was a fantastic student,” he remembers. “I could do all the exercises. I
did well on exams. But I didn’t understand anything.” Achievement came to feel
like hollow orb: a brittle shell of accolades, soon to shatter and reveal the
horrible vacuum inside.
Then came a turning point. He stopped trying to be the best, and began to
learn from them.
He shovels credit on his PhD adviser Gérard Laumon. “I had one of the best
advisers in the world,” Châu says, glowing. “I would come to his office every
week. He would read with me one or two pages every time.” They went line by
line, equation by equation, settling for nothing less than full comprehension.
Châu soon began work on the famous Langlands program. Think of it as the
transcontinental railroad of modern mathematics: a sweeping vision for how to
connect several distant branches of the discipline. The project has drawn
generations of ambitious mathematicians like Châu into its orbit, and Châu
found himself attracted to a particularly vexing piece of the Langlands program:
proving the “fundamental lemma.”
Was it the Olympiad all over again? Rival mathematicians jockeying for
primacy, racing to be the first to prove it?
No, says Châu.
“I was helped a lot by people in my field,” he says. “Many people
encouraged me to do so, in a very sincere way. I asked them for advice, and they
would tell me what to learn. It was very open. I did not feel competition.” With
help from these collaborators, Châu managed to prove the fundamental lemma.
It was the work than won him his Fields.
For such an exceptional scholar, Châu’s tale is pleasantly ordinary. People
who thrive in the tournament atmosphere of school—with its clear rankings,
easy lateral comparisons, and steady diet of rewards—find they need a new
attitude as they advance into the open-ended world of scholarship. Drawn in as
competitors, they evolve into collaborators.
II
DESIGN
Here’s a tough-love life lesson: just because you set your mind to something
doesn’t mean you can do it.
Say you’re creating a square, and you want its diagonal to be the same length
as its sides. Well, I hate to break it to you, but that’s not how squares work.
Or say you’re tiling a floor, and you want to use a pattern of regular
pentagons. I’m sorry to report that this will never pan out. There will always be
gaps between your tiles.
Or say you’re building an equilateral triangle, and you fancy a bespoke angle
size—say 70°, or 49°, or 123°. I’m afraid the cosmos doesn’t care: Those angles
are going to be 60°, no more and no less, or else your triangle won’t be
equilateral.
Human laws are flexible, subject to repeal and renegotiation. The drinking
age is 21 in the US, 18 in Britain, 16 in Cuba, “never” in Afghanistan, and “step
right up” in Cambodia. Any country can stiffen or relax its drinking laws at a
whim (and of course, the stiffer your drink and the laxer your laws, the more
whim-prone you become). Geometry’s laws are not like that. There’s no wiggle
room: no president to issue pardons, no jury to acquit, no officer to let you off
with a warning. Math’s rules are self-enforcing, unbreakable by their nature.
Yet as we’ve seen, and shall see many times again, that’s not a bad thing.
Restrictions birth creativity. The laws on what shapes can’t do come packaged
with case studies illuminating what they can. In design projects ranging from
sturdy buildings to useful paper to planet-destroying space stations, geometry
inspires even as it constrains.
So forget the sentiment that “anything is possible!” It is very sweet but
deeply unnatural, like most things we feed to children. The reality is harsher—
and more wondrous.
Chapter 6
WE BUILT THIS CITY ON TRIANGLES
It’s not your typical protagonist. Snooty literary types may dismiss it as two-
dimensional. Yet this atypical hero will embark on a typical hero’s journey:
rising from humble origins, learning to harness an inner strength, and ultimately
serving the world in a time of crisis.
Now, if your mind is so narrow that it cannot encompass the idea of a valiant
polygon, then by all means, read no further. Don your blindfold of prejudice.
Just make sure to squeeze your eyelids tight, for only the deepest darkness can
guard your shuttered thoughts from the radiant truth, the penetrating light, of
planar geometry. Don’t you remember? We built this city. We built this city on
triangles…
1. TWELVE KNOTS IN AN EGYPTIAN ROPE
Welcome to ancient Egypt: a prosperous kingdom, a teeming bureaucracy, rigid
of faith and right-angled of elbow. It shall endure for millennia as the sun rises
and sets on empires with lesser headgear.
Come, let’s stroll. The year is 2570 BCE, and the Great Pyramid of Giza is
halfway built. Three and a half million tons of brick rise out of the desert, with 3
million more to follow. The heaviest blocks outweigh two bull elephants. The
square base is 756 feet per side, the length of three New York City blocks.
Already, it is the tallest structure in the world. Upon its completion a decade
from now, when its workforce of 80,000 can finally rest for lemonade, it will
stand 481 feet tall. Five millennia from today, it will still be here, history’s most
enduring skyscraper, the greatest triumph of triangular architecture.
Except it isn’t.
I mean, it is still standing, last I checked. But it is not a victory for triangles.
If you want to see the triangle at work, forget the Great Pyramid, and walk with
me to a vacant lot nearby. There, we find a small team of surveyors carrying a
peculiar loop of rope, in which they have tied 12 equally spaced knots.
What for? Just watch. After a few steps, three men each grab a knot (#1, #4,
and #8, respectively) and pull the rope tight. As if by magic, it forms a right-
angled triangle. A fourth worker marks out the right angle. They let the triangle
relax once again into a knotty loop. The scene repeats again and again, until the
whole expanse is subdivided into perfect sections of equal size.
If you’ve ever stayed awake through a geometry lesson (and even if you
haven’t), the scene may call to mind Pythagoras’s theorem. This rule states that
if you draw a square on each side of a right-angled triangle, the two smaller
areas taken together will equal the larger one. Or, in modern algebraic terms: a2
+ b2 = c2.
There is an endless roster of such triangles. For example, the sides can have
lengths 5, 12, and 13; or 7, 24, and 25; or 8, 15, and 17; or my personal favorite,
20, 99, and 101. The Egyptians wisely chose the simplest example, the triangle
with side lengths 3, 4, and 5. Hence, the 12 knots.
But this chapter is not about Pythagoras and “his” rule (which earlier
civilizations found without his help). It is about a simpler and more fundamental
property of triangles, a hidden elegance we shall soon discover. The story of the
triangle begins not in a Pythagorean temple, nor atop the Great Pyramid, but
here, in a vacant field. A slack rope transforms into a surveyor’s tool. It’s our
first hint of a power so great it will make the pyramids look crude by
comparison.
It’s a universal truth: a triangle’s longest side must be shorter than the other
two put together.
This law is self-evident to a fly walking the perimeter of the triangle. It
knows that a straight path (from A to B) will always be shorter than a detour
(from A to C to B). Thus, the two shorter sides, taken together, must be longer
than the third side.
This law has a companion, even deeper and more powerful: If three sides do
form a triangle, then they will form exactly one. Given three sides, there is no
room to embellish or improvise. There is a single mold to follow.
For example, let’s agree in advance upon three sides (such as 5 meters, 6
meters, and 7 meters), then build our personal triangles in separate rooms. I
guarantee that we will emerge with identical creations.
Watch: I’ll lay my longest side flat along the ground, then lean the other two
against each other to meet at their tips. Done! Slide the corner to the left, and one
side sticks out; to the right, and the other does. The solution is what
mathematicians call unique. Even without anticipating your method, I know
you’ll arrive at the same solution, because there is no other.
This truth belongs to the triangle alone. No other polygon can claim the
same.
Try it with the triangle’s closest cousin, the four-sided quadrilateral. I lay one
side flat. I stand up the next two vertically. And I lay my final side across the
top, fastening the corners with duct tape for good measure. But then, a wind
begins to blow. My square sways. The uprights tilt, hinging at the corners, and
the whole apparatus starts to collapse like a folding chair. Every moment brings
a new shape, from “square” to “almost square” to “kind of diamondy” to “skinny
rhombus of super pointiness.”
The four sides haven’t yielded a unique shape. Rather, they’ve given us an
infinite family of possibilities. Any member can be rearranged into another by
applying a little pressure.
Thus, we see the triangle’s hidden magic, its secret identity: not its mere
three-sided-ness, but the rigidity it confers.
The Egyptian rope pullers knew this power. By pulling their 12-knot rope
tight, they summoned a Pythagorean triangle into existence, conjuring a right
angle from rope. You might try instead to summon a square, with its four right
angles, but beware: a whole family of unwanted shapes will answer that call.
Even pulled tight, the quadrilateral will twist at its corners, shifting forms,
refusing definition. The same is true of pentagons, hexagons, heptagons, and all
the other polygon cousins. None can do what the triangle can.
The pyramids, being solid, did not take advantage of this special power.
Cubes, cones, frustums—any of these would have suited the pharaohs’ purpose
just as well. The blunt language of stone doesn’t much care what shape it
articulates.
Now, I don’t mean to slight the pyramids. For one, “slight” is not the right
word for a brick pile weighing 20 billion pounds. I admire their otherworldly
precision: sides equal to within 8 inches, facing the cardinal directions with an
error below 0.1°, with corners less than 0.01° from right-angled perfection.
Those Egyptian cats knew their math.
But I must point out that these are a surveyor’s triumphs, not an engineer’s.
The Great Pyramid remains, fundamentally, a big stack of blocks. That’s cool
for a towering symbol of your pharaoh’s immortality, but not great for a building
you’re actually hoping to, y’know, use. The pyramid’s modest chambers and
slivers of passageway occupy less than 0.1% of its internal volume. Imagine if
the Empire State Building were solid steel except for a single floor with a 2-foot
ceiling, and you too will begin aspiring to a more efficient building plan.
In centuries to come, architects would seek new poetries of structure. They
would build bridges wider than the sky, towers taller than Babel. And for this,
they would need a shape of extraordinary resilience, a shape of unique and
unyielding character: a three-sided, three-cornered hero.
1. “Outside” is a bad place to live. It can get very cold, there’s nowhere to
keep your stuff, and sometimes bears show up. That’s why humans
invented “inside.”
2. To create “inside,” you make a big empty shape and then live in the
middle of it.
3. If your shape is friendly and made of the right things, then it will be nice
to live inside and won’t fall down on you. This is called “architecture.”
Okay, now that we’re up to speed, I can introduce you to a crucial supporting
character in the triangle’s story: the beam. If you’re an architect aiming to avoid
both (a) pyramid-like monoliths and (b) collapsing floors, then beams will likely
factor in your design.
The beam’s effect is to take vertical forces and move them horizontally. For
example, picture a plank stretched over a ditch. When you stand on the beam,
your weight pushes it downward. But the real support isn’t below: it’s off to the
sides, where the plank meets the dirt. The beam takes a force in the middle and
moves its impact to the sides.
There’s just one problem: beams are inefficient.
Architecture, like life itself, is all about stress management. Whereas life
offers many types of stresses (deadlines, child-rearing, low phone battery, etc.),
structures experience just two: pushes and pulls. Pushes create compression,
when an object is squeezed tighter. Pulls create tension, when an object is
stretched apart. Each type has its own peculiar character, and different materials
cope in different ways. Concrete can withstand fantastic degrees of compression,
but it crumbles under tension. At the other extreme, steel cables can endure
incredible amounts of tension, but they buckle under the slightest compression.
Now, picture a beam sagging under a load. It curves into a smile (or, more
fitting, a grimace). What’s the nature of its strain: tension or compression?
The answer is “both.” Look at the upper surface of the beam: like a runner on
the inside lane of a track, it has a shorter distance to curve. Thus, its material
squeezes together, creating compression. Now, move your gaze to the bottom:
like a runner on the outside lane of a track, it has farther to curve, so its material
stretches and experiences tension.
So far, nothing to worry about: many materials, such as wood, cope well with
both pushes and pulls. The problem isn’t that the beam experiences both
stresses; it’s that a large fraction of the beam experiences neither.
Look to the center. Halfway between the compression on top and the tension
on bottom, the beam’s middle experiences no strain whatsoever. Its curve is the
carefree smile of a guy who’s not helping. The middle is wasted substance, no
better than the useless bulk of a pyramid. A generic beam squanders half of its
strength, like a student coasting on 50% effort.
Every teacher knows the next two words: That’s unacceptable. In
architecture, every ounce counts, whether you’re building a sky-tickling tower, a
canyon-spanning bridge, or a life-affirming roller coaster.
Rest assured: architects are no fools. They’ve got a plan.
This is a good start. But we’ve still got wasted material in the center. Thus,
we activate phase two of the architects’ plan: start punching holes in the I-beam.
Every gap in the material saves precious resources while costing us almost
nothing in strength. The emptier the space, the greater our savings, which means
our best hope is to leave the I-beam’s middle a hole-riddled web, more vacant
space than solid substance.
But hang on. Before we start punching holes willy-nilly, we need a plan.
What patterns of gaps will minimize the material required while preserving the
structure’s strength and rigidity? Where can we turn for a design that’s simple
and resilient, not to mention well suited to the flat, nearly two-dimensional realm
of the I-beam’s middle?
There is but one shape that can answer this call. The weak-willed square
twists at its corners. The cowardly pentagon collapses under pressure. And don’t
get me started on that spineless turncoat known as the hexagon. Only a
polygonal Superman can absorb the strain with a stoic and unyielding form.
Get me Triangle on the phone.
When you connect triangles into a single structural unit, you create a truss
(from the French word for “bundle”). In a truss, every member experiences
tension or compression. Trusses waste no material, like hunters using every part
of the animal.
In ancient Egypt, the triangle labored in a vacant lot, offering a nifty trick to
surveyors while the spotlight shone elsewhere. Then, millennia later and oceans
away, the triangle moved from the background scenery to center stage.
The Baltimore and Pennsylvania trusses, Pratt variants with nested triangles,
became common for rail bridges.
The Bailey truss arrived with the military needs of World War II. Its
standardized modular triangles could be dismantled, shipped, and reassembled to
meet the shifting urgencies of wartime.
Still, when I saw their stuff—which goes by the even less appealing name of
“A4”—I developed a hot and immediate loathing. It’s slightly too skinny, like a
trendy pair of jeans. I was long accustomed to a loose-fit boot-cut style, and my
whole being chafed at this svelte European nonsense. Letter paper has a length
that’s about 30% greater than its width; clearly this ratio was different. And, just
as clearly, worse.
So I went to look up its measurements. I figured it’d be 22.5 centimeters by
28 centimeters, or perhaps 23 centimeters by 30 centimeters. Something nice and
tidy for these metric-minded folk, right?
Nope. It’s 21 centimeters by 29.7 centimeters.
What the heck?
I divided 29.7 by 21 to figure out the simplified ratio: roughly 1.41. As a
math teacher, I recognized this number immediately: It’s (approximately) √2,
a.k.a. “the square root of 2.” And in that instant, my befuddlement gave way to a
snarling, smoking outrage.
√2 is irrational: a word that literally means “not a ratio.”
The paper makers had chosen a ratio that—I can’t put too fine a point on this
—is not a ratio.
We tend to run across two kinds of numbers in life: (1) whole numbers, as in
“I have 3 children,” “My children consume 5 entire boxes of cereal each
morning,” and “The stains on my children’s clothing range over 17 different
colors”; and (2) ratios of whole numbers, as in “We spend 1/4 of our
disposable income on Legos,” “Houses with children are 17½ times more likely
to be the target of Magic Marker graffiti,” and “Hey, when did 2/3 of my hair go
gray?”
(Your everyday decimals, I should point out, are just ratios in disguise. For
example, $0.71 is simply 71/100 of a dollar.)
But some wild and exotic numbers don’t fit into either category. They are not
only unwhole, but unholy: They cannot be written as ratios. (My 12-year-old
student Aadam, who has a far more brilliant mind than I, dubbed them the
“disintegers.”) No fraction or decimal that you can write down will ever quite
nail these numbers. They’ll always slip between the cracks.
And √2—the number that, when multiplied by itself, yields 2—turns out to
be just such a number. Take a look:
There’s no nice decimal that quite equals √2, and there’s no nice ratio, either.
7/5? Close. 141/100? Even closer. 665,857/470,832? So close you can almost
taste it. But never quite right. Never quite √2.
√2 isn’t merely irrational. It is, alongside π, one of the most famously
irrational numbers in all of mathematics. Legend has it that the members of the
ratio-worshipping cult of Pythagoras were so dismayed to find that √2 couldn’t
be written as a fraction that they drowned the mathematician who reported the
discovery.
If European paper aspires to √2, it has chosen a goal that it can never attain.
To put it in a language my British colleagues will understand: That’s bloody
shortsighted, innit?
For a few days, I lived in a state of heightened irritation. To touch this foolish
paper was as loathsome as touching poison ivy or underdesk gum. I joked darkly
about it, the kind of quip that’s intended as charming and world-weary but
contains such concentrated bitterness that people recoil to hear it.
And then, I realized I was wrong.
I didn’t realize it all by myself, of course. I never do. Instead, someone
pointed out to me the wonderful property that A4 paper has.
It belongs to a team.
It’s exactly double the size of A5, quadruple the size of A6, and eight times
the size of the adorably tiny A7. Meanwhile, it’s exactly half the size of A3, a
quarter the size of A2, and one-eighth the size of the impressively ginormous
A1.
As befits the individualistic culture that employs it, letter paper is an island.
Our 8½" × 11" size bears no particular relationship to the smaller and larger
paper types in circulation. It’s a one-off.
Gwen and Sven, two beautiful paper scientists, are working on a top-
secret research project. It is code-named “The Prince and the Paper,” or
perhaps “Paper Caper”—whichever sounds cooler in their vague, foreign
accents. It is late. They are exhausted but devoted to their work.
GWEN: All right, Sven. I may not know our precise nationalities, but I
know one thing: The fate of civilization depends on our ability to create a
series of papers where folding one size in half gives you the next size in
the series.
SVEN: The stakes could not be higher. But… what dimensions will such
paper have?
SVEN: Blast it, Gwen! We know that’s the same ratio, but we still don’t
know what it is.
GWEN: Okay. Let’s say that the “medium” is r times longer than the
“short.”
SVEN: But what’s r?
GWEN: I don’t know yet. I only know that it’s more than one but less
than two, because the “medium” is longer than the “short,” but it’s not
twice as long.
SVEN: All right. And I suppose that makes the “long” also r times longer
than the “medium.”
GWEN: So if you want to turn “short” into “long,” you’d have to multiply
by r (to get “medium”) and then multiply by r again. That’s r squared.
SVEN (slapping the table): You genius with perfect pores and
inexplicable high heels—Gwen, you’ve done it!
GWEN: Have I?
SVEN: The “long” is r2 times longer than the “short.” But look: it’s also
double the length of the short!
SVEN: Yes. r2 is 2.
GWEN: So r is the square root of 2! That’s the secret ratio that will end
suffering and unify humanity!
SVEN (suddenly changing accents): All right, Gwen. Hand over the ratio.
Contrary to initial appearances, the makers of A4 paper didn’t choose this ratio
to upset me personally. Nor did they choose it in deference to arbitrary fashions,
or in obstinate defiance of American hegemony, or even for the sadistic pleasure
of selecting an irrational ratio.
In fact, they didn’t really choose it at all.
What they chose was to create a system of paper where folding one size in
half yields the next size in the series. That’s a pretty cool and hard-to-fault
feature. But once they committed to this path, the decision was out of their
hands. There is only one ratio that achieves this property, and it just so happens
to be the famously irrational √2.
Now, I know we all like to imagine paper designers as unshackled fantasists,
bound only by the limits of their imagination. The reality is actually more
interesting. Designers move through a space of possibilities that are governed by
logic and geometry. This landscape has immovable features: some numbers are
rational, some aren’t, and there’s nothing a designer can do about it. Instead, the
designer must navigate around these obstacles—or, better yet, turn them into
assets, like an architect whose building is in harmony with the surrounding
environment.
Long rant made short: I’ve come around on A4 paper. Now that I know why
they’re aiming for √2, the fact that they’re doomed to miss the target
microscopically doesn’t much bother me. To be honest, A4 paper doesn’t even
look wrong to me anymore. Now it’s letter paper that looks off: a little stout and
old-fashioned.
It seems that I’ve completed the transition from one type of insufferable
American to another. From jingoistic defender of my own arbitrary national
customs, I’ve become a fervent evangelist for arbitrary foreign customs. I’m
even drinking fewer caramel buckets these days, although—much like letter
paper—I’m sure I’ll never give it up altogether.
Chapter 8
THE SQUARE-CUBE FABLES
Fables and math have a lot in common. Both come from dusty, moth-eaten
books. Both are inflicted upon children. And both seek to explain the world
through radical acts of simplification.
If you want to reckon with the full idiosyncrasy and complexity of life, look
elsewhere. Ask a biologist, or a painter of photorealistic landscapes, or someone
who files their own taxes. Fable tellers and math makers are more like
cartoonists. By exaggerating a few features and neglecting all the rest, they help
explain why our world is the way it is.
This chapter is a brief collection of mathematical fables. They show how
diverse settings, from baking to biology to the financing of the arts, are governed
by the constraints of geometry. At these tales’ collective heart is a single
fundamental idea, a moral so simple that even Aesop neglected to spell it out:
size matters.
A big statue is not just a big version of a small statue. It is a different object
entirely.
Needing to fill a double-sized pan, we are tempted to double the recipe. But
this would be a mere half measure. Look carefully and you will see: We need to
quadruple the recipe.
What happened? Well, the flat pan has two dimensions: length and width. By
doubling the length, we doubled the pan’s area. And by doubling its width, we
doubled the area again. That means the area has been double-doubled. Hence, a
multiplier of four.
This happens any time that you grow a rectangle. Triple the sides? Nine times
the area. Quintuple the sides? Twenty-five times the area. Multiply the sides by
9 bajillion? That’ll yield 81 bajillion bajillion times the area.
Or, more precisely: Multiply the lengths by r, and you multiply the area
by r2.
It’s not just rectangles. The same principle holds for all two-dimensional
shapes: trapezoids, triangles, circles, and any other vessel into which you might
pour the sacred brownie batter. As the length grows, the area grows much faster.
Back in the kitchen, we have just finished quadrupling our recipe when a
forgotten cupboard reveals the pans we’d been looking for all along. We blame
each other, then laugh, because who can stay mad when chocolate glory is so
close at hand?
We now face a choice: Shall we bake the brownies in the one big pan, or in
four small ones?
This is a fable, so we shall ignore the details. Forget oven temperature,
cooking times, heat flow, and minimizing the dishes to clean. Focus instead on
one matter: size itself.
As a brownie pan grows, its exterior (the length of the edges, which is one-
dimensional) gets larger. But its interior (which is two-dimensional) gets larger
faster. This means that small shapes tend to be “edge-heavy” while large shapes
are “interior-heavy.” In our case, the four small pans will have the same area as
the single big one, but they’ll have double the edge length.
The small pans will maximize our quota of edge brownies, while the big pan
will minimize it.
No matter how I try, I cannot fully embrace the humanity of people who
prefer edge brownies. Who would forsake a fudgy wonder for a chewy chore, a
crispy mistake? I can only imagine that they would prefer bones over meat,
crumbs over crackers, side effects over pain relief. Such people lie beyond
explanation and exculpation. Go big pan, or go home.
Well, you’ve doubled the length; that doubles the volume. You’ve doubled
the height, which doubles the volume again. And you’ve doubled the depth,
which doubles the volume a third time. That’s a triple double, although not in the
Russell Westbrook sense: here, double-double-double amounts to “multiply by
eight.”
The results are as clear as they are startling: Volumes get big fast. Triple the
sides of a cube? That’s 27 times the volume. Multiply the sides by 10? You’ll
multiply the volume by an absurd factor of a thousand. And what’s true of cubes
is true of every shape: pyramids, spheres, prisms, and (unlucky for Chares)
exquisite statues of the sun-god Helios. In precise terms: Multiply the length by
r, and you multiply the volume by r3.
If Chares had created a one-dimensional artwork (The Colossally Long String
of Rhodes), then the aesthetics might have suffered but his price scheme would
have worked fine: doubling the length would double the required material. Or
suppose he’d been commissioned to create a two-dimensional painting (The
Colossal Portrait of Rhodes). He’d still be lowballing the contract, but less so:
doubling the height of a canvas quadruples its area, demanding four times the
paint. Alas, Chares had the misfortune to work in a full three dimensions.
Doubling the height of his statue multiplied the bronze requirements by eight.
When a 1D length grows, a 2D surface grows faster, and a 3D volume grows
faster still. The Colossus of Rhodes—one of the wonders of the ancient world—
doomed its own creator for the simple reason that it was three-dimensional.
3. WHY THERE AREN’T GIANTS
King Kong, the three-story ape. Paul Bunyan, the lumberjack whose footsteps
carved out lakes. Shaquille O’Neal, the mythical seven-foot one, 325-pound
basketball player who could do everything except make free throws. You know
these stories, and you know just as well that they are fantasies, legends, wide-
eyed fictions. There’s no such thing as giants.
Why? Because size matters.
Suppose we take exemplary human specimen Dwayne Johnson and double
his dimensions. Now with double the height, double the width, and double the
depth, Dwayne’s total body mass has grown by a factor of eight.
So far, so good. But take a glance at his legs. To keep our man standing, his
bones will need eight times the strength. Can they muster it?
I doubt it. They’ve undergone two helpful doublings (in width and depth) but
one useless doubling: length. Just as you don’t reinforce a column by making it
taller, you can’t bolster a leg by making it longer. Extra length confers no extra
strength, just extra strain, as the bone’s base must now support the greater mass
above.
Dwayne’s leg bones won’t keep pace with the demands placed on them: A
factor of four can’t match a factor of eight. If we keep growing Dwayne
Johnson, doubling and tripling and quadrupling his size, then eventually he will
reach a critical breaking point. His limbs will buckle and splinter beneath the
overwhelming weight of his torso.
We’re dealing with a process called isometric scaling: growing a shape while
preserving its proportions. (Iso- means “same”; -metric means “measurement.”)
It’s a lousy method for making big animals. Instead, we need allometric scaling:
growing a shape while altering its proportions accordingly.
When an animal’s height grows by 50%, its legs can keep pace with the
strain only by growing 83% thicker. That’s why cats can survive on slender
limbs, whereas elephants need pillars to stand on.
A constraint upon Dwayne Johnson is a constraint upon us all, and thus,
giants shall always belong to the realm of myth. Paul Bunyan’s tibiae would
shatter with every lake-punching step. King Kong’s muscles (their strength
growing with the square of his height) could never carry his bulk, which grows
“with the cube”: he’d merely sit there, a giant monkey puddle suffering
perpetual heart failure. And Shaquille O’Neal? Well, his tale is so implausible
that I doubt anyone truly believes it.
Once again, ignore the details: the tiny toothless mouth, the eensy wiggling
toes, the small bald head that smells so amazing. Think of a baby the way you’d
think of any organism: as a homogenous bundle of chemical reactions. Every
activity of the body is built on such reactions; in some sense, the reactions are
the creature. That’s why animals are so temperature-sensitive: Too cold, and the
reactions slow to a halt; too hot, and some chemicals deform, making key
reactions impossible. You’ve got to keep a close eye on the thermostat.
Heat is created by reactions in each cell (i.e., in the interior). And heat is lost
through the skin (i.e., at the surface). This creates a familiar tug-of-war: interior
vs. surface.
Bigger animals, being more interior-heavy, will have an easy time keeping
warm. Smaller ones, being surface-heavy, will struggle. That’s why you’re most
vulnerable to cold in your surface-heavy extremities: fingers, toes, and ears. This
also explains why cold climates support only big mammals: polar bears, seals,
yaks, moose, mooses, meeses, and (depending on your zoology professor) the
Sasquatch. A surface-heavy mouse wouldn’t stand a chance in the Arctic. Even
at moderate latitudes, mice cope with heat loss by eating a quarter of their body
weight in food each day.
A baby isn’t a mouse, but it’s definitely not a yak. Its tiny body expends heat
like governments spend money. And to stifle that heat loss, there’s no cuddlier
option than a blankie.
Let’s tune our Digges shades to “100 light-years.” Gazing up at the night sky,
we say goodbye to Rigel, Betelgeuse, Polaris (a.k.a. the North Star), and many
other familiar orbs. The remaining stars, those within a hundred-light-year
radius, number about 14,000—a sparser, dimmer heavens than the usual.
Altogether, these stars will have some total brightness. We’ll give a lowball
estimate according to the following equation:
Turn the dial on your Digges shades, doubling the distance: 200 light-years.
Stars reappear. The sky brightens. But by how much?
Well, the visible sky forms a 3D hemisphere around Earth. By doubling its
radius, we’ve octupled its volume. Assuming (as Digges did) that the stars
follow an even distribution, like trees in a forest, then we’re now seeing eight
times as many stars as before. Their population leaps from 14,000 to more than
100,000.
But these new stars are farther, which means they’ll be dimmer. The
question, once again, is: By how much?
Each star appears in the night sky as a tiny circle. The bigger the circle, the
more light hits our eye. Since this is a fable, we can ignore personality
differences among stars—their temperatures, their colors, whether they have
cool names like Betelgeuse and Polaris. We’ll assume, like unapologetic star-
racists, that all stars are identical. The only thing that matters is their distance.
If Star A is twice the distance of Star B, then its circle will be half the height
and half the width. That leaves it with a quarter of the area—and thus, a quarter
of the brightness.
So, what can we make of our new and enlarged night sky? The number of
visible stars has been multiplied by eight; their minimum brightness has been
divided by four; and (correct me if I’m wrong) eight divided by four is two,
meaning that the total brightness, according to our simple equation, has doubled.
Double the radius on your Digges shades? Double the brightness of the sky.
Thank you for purchasing the Game of Dice! This fun pastime for the whole
civilization is suitable for ages from “stone” to “digital,” and beloved by gamers
from the commoner to the tyrant. Don’t take my word for it! Just ask these
Roman emperors:
This instruction manual will introduce you to the basic rules. The Game of
Dice is equal parts theory and practice, stretching the mind and the fingertips
alike. Let’s play!
This snout-faced mole rat of a polyhedron dashes our hopes. Its 12 faces are
equilateral triangles, all identical. Yet it does not make a fair die. At some
corners, four triangles meet. At others, five do. When you roll the little monster,
some faces come up more than others. I’m afraid congruence just isn’t enough.
We need symmetry.
In ordinary speech, “symmetry” refers to a vague, pleasing uniformity. Its
mathematical meaning is more specific: a geometric action that transforms an
object without really changing it. For example, a square table has eight
symmetries:
The symmetries leave the table unchanged, though on careful inspection, they
scramble the locations of the corners. For example, the 180° rotation swaps pairs
of opposite corners: #1 and #3 trade places, as do #2 and #4. Contrast this with
the diagonal reflection, which swaps #2 and #4, but leaves #1 and #3 where they
are. Symmetries on dice work in much the same way: they rearrange the faces
while leaving the overall shape unchanged.
Symmetry offers a surefire path to fair dice. Just pick a shape with enough
symmetries that every face can be swapped with every other.
One example: the dipyramid. Take two identical pyramids and glue their
bases together. The right combination of rotations can swap any triangular face
with any other, meaning the faces are geometric equivalents, and so the die is
fair.
Another example: the trapezohedron, which looks like a dipyramid with
some nifty carving at its equator, to turn the three-sided triangles into four-sided
kites.
With a dipyramid or a trapezohedron, you can create any even-sided die:
eight-sided, 14-sided, 26-sided, 398-sided. In theory, every single one will be
fair, each face appearing with equal likelihood. You might think that we’ve
solved the problem. The Game of Dice is over, yes?
Not so fast! People are fickler than that. It’s not enough for dice to be fair…
The best dice need more than symmetric faces. They need symmetric
everything. And if you’re a polyhedron aficionado, you know what that means.
Platonic solids!
Of all the straight-edged 3D shapes, the Platonic solids are the most perfect.
Their symmetries can swap any two faces, any two corners, or any two edges—
symmetries so magnificent that even a hardened cynic cannot doubt their
eminent fairness.
There are five Platonic solids: no more, no less. And each member of this
geometric pantheon has graced Earth in the form of a die.
In the Game of Dice, the Platonic solids are the trump cards. One cannot
picture them without heavenly choirs piping up in the background. But, as an
exclusive pantheon of five, they are all too rare. They let us randomize among 4,
6, 8, 12, or 20 outcomes… but no others.
It’s worth trying to break out of the box. Why not a paradigm-busting design,
a fresh and innovative way to randomize any number of outcomes?
Spoiler: It’s harder than it sounds.
RULE #3: GOOD DICE WORK WELL ANYWHERE.
One alternative is “long dice.” Instead of worrying about every face having an
equal chance, build a long prism instead.
These dice work not because all faces come up with equal likelihood, but
because two of them never come up at all. Long dice play fair, look fair, and
allow you to randomize among any number of outcomes. So why aren’t they
more popular?
Well… they roll too much.
Whereas Platonic solids treat the table like a dance floor, leaping this way
and that, long dice roll in a single direction. You’ve got to clear a whole
bowling-alley path for them. What kind of self-important die demands a red
carpet?
Back to the chalkboard we go, to recruit another mathematical principle:
continuity.
Grab your long die from wherever it rolled to a stop. I’ll wait. (And wait.
And wait.) Now, as we can see, the two bases almost never come up. But
imagine shortening the die, to make a “less long” variant. The shorter it grows,
the more probable the two bases become. After a while, you’re left with a die so
short that it’s basically a coin, at which point the roles reverse. Now, almost
every toss lands on a base.
Somewhere in the middle, there is a moment of crossing over: a length at
which the bases and the lateral faces enjoy the same likelihood. Somewhere in
the middle is a fair die.
In theory, you can pull this trick with any polyhedron, achieving shapes that
are funky-looking yet fair. So where are they? Why aren’t novelty shops selling
nifty trick dice that look wild but play well—the bizarro antithesis of loaded
dice?
It’s because the adjustments are too sensitive. Fair on hardwood? Unfair on
granite. Fair at one size? Unfair when built double-sized. Fair on one roll?
Unfair at a different strength, or with a different spin. Change any condition, no
matter how incidental, and you alter the physics. Such dice would be chained to
the hyperspecific circumstances of their birth. People want portable, durable dice
—not fussy prima donnas.
This procedure works for any randomization scenario. Say we want to pick a
random word from the Lord of the Rings trilogy. There are about 450,000 to
choose from. So flip 19 coins, yielding more than 500,000 possible outcomes.
Assign a word to each. If you land on one of the unassigned outcomes, just flip
all 19 coins again.
Heck, you don’t even need 19. Just flip a single coin 19 times.
By this logic, there is nothing dice can accomplish that a single coin cannot
match. And yet it’s hard to picture a Vegas casino drawing in crowds with a
coin-only craps table or a penny-based roulette wheel.
The problem is obvious: these systems are too complicated. It’s an awful lot
of trouble to record a sequence of coin flips, look up the result in an index, and
then (if necessary) repeat the whole process. You want a single roll. No sides
wasted. No user manual required.
This principle nixes some clean mathematics. For example, you could
randomize among four outcomes with a simple cubical die. Just label two of the
faces “roll again.” But this approach bugs people. It feels inelegant to waste
faces. When sharing a cake among four friends, you would never cut six pieces
and then throw out the two extras.
I suspect that’s why Dungeons & Dragons players roll a four-sided die
instead. I read this as a sign of desperation, because of all the Platonic solids, the
tetrahedron is history’s least popular die. Again, any human can see why: It
lands with a face down but not a face up. This feels wrong, like asking someone
to guess the number that you’re not thinking of.
Across the millennia, humans have shunned the tetrahedron in favor of dice
that have parallel pairs of opposite faces, so that every “down” face corresponds
to an “up.” Mathematics doesn’t care. But humans are the ones calling the shots.
RULE #5: THEY’RE HARD TO CONTROL.
Remember the whole purpose of the die? It puts the human body in contact with
higher forces: random chance, karmic fate, the will of the gods. This enables
brainstorming, games of luck, the telling of fortunes, and other profound
expressions of our humanity.
So of course, people try to cheat.
One path: manipulate the exterior of the die. For example, extend it by an
undetectable increment, forming a brick shape. Or create faces with slight bulges
(making them less likely) or indentations (making them likelier). Or you can cap
some faces with a bouncy material, or sandpaper them down so the die prefers
them. These tricks are as old as ruins. I mean that literally: loaded dice with
shaved corners have been excavated in Pompeii.
A second approach: manipulate the interior of the die. “Trappers” are dice
with two hidden compartments; the right motion will move a heavy blob of
liquid mercury from one compartment to the other, shifting the probabilities. (If
poison metals aren’t your cup of tea, then use wax that melts just below body
temperature.) Another scheme: Back when wooden dice were popular, some
cheaters grew small trees with pebbles embedded in the branches. Then they’d
carve a die from the wood, leaving that invisible rocky weight within. This con
job demands not just extraordinary patience but above-average botanical skill.
A third approach: renumber the faces. A normal die features opposite faces
that sum to seven. (The pairs are 1 and 6, 2 and 5, and 3 and 4.) In cheat dice
known as “taps,” some numbers are duplicated, yielding (for example) opposite
pairs of 6 and 6, 5 and 5, and 4 and 4. From any angle, your victims will see only
three faces, so nothing appears to be wrong.
Though all of these cheating methods target cubical dice, it’s not that cubes
are extra vulnerable. They’re just extra popular. Evidently, there’s more money
in craps than in D&D.
RULE #6: GOOD DICE FEEL NICE.
The Game of Dice is like most games, in that nobody really needs it. We live in
the 21st century. I jetpack to work and take vacations by flying car. Okay, I
don’t, but I’ve got half the world in my pocket in the form of a 5-ounce
computer. Technology is making us all obsolete, and that includes the blue-
collar workers called dice. Watch: I’m going to stop writing and go simulate a
million rolls of a cubical die in Microsoft Excel. I’ll let you know how long it
takes.
Okay, done. Took about 75 seconds. Here are the results:
Not only is computer randomization faster and easier than throwing a plastic
cube across a table, it’s more random, too. Casinos could eliminate craps tables
and roulette wheels tomorrow, and their digital replacements would outperform
the old dinosaur randomizers by leaps and bounds.
But what fun would that be?
Dice are meant to be held. The first time I played Dungeons & Dragons
(okay, the only time), I encountered something more alluring than any of the
orcs and mages in the game itself: the Zocchihedron, the 100-sided die. Can you
imagine? A hundred sides! A die that rolls for 30 seconds before coming to a
stop! I knew that two d10s (one for the tens digit, one for the units) served better
and was fairer than Lou Zocchi’s lumpy shaved golf ball. I didn’t care. I wanted
to roll that d100.
Greeks must have felt that same allure when tossing the sheep knuckle they
called the astragalos. They valued its four sides at strange values (1, 3, 4, and 6)
and threw them by the handful. A throw of all 1’s was “the dogs,” the worst
throw imaginable. The best throw (either all 6’s or one of each face, depending
who you ask) was called “Aphrodite.” The astragali weren’t fair; they were
something better. They were skeletons held in a living hand to foretell fate.
When Julius Caesar crossed the Rubicon, tipping history toward the end of the
Roman Republic and the dawn of empire, his comment was “Alea iacta est”:
The die is cast.
The Game of Dice, I suspect, will never end. These objects speak to
something deep in our bones. Just remember to follow those six rules:
Our aim: To cow the galaxy with the greatest structure it had ever seen. Our
funding: virtually infinite, thanks to the Emperor’s unflinching approach to
taxation. The sky was—quite literally—the limit. So our first question was:
What should this thing look like?
IMPERIAL GEOMETER
I was told to find a simple, elemental design. Something so unnerving and
spectacular that it would make droids weep and bounty hunters lose bladder
control. Compared to peer-reviewing articles, it was a tough assignment.
DARTH VADER
A hexagonal prism? Was this to be an empire of honeybees?
IMPERIAL GEOMETER
Even in my deepest valleys of self-doubt, I was grateful for Lord Vader’s
feedback. Like a lot of visionaries, he is an exacting manager. But I know it’s all
meant as constructive criticism.
DARTH VADER
Imbecile.
IMPERIAL GEOMETER
Eventually, we identified a single surpassing goal: symmetry.
Most people use the term casually, but in mathematics, “symmetry” has a
precise meaning: it’s a thing you can do to a shape that leaves it looking the
same.
For example, a Wookiee face has a single symmetry: you can reflect it in a
vertical mirror. That’s it. If you do anything else—say, rotate it 90°, or reflect it
in a horizontal mirror—then you rearrange the face entirely, after which the
Wookiee may undertake to rearrange YOUR face.
IMPERIAL GEOMETER
One day, Lord Vader swept my illustrations from the table and roared, “MORE
SYMMETRY!” We’d been considering an icosahedron, which has 120
symmetries. How could I top that? But then, in the proudest moment of my
career—heck, of my LIFE—it came to me: the maximally symmetric shape.
DARTH VADER
Why did he not think of a sphere to begin with? One loses so much time to the
glacial workings of feeble minds.
IMPERIAL GEOMETER
I still maintain it was a mistake to include that weapon. I mean, which is more
effective for intimidation: a little laser light show, or an INFINITE number of
symmetries?
Right away, we encountered a problem. Ran into a headwind, you might say.
Back then, everybody was into Star Destroyers: these sleek, angular designs.
They were galactic steak knives, ready to pop a star like a helium balloon. I
thought the appeal was aesthetic, but when I learned the design was functional,
too, it spelled trouble for our sphere.
IMPERIAL PHYSICIST
Imagine you’re flying an airplane. No matter how good a pilot you are, you’re
going to have A LOT of collisions. I’m referring, of course, to air molecules.
Best-case scenario? The air molecules travel PARALLEL to your surface.
Then, they won’t impact you at all. They’re like passing traffic in the
neighboring lane. The worst-case scenario is that the air molecules hit
PERPENDICULAR to your surface, at 90-degree angles. Then, your vessel
bears the full force of the impact. That’s why you don’t build airplanes with big,
flat fronts: it’d be like trying to weasel through a crowd while wearing a giant
sandwich board on your torso.
Hence, the tapered design of the Star Destroyer. When it’s traveling through
atmosphere, the air molecules mostly glance off the sides—not quite parallel to
the surface, but close enough. The Death Star, by contrast, gave us all
aerodynamics nightmares. There’s a huge surface hitting the air at near-perfect
right angles.
IMPERIAL ENGINEER
Imagine your friends are launching paper planes, and instead of joining in, you
decide to throw a desk across the room. It’s going to cost a lot more energy, and
the flight won’t be pretty.
DARTH VADER
It was a harsh sacrifice. But leadership is no place for the fainthearted.
Soon, another obstacle leapt into our path: our physicist kept insisting that the
Death Star really ought to be lumpy and asteroid-shaped.
IMPERIAL PHYSICIST
Glance around the galaxy. Where do you see spheres? It’s the big, heavy stuff.
Stars, planets, a few larger moons. Now look at the smaller, less dense objects:
asteroids, comets, dust clouds. You find a lot of weird potato shapes.
That’s not a coincidence. It’s a fact of gravity. I told them from the start: The
Death Star is too small to become a sphere.
DARTH VADER
I was about to break ground on the most ambitious construction project in the
history of evil, and a lab coat with glasses was telling me it was TOO SMALL.
Was I angry? You tell me.
IMPERIAL PHYSICIST
Look, I’m no propaganda expert, but the physics is pretty clear here. All matter
attracts all other matter. More matter, more attraction. That’s gravity.
So, toss a bunch of ingredients together in the mixing bowl of space, and
every bit is mutually drawn toward every other bit. They congregate around a
kind of 3D balancing point: the center of mass. Over time, the outlying clumps
and more distant protrusions are drawn toward this center, until it reaches the
final equilibrium shape: a perfect sphere.
But that’s only if you’ve got enough matter. Otherwise, the attraction is too
weak to conquer those lumpy outcroppings. Hence, big planets go spherical, but
small moons stay potatoes.
DARTH VADER
I wondered: Are all physicists this insolent? Should I drain the life out of every
scientist right now, to spare myself future encounters?
IMPERIAL PHYSICIST
The magic size, where you’re big enough to go spherical, depends on what
you’re made of. Ice will go spherical at a diameter of about 400 kilometers,
because it’s pretty malleable. Rock is much stiffer and needs more gravitational
persuasion—so it won’t go spherical until the diameter hits 600 kilometers. For a
material like imperial steel, designed to withstand tectonic-level forces, it’d be
even larger. Maybe 700 or 750 kilometers.
And the Death Star? It was only 140 kilometers across. A pebble.
IMPERIAL PHYSICIST
You could see Lord Vader’s face light up. Well… I think you could. His mask
looked smiley. Anyway, he clearly loved the idea that a spherical shape could
imitate and evoke larger celestial bodies.
DARTH VADER
There is a fearsome sea creature that inflates itself into a sphere in order to
intimidate its enemies. That is why, in my journals, I call the Death Star “the
Pufferfish of the Skies.”
Ever seen footage of the Death Star? You get the sense it was teeming with
stormtroopers. Densely packed as a submarine.
Ha. In reality, the Death Star was the loneliest, emptiest place I’ve ever been.
STORMTROOPER
And the ultimate insult? They slept us 60 guys to a room. Each dorm had three
toilets and two showers. Forget atomizing planets; if you want to give people
nightmares, just show them the lines outside those bathrooms every morning.
The Emperor can afford a multi-quadrillion-dollar space orb, and here I am
on board, sleeping in a triple-decker bunk bed and walking half a kilometer to
pee? Ugh. It still makes me mad.
5. CURVES YOU CAN’T SEE
Fundamentally, the Death Star wasn’t about personnel. All the humans, all the
machines—reactor core, sublight engine, hyperdrive—they were merely a
support system for the station’s real purpose: the superlaser.
IMPERIAL GEOMETER
Ah, that’s another reason that a sphere was the perfect shape! For a given
volume, it’s got the lowest possible surface area. If you’re building a casing for
something huge—say, a planet-demolishing laser gun—then a sphere will get it
done with the minimum of material.
IMPERIAL ENGINEER
I know the geometer will tell you that the spherical shape saved us money,
because a cube would have required 24 percent more steel. That’s a typical
mathematician for you: all theory, no practicality.
There’s a reason we don’t tend to build spherical ships: Curves are a pain!
Ever tried to arrange furniture in a curved room? Good luck positioning the sofa.
Now, the Death Star was too big for you to notice the curvature on that level.
No problem for the interior decorators. But on a construction level, that
curvature gave me a multiyear migraine. You needed the steel beams to bend by
about zero degrees, zero minutes, and 3 seconds of arc per meter—that’s less
than 1 degree of curvature per kilometer.
What a mess. The curvature was low enough that you couldn’t see it with the
naked eye, but high enough that every single part had to be custom-built.
IMPERIAL GEOMETER
Like I said, a sphere minimizes surface area. But I’ll admit that it had other
drawbacks.
IMPERIAL ENGINEER
The problem that still haunts me is the heating. It’s outer space, right? Cold. You
want to retain heat, and a sphere is great for that. Minimum surface area means
minimum heat loss. But apparently, we did our jobs TOO well, because early
simulations showed that the station would be prone to overheating.
IMPERIAL ARCHITECT
We had to dispose of some excess heat, so I put in thermal vents. Nothing big. A
few meters wide. Release the heat into space; problem solved.
I didn’t think…
I mean, when I heard that the rebels had destroyed the station by exploiting a
thermal vent…
DARTH VADER
They also succeeded in transmitting a proton torpedo INTO the station.
IMPERIAL GEOMETER
The Death Star achieved this eerie fusion: so large it must have been natural, but
so perfect it couldn’t be. It was disturbing, compelling, horrifying. That’s the
power of geometry.
DARTH VADER
As often happens, our critics supplied our finest slogan. When my old frenemy
Obi-Wan said, “That’s no moon,” we knew we had the tagline for the
advertisements.
III
PROBABILITY
Ever flipped a coin? Unless you’re too poor to have touched one, or too rich to
have bothered, I’m going to guess “yes.” I’ll also guess that, even though the
odds are 50/50, your most recent flip did not show a 50/50 mix of outcomes. It
showed “heads” or “tails.” All or nothing.
That’s how life is: full of random one-off events. Unexpected train delays.
Come-from-behind victories. Magical out-of-nowhere parking spots. In our
storm-tossed world, anything can happen, and fate never RSVPs.
But if you could flip a trillion coins, you’d find yourself approaching a
different world altogether: a well-groomed land of long-term averages. Here,
half of all coins land on heads, half of all newborns are boys, and one-in-a-
million events happen a millionth of the time, give or take. In this blue-skied
theoretical realm, quirks and coincidences don’t exist. They vanish against the
totality of all possibilities, like stones thrown into the sea.
Probability bridges these two realms. The wild, woolly world we know is
what makes probability necessary. The calm, measured world we can never quite
reach is what makes probability possible. The probabilist is a dual citizen,
aiming to understand every breaking headline and celebrity fiasco as a single
card drawn from an infinite deck, a cup poured from a bottomless pitcher. As
mortal creatures, we’ll never set foot in the land of eternity—but probability can
offer us a glimpse.
Chapter 11
THE 10 PEOPLE YOU MEET IN LINE FOR THE LOTTERY
Ah, the lottery ticket. It’s a certificate of optimism, a treasury bond from the
Department of Hope. Why cling to a drab, dog-eared $1 bill when you could
exchange it for a thrilling mystery bill worth anywhere from zero to $50 million?
If that sounds unappealing to you, well, you and humanity will just have to
agree to disagree.
I should confess that I’ve spent less on lottery tickets in my life ($7) than I
spent on croissants this month ($don’t ask). Nevertheless, each year, roughly
half of US adults play the lottery. It’s not the half you might guess. People
earning at least $90,000 are likelier to play than those earning below $36,000.
Those with BAs participate more than those without. The state with the highest
rate of lottery spending is my home state of Massachusetts: a haven of wealthy,
overeducated liberals, who spend $800 per person each year on the lottery.
Playing the lottery is like watching football, suing your neighbors, or butchering
the national anthem: an American pastime, pursued for a variety of reasons.
Come, join me in line for a ticket, and let’s investigate the multifaceted
appeal of turning your money into commodified chance.
1. THE GAMER
Behold! It’s the Gamer, who buys lottery tickets for the same reason I buy
croissants: not for sustenance but for pleasure.
Take the Massachusetts game entitled $10,000 Bonus Cash. It’s a genius
name. Put “10,000” and “bonus” in front of any word and there’s no way you
can go wrong. Plus, the graphics on the $1 tickets for this game look like rave
night at the color printer. On the back, you’ll find the following complicated
odds of victory:
What is this ticket worth? Well, we don’t know yet. Maybe $10,000; maybe
$5; maybe (by which I mean “very probably”) nothing.
It’d be nice to estimate its value with a single number. So imagine that we
spent not a mere $1, but $1 million. This lets us escape the rowdy dance floor of
the short-term world into the peace and quiet of the long-term one, where each
payout arrives in the expected ratio. In our million tickets, a one-in-a-million
event will occur roughly once. A 1-in-100,000 event will occur roughly 10
times. And a 1-in-4 event will occur 250,000 times, give or take.
Sifting through our stacks upon stacks of overstimulating paper, we’d expect
the results to look something like this:
About 20% of our tickets are winners. Totaling up all of the prizes, our $1
million investment yields a return of roughly $700,000… which means that we
have poured $300,000 directly into the coffers of the Massachusetts state
government.
Put another way: on average, these $1 tickets are worth about $0.70 each.
Mathematicians call this the ticket’s expected value. I find that a funny name,
because you shouldn’t “expect” any given ticket to pay out $0.70, any more than
you would “expect” a family to have 1.8 children. I prefer the phrase long-run
average: It’s what you’d make per ticket if you kept playing this lottery over and
over and over and over and over…
Sure, it’s $0.30 less than the price you paid, but entertainment ain’t free, and
the Gamer is happy to oblige. Poll Americans about why they buy lottery tickets,
and half will say not “for the money” but “for the fun.” These are Gamers.
They’re why, when states introduce new lottery games, overall sales rise.
Gamers don’t see the new tickets as competing investment opportunities (which
would lead to a corresponding drop in sales for old tickets), but as fresh
amusements, like extra movies at the multiplex.
What’s the precise attraction for the Gamer? Is it the gratification of victory,
the adrenaline rush of uncertainty, the pleasant buzz of watching it all unfold?
Well, it depends on each Gamer’s appetites.
I’ll tell you what it isn’t: the financial benefit. In the long run, lottery tickets
will almost always cost more than they’re worth.
2. THE EDUCATED FOOL
Wait… “Almost always”? Why “almost”? What state is dumb enough to sell
lottery tickets whose average payout is more than their cost?
These exceptions emerge because of a common rule by which big-jackpot
lotteries sweeten the deal: if no one wins the jackpot in a given week, then it
“rolls over” to the next week, resulting in an even bigger top prize. Repeat this
enough times, and the ticket’s expected value may rise above its price. For
example, in January 2016, the UK National Lottery draw offered a juicy
expected value of more than £4, compared to a ticket price of only £2. Weird as
they appear, such schemes typically spur more than enough sales to justify their
cost.
In the queue for lotteries like this, you’ll meet a very special player, a treat
for gambling ornithologists like us. See it there, high in the branches, grooming
itself? It’s the Educated Fool, a rare sap-brained creature who does with
“expected value” what the foolish always do with education: mistake partial
truth for total wisdom.
“Expected value” distills the multifaceted lottery ticket, with all its prizes and
probabilities, down to a one-number summary. That’s a powerful move. It’s also
simplistic.
Take these two tickets, each costing $1.
Spend $10 million on A, and you expect to earn back only $9 million,
amounting to a $0.10 loss per ticket. Meanwhile, $10 million spent on B should
yield $11 million in prizes, and thus a profit of $0.10 per ticket. So to those
enamored with expected value, the latter is a golden opportunity, while the
former is pyrite trickery.
And yet… would $11 million bring me any greater happiness than $9
million? Both values exceed my current bank account many times over. The
psychological difference is negligible. So why judge one a rip-off and the other a
sweet bargain?
Even simpler, imagine that Bill Gates offers you a wager: for $1, he’ll give
you a 1 in a billion chance at $10 billion. Calculating the expected value, you
start to salivate: $1 billion spent on tickets would yield an expected return of $10
billion. Irresistible!
Even so, Educated Fool, I beg you to resist. You can’t afford this game.
Scrape together an impressive $1 million, and the 99.9% likelihood is still that
rich man Gates walks away $1 million richer while you walk away broke.
Expected value is a long-run average, and with Gates’s offer, you’ll exhaust your
finances well before the “long run” ever arrives.
The same holds true in most lotteries. Perhaps the ultimate repudiation of
expected value is the abstract possibility of $1 tickets like this:
If you buy 10 tickets, you’re likely to win $1. That’s pretty terrible: only
$0.10 per ticket.
If you buy 100 tickets, you’re likely to win $20. (That’s 10 of the smallest
prize, and one of the next smallest.) Slightly less terrible: now $0.20 per ticket.
If you buy 1000 tickets, you’re likely to win $300. (That’s a hundred $1
prizes, ten $10 prizes, and a single $100 prize.) We’re up to $0.30 per ticket.
Keep going. The more tickets you buy, the better you can expect to do. If you
somehow buy a trillion tickets, the likeliest outcome is that you will win $1.20
per ticket. A quadrillion tickets? Even better: $1.50 per ticket. In fact, the more
tickets you buy, the greater your profit per ticket. If you could somehow invest
$1 googol dollars, you’d be rewarded with $10 googol in return. With enough
tickets, you can attain any average return you desire. The ticket’s expected value
is infinite.
But even if you trusted a government to pay out, you could never afford
enough tickets to glimpse that profit. Go ahead and spend your life savings on
these, Educated Fool. The overwhelming likelihood is bankruptcy.
We humans are short-run creatures. Better leave the long-run average to the
immortals.
3. THE LACKEY
Ooh, look who just stepped into the queue. It’s the Lackey!
Unlike most folks here, the Lackey’s windfall is guaranteed. That’s because
the Lackey isn’t keeping the tickets, just pulling down minimum wage to stand
in line, purchase them, and pass them off to someone else. Modest
compensation, but ultrareliable.
Who would pay such a Lackey? Well, that’s a question for our next player…
In the long run, it doesn’t matter. Choose B 100 times, and (on average)
you’ll get 90 windfalls, along with 10 disappointing zeroes. That’s a total of
$90,000 across 100 trials, for an average of $900. Thus, it shares the same
expected value as A.
Yet if you’re like most people, you’ve got a strong preference. You’ll take
the guarantee, rather than grab for a little extra and risk leaving empty-handed.
Such behavior is called risk-averse.
Now, try this one:
It’s the mirror image of Question #1. In the long run, each yields an average
loss of $900. But this time, most people find the guarantee less appealing.
They’d rather accept a slight extra penalty in exchange for the chance to walk
away scot-free. Here, they are risk-seeking.
These choices are characteristic of prospect theory, a model of human
behavior. When it comes to gains, we are risk-averse, preferring to lock in a
guaranteed profit. But when it comes to losses, we are risk-seeking, willing to
roll the dice for the chance to avoid a bad outcome.
A crucial lesson of prospect theory is that framing matters. Whether you call
something a “loss” or a “gain” depends on your current reference point. Take
this contrast:
The questions offer identical choices: (a) you walk away with $1500, or (b)
you flip a coin to determine whether you end with $1000 or $2000. But folks
don’t give identical responses. In the first case, they prefer the $1500
guaranteed; in the second, they favor the risk. That’s because the questions
create different reference points. When the $2000 is “already yours,” the thought
of losing it stresses you out. You’re willing to take risks to avoid that fate.
When life is hard, we roll the dice.
This line of research has a sad and illuminating implication for lottery ticket
purchasers. If you’re living under dire financial conditions—if every day feels
like a perpetual loss—then you’re more willing to risk money on a lottery ticket.
Think of how a basketball team, trailing late in the game, will begin fouling
its opponents. Or how a hockey team, down a goal with a minute to play,
replaces its goalie with an extra forward. Or how a political candidate, behind
with two weeks until the election, will go on the attack, hoping to shake up the
campaign. These ploys harm your expected value. You’ll probably lose by an
even larger margin than before. But by heightening the randomness, you boost
your chances at winning. In times of desperation, that’s all people want.
Researchers find that those who buy lottery tickets “for the money” are far
likelier to be poor. For them, the lottery isn’t an amusing way to exercise wealth,
but a risky path to acquire it. Yes, on average, you lose. But if you’re losing
already, that’s a price you may be willing to pay.
7. THE KID WHO JUST TURNED 18
Hey, look at that fresh-faced young lottery player! Does the sight fill you with
nostalgia for your own youth?
Or does it make you think, “Hey, why are states in the business of selling a
product so addictive that they feel compelled to protect minors from buying it?”
Well, there’s someone else I’d like you to meet…
9. THE DREAMER
Now, allow me to introduce you to a starry-eyed hopeful, who holds as special a
place in my heart as the lottery holds in theirs: the Dreamer.
For the Dreamer, a lottery ticket isn’t a chance to win money. It’s a chance to
fantasize about winning money. With a lottery ticket in hand, your imagination
can go romping through a future of wealth and glory, champagne and caviar,
stadium skyboxes and funny-shaped two-seater cars. Never mind that winning a
lottery jackpot tends to make people less happy, a trend well documented by
psychologists. Daydreaming about that jackpot affords a blissful few minutes
while driving your disappointing regular-shaped car.
The prime rule of fantasy is that the top prize must be enough to change your
life, transporting you into the next socioeconomic stratum. That’s why instant
games, with their modest top prizes, draw in low-income players. If you’re
barely scraping together grocery money each week, then $10,000 carries the
promise of financial transformation. By contrast, comfortable middle-class
earners prefer games with multimillion-dollar jackpots—enough kindling for a
proper daydream.
If you’re seeking an investment opportunity, it’s crazy to fixate on the
possible payout while ignoring the probability. But if you’re seeking a license
for fantasy, then it makes perfect sense.
This Dreamer tendency helps explain the lottery’s odd economy of scale,
whereby big states fare better than small ones. Imagine a simplified lottery
game: Half of all revenue goes to the state, and the other half goes to the jackpot
winner. If a state sells $1 million in tickets, then each player has a 1-in-a-million
chance at a $500,000 jackpot. Meanwhile, if a state can sell $100 million in
tickets, then the odds drop to 1 in 100 million, and the jackpot leaps to $50
million.
The expected value doesn’t change. In each state, an average ticket is worth
$0.50. But the psychology of the Dreamer will favor the megajackpot with the
miniscule odds.
10. THE ENTHUSIAST FOR SCRATCHING THINGS
Ah, this citizen knows where it’s at. Forget the cash prizes and all that
probability mumbo-jumbo. The chance to rub a quarter across some cardboard is
winnings enough. The state lottery: it’s like a scratch-and-sniff for grown-ups.
Chapter 12
CHILDREN OF THE COIN
If I’ve been stumped and impressed by one thing in my teaching career, it’s
human genetics. I’m not referring to that unfortunate semester in 2010 when I
was pressed into teaching 10th-grade biology. Rather, I refer to the job’s finest
long-term perk: getting to know families.
Every time you teach a pair of relatives—brother and sister, cousin and
cousin, aunt and equal-aged nephew—you renew your wonder at the scattershot
nature of biological inheritance. I’ve taught siblings who looked like twins, and
twins who looked like strangers. Parents’ nights always mess with my head.
Meeting after meeting, my mind performs a real-time face-mash of the two
adults before me, and I discover that their kid is a seamless Photoshop job: his
ears flanking her eyes, his hair on her head shape. All families are alike, and yet
no two families are alike in quite the same way.
The riddles of resemblance strike at the heart of biology. But as my students
can tell you, I am not a biologist. I have no DNA sequencer, no special
knowledge of introns or histones, and (more to the point) no clue. No, I’m a
mathematician. What I’ve got is a coin, a theorem, and a Brave Little Toaster–
esque faith in the power of figuring things out from first principles.
And maybe that’s enough.
This chapter is about two seemingly unrelated questions in the field of
probability. First is the question of genetic inheritance, which could fill grad-
school textbooks. Second is the question of coin flips, which feels so trivial it’s
barely worth stooping to pick up off the ground.
Can we bridge the two realms? Can the coin’s monosyllabic grunts give
expression to the complexity of the human race?
I want to say yes. And I’m pretty sure I got that from my dad.
Okay, let’s start with the easier of the two questions: What happens when you
flip a coin?
Answer: The two outcomes, heads and tails, occur with equal likelihood.
Solved!
Mmm, you feel that? It’s the pleasure of turning away from an urgent,
intractable real-world problem to focus instead on a puzzle that nobody cares
about. Savor the sensation. It’s why people become mathematicians.
Okay, so nothing much happens when we toss a single coin. But what if we
toss a pair? You’ll find that four outcomes are equally likely:
A stickler for specifics will see the two middle outcomes as different: head
then tail vs. tail then head. (Imagine that one coin is a penny and the other is a
nickel.) But if you’re like most coin flippers, you won’t care. Your mind
collapses both into a single outcome (“one head, one tail”) that happens to have
double the probability.
Keep going: What if we toss three coins?
Eight outcomes in all. If you’re hoping for all heads, there’s only one path:
every coin must perform its duty, so the probability is 1 in 8. But if you’re
aiming for two heads and a tail, then you’ve got three ways to do it—the tail can
appear on the first, second, or third coin—giving a probability of 3 in 8.
Now, what about four coins—or five, or seven, or 90?
In fact, data on nearly 800 sibling pairs, compiled by Blaine Bettinger at his blog
The Genetic Genealogist, matches our crude prediction pretty darn well:
There’s a caveat. I’ve omitted a factor so important that I can’t blame
biologists if they’re rage-tearing the pages of this book right now. That’s
crossing over, or recombination.
I’ve claimed that chromosomes are passed down undisturbed and intact. Like
many things I said while teaching biology, that’s false. Before a single version of
each chromosome is selected, the two experience a splicing-together—for
example, swapping the middle third of each. Thus, a given chromosome may
come mostly from Grandpa, but it’ll feature a handful of recipes exchanged for
the equivalents from Grandma.
Crossing over happens roughly twice per chromosome. Thus, to enhance the
accuracy of our model, we can triple the number of coin flips (since two
crossovers will split a chromosome into three pieces).
How does this impact the number of possible offspring? Remember, linear
growth in the number of objects yields exponential growth in their combinations.
So the variety of offspring does far more than triple. In fact, it grows from 246
(or 70 trillion) to 2138 (which is a staggering 350 duodecillion).
I’m arguing, in short, that a newborn baby has a lot in common with a
handful of scattered pennies. That’s not to say you should revalue your infant’s
life at $0.46. To the contrary, you should admire $0.46 of spilled pocket change
as a miracle on par with human birth.
Getting to know families, you feel that they mix like liquids, blue and yellow
paint sloshing together to create green. But they don’t. Families mix like decks
of cards, with elements shuffled and reshuffled, tucked up one’s sleeve and then
brought back into play. Genetics is a game of combinations: the lovely and
elusive patterns of resemblance between siblings can all be traced back to the
combinatorial explosion. Flip enough coins, and the stark, distinct outcomes
(heads vs. tails) begin to blur and blend. The jagged edges of the graph fuse to
form something as liquid and flowing as humanity itself. In this way, we are the
offspring of combinatorics, the progeny of a shuffled deck, the children of the
coin.
Chapter 13
WHAT DOES PROBABILITY MEAN IN YOUR
PROFESSION?
No big deal, right? I mean, does probability ever come up in the real world?
It’s not like we spend our lives clawing for intellectual tools that might offer the
slightest stability in the swirling miasma of uncertainty that surrounds us every
waking moment…
Well, just in case—this chapter is a handy guide on how various sorts of
human beings think about uncertainty. Just because it’s hard doesn’t mean we
can’t have some fun with it.
Hello! You are a person. You have two eyes, a nose, and sentience. You
dream, laugh, and urinate, not necessarily in that order.
Also, you live in a world where nothing is certain.
Take a simple fact: How many planets does your solar system have? These
days, you say “eight.” From 1930 until early 2006, you answered “nine”
(counting Pluto). In the 1850s, you wrote textbooks listing twelve: Mercury,
Venus, Earth, Mars, Jupiter, Saturn, Uranus, Neptune, Ceres, Pallas, Juno, and
Vesta. (You now call the last four “asteroids.”) In 360 BCE, you named seven:
Mercury, Venus, Mars, Jupiter, Saturn, the moon, and the sun. (You now call the
last two “the moon” and “the sun.”)
You keep changing your mind as new data and new theories emerge. That’s
classic you: When it comes to knowledge, you’ve got lots of good ideas but no
guarantees. Teachers, scientists, politicians, even your sense organs—any of
these could deceive you, and you know it.
Probability is the language of your uncertainty. It lets you quantify what you
know, what you doubt, what you doubt you know, and what you know you
doubt, expressing these nuances of confidence in a clear and quantitative
language. At least, that’s the idea…
Hello! You are a political journalist. You write about elections coming up.
You write about elections going down. On rare special days, you even write
about things like “policy” and “governance.”
Also, you seem baffled by the occurrence of slightly unlikely events.
This wasn’t always true. Once upon a time, you saw elections as magical
moments of infinite possibility. You downplayed the likeliest outcome so as to
build suspense, to make it seem as if each race was won on a half-court buzzer-
beater. On election night 2004, when George W. Bush led Ohio by 100,000
votes with barely 100,000 left to count, you said the state was “too close to call.”
When probabilistic models gave Barack Obama a 90% chance to win in 2012,
you dubbed the race “a toss-up.”
Then, 2016 turned your world to pineapple upside-down cake. Donald Trump
defeated Hillary Clinton. Waking up the next day, you felt you’d experienced a
quantum singularity, no more foreseeable than a rabbit materializing out of thin
air. But to probabilist Nate Silver and like-minded others, this was only a
moderate surprise, a 1-in-3 event—like rolling a die that lands on 5 or 6.
Hello! You are an investment banker. You bank investments. You invest in
banks. And your suits are worth more than my car.
As recently as the 1970s, your job was pretty boring. You were a funnel for
funds, pouring money into “stocks” (i.e., pieces of companies) or “bonds” (i.e.,
pieces of debt). Stocks were exciting; bonds were bland. You were a human
bond.
Now, your job is as thrilling as a roller coaster that has failed its safety
inspection. Starting in the 1970s and 1980s, you began inventing complex
financial instruments that no one fully understands, least of all the government
regulators. Sometimes, these devices earn you big bonuses. Sometimes, they
make your century-old firm collapse, leaving a vast crater across the economy as
if a dinosaur-killing asteroid has struck. Exciting times!
To be fair, allocating capital is a pretty important job in capitalism, and has
the potential to create lots of value. It annoys you when snarky math teachers
slight your profession. If so, take a moment to calculate the factor by which your
salary exceeds that of the snarky math teachers, and I think you’ll find your
spirits rising.
Hello! You are a local news anchor. You have great hair, crisp enunciation,
and a long history of artificial banter with your coanchor.
Also, you are obsessed with very unlikely events.
You like to report on chilling dangers. Local murders. Airborne carcinogens.
Defective toys that attach themselves like the creature in Alien to your baby’s
chubby cheeks. Ostensibly, you do this to inform your audience. But be honest.
You do it to hold their attention. If you hoped to raise awareness about the
statistical dangers to children, you’d warn about household accidents caused by
guns and swimming pools. Instead, you paint vivid landscapes in which the only
danger more pressing than abduction is crocodile attacks. You know that we
can’t tear our eyes away from such horrors—especially when they’re leavened
with artificial banter.
Hello! You are a weather forecaster. A televised cloud prophet. You gesture
with conviction, and end every conversation with “And now, back to you.”
Also, you fudge probabilities so people don’t get mad at you.
Of course, you try to be honest. When you say that there’s an 80% chance of
rain, you’re exactly right: It rains on 80% of such days. But when rain is less
likely, you inflate the numbers. You fear angry tweets from people who leave
their umbrellas at home and then blame you for the sky’s damp doings. So when
you forecast a 20% chance of rain, it rains only 10% of the time. You’re
increasing the odds so as to decrease your hate mail.
Perhaps, if people understood probability better, you could speak your truth.
People seem to act as though “10%” means “not going to happen.” If only they
embraced its true meaning (“happening one time in 10”) then you could loosen
your tongue and divulge the numbers of your inner heart. Until then, you remain
a merchant of half-truths.
And now, back to you.
Hello! You are a philosopher. You read strange books, write even stranger
ones, and enjoy walking into a bar alongside a priest and a rabbi.
Also, you do not let probability intimidate you.
Leave “probable” to the empiricists, you say. You seek ideas that are
improbable. You ask questions no one else will ask and think thoughts no one
else will think—mostly because they are pedantic, technical, and probably false.
But this is to your credit! At your best, you birth new fields. Psychology, from
Aristotle to William James, has its ancestry in philosophy. Even at your worst,
your inquiries provoke and refresh us as no one else can.
Hello! You are a Mission: Impossible agent. You dangle from ceilings into
locked vaults, cling via suction cup to skyscrapers, and unmask your true
identity to the double-crossers you’ve just triple-crossed.
Also, you do not seem to get what “impossible” means.
“Impossible” does not mean “as routine as a quarterly earnings report.” Nor
does it mean “very rare” or “rather difficult” or “whew, it’s a good thing that
knife blade halted a millimeter from Tom Cruise’s eye.” It means “not possible.”
And yet it keeps happening. Your film’s title is as dishonest as its theme song is
catchy.
In this, you are not alone. All fiction bends plausibility. One of my favorite
shows is Friday Night Lights, a humble exploration of life in a small Texas
town, depicting the human-scale struggles of ordinary people. Yet even this
paragon of plausibility concludes each football game with once-in-a-lifetime
dramatics: a 90-yard touchdown pass, a goal-line fumble, or a would-be field
goal bouncing off the crossbar. It’s enough to make me wonder: are we
projecting our improbable fantasies onto the TV screen, or is the TV screen
planting those fantasies in us to begin with?
Hello! You are the captain of the Millennium Falcon. You are a ruffian, a
rascal, and a gold-hearted rogue. Your life partner is an 8-foot space dog
wearing a bandolier.
Also, you reject probability altogether.
You’re not a man of sober reflection and strategic consideration. You’re a
contraband smuggler, an empire toppler. You’re a quick-draw Greedo-killing
daredevil for whom doubt is fatal and hesitation is worse. There are no
probabilists in foxholes, and you have spent your life in foxholes. For you,
laborious probabilistic calculations would be as cumbersome as a neurotic
golden robot who keeps saying, “Oh my!” and “If I may say, sir…”
I like to think that there’s a bit of you in each of us. Probability serves us well
in times of cool and careful assessment, but sometimes we need a confidence
that’s unwarranted by the hard quantitative facts. In moments of instinct and
action, the soul leashed to plausibility may shrink from a necessary leap. At such
times, we need to forget the numbers and just fly.
Chapter 14
WEIRD INSURANCE
Despite my best efforts and worst drawings, I have become a grown-up. I drink
unsweetened coffee. I wear unironic ties. Children call me “mister.” Most
discouraging of all, I spend a nonzero percentage of life thinking about
insurance: health, dental, home, auto. Before long, I’ll be insuring my ties
against coffee stains, all without a shred of irony.
Beneath the suffocating weight of adulthood, there’s something funny going
on here. Insurance companies don’t really sell products. They’re vendors of
psychology, offering peace of mind. They let short-run folks like me purchase a
bit of stability from the long run of eternity.
So, for this chapter, let us spit our coffee back in the face of adulthood and
reach for the chocolate milk instead. Let us gaze upon insurance with fresh eyes
and explore the mathematics of this quirky industry, not through its traditional
products, but through a series of peculiar and unconventional plans. Let us
investigate the promise of mutual gain, the danger of exploitation, and the slow
data-driven tug-of-war between the insurer and the insured—all while talking
about college football and aliens.
Even in the darkest depths of grown-up-ness, it’s worth remembering that
insurance is a strange and silly thing. Kind of like adulthood itself.
SHIPWRECK INSURANCE
Ever since there have been ships, there have been shipwrecks. In China five
millennia ago, merchants traded goods downriver in boats. If all went per usual,
then the cargo would arrive, no problem. But if rapids and rocks intervened,
you’d find your shipment rerouted to the riverbed, which can rarely be trusted to
pay up.
Fortune dealt in two cards: moderate success and utter disaster. The former
gave you reason to play the game, but the latter could ruin you.
What was an ancient Chinese trader to do?
Simple: redistribute the danger. Instead of loading one’s own boat with one’s
own goods, merchants would scatter their wares across each other’s ships so that
each vessel carried a mixed cargo. By coming together, they turned the small
likelihood of a high cost into the high likelihood of a small one.
This kind of risk redistribution predates history. You see it in any tight-knit
community where people sacrifice to help their neighbors, knowing they might
one day draw on the same communal kindness. A minor guaranteed cost helps
eliminate an existential risk.
Today, “tight-knit community” has given way to “for-profit insurance
company.” Old-school social mechanisms of enforcement (e.g., not helping the
stingy people) have yielded to precise formula-driven ones (e.g., setting rates
based on risk factors). History has distilled a social arrangement into a
mathematical one.
Still, in principle, insurance today works much as it did 5000 years ago. I find
fellow souls facing the same mortal risks. We exchange pieces of our livelihood.
And we float downriver knowing that now, no shipwreck can sink us.
It’s a common scene: workers pooling their money to buy lottery tickets, hoping
to quit their jobs en masse. They are comrades unified by the dream of no longer
being unified. But suppose they win. What happens to their poor boss, awaking
one morning to the abrupt resignations of an entire department, like some kind of
unnatural disaster?
Fear not, forsaken boss. There’s a cure for what ails you.
In the 1990s, over a thousand British businesses invested in the National
Lottery Contingency Scheme. If your staff won the jackpot, insurance payouts
would help to cover recruitment, retraining, revenue losses, and temporary labor.
(I reckon the government, which runs these lotteries, should sell the
insurance plans, too. It’d be like Godzilla Insurance Co. offering fire-breath
protection.)
As you’d guess, the expected value favors the insurer. What might surprise
you is the margin. The odds of a single ticket winning the jackpot are 14 million
to 1, and yet the payout ratios max out at 1000 to 1. Smaller businesses fare even
worse. With two employees, the expected value remains unfavorable even if
each staffer buys 10,000 tickets.
Of course, insurers can’t just charge the expected value. They’ve got to
consider their own costs, not to mention the risk involved. An insurer that sets its
margins too narrow will go bankrupt the first time a slightly unusual number of
clients demand payouts.
As for the fearful bosses, is there a better option? Well, if you can’t insure
’em, join ’em. Tossing a little money into the pool is an insurance plan in itself:
a small investment that will pay out if (and only if) they win.
MULTIPLE-BIRTH INSURANCE
I view “one” as a formidable number of babies to care for. “Two” sounds like
quartering a child army. “Three” is unimaginable—a childbirth googol. So it’s
dizzying to learn that 1 in 67 pregnancies results in multiple births. In the UK,
expectant parents who share my trepidation can purchase “multiple-birth
insurance.”
You pop out twins, and the insurance company pays out £5000.
One analyst and personal finance adviser, David Kuo, scoffs at this plan,
saying: “You’re better off putting money on the horses.” I see his point: The
expected value here is pretty lousy. The lowest premium is £210, for a mother
with an age below 25 and no family track record of twins. If a hundred such
mothers buy the insurance, then the company will collect £21,000, and (most
likely) pay out just one sum of £5,000. That’s a comfy profit margin, which is
the polite way to say “their customers are paying through the nose.”
Parents with a greater likelihood of twins fare little better. One 34-year-old
mother who was a twin herself paid £700 for coverage. That’s a payout ratio of
just 7 to 1.
If you’ve had fertility treatments, or waited until after the 11-week
ultrasound, then you’re not eligible at all.
It’s the basic problem of all psychological insurance. Why pay good money to
offload a risk that you could handle yourself?
ALIEN-ABDUCTION INSURANCE
In 1987, a fellow in Florida named Mike St. Lawrence rolled out a $19.95 alien-
abduction policy. It provided $10 million in coverage, paid out at a rate of $1 per
year (until completion or death, whichever comes first). It also entitled abductees
to “psychiatric care” and “sarcasm coverage” (“limited to immediate family
members”). The slogan: “Beam me up—I’m covered.” Applicants were rejected
if they answered yes to the question “Do you take this seriously?” or “Were your
parents related before they married?” It was hard to miss the winking.
Then a British company began selling abduction coverage in earnest. Soon,
they had sold 37,000 plans at £100 apiece—and (surprise!) not made a single
payout. That’s nearly £4 million in pure profit. Said one managing partner: “I’ve
never been afraid of parting the feeble-minded from their cash.”
The silly story points toward a real danger. Psychological insurance can
quickly turn predatory. When we buy protection against fear itself, we invite
companies to make us afraid.
FAILED-EXAM INSURANCE
During your mathematics PhD at UC Berkeley, you take a high-stakes qualifying
exam known as “the Qual.” This is a monster you face alone: three hours of oral
questioning from a trio of expert professors. Though preparation is exhausting,
the passing rate is high—at least 90%—because it isn’t scheduled until you seem
battle-ready. The expert professors desire a retake no more than you do.
Qual failure, then, fits the following criteria: (1) it’s rare; (2) it hurts; (3)
many individuals face it; and (4) it strikes more or less at random. That would
seem to make it ripe for insurance.
Some PhD students have hashed out the idea. Each participant pays $5 into
the pool. If anyone fails their exam, then they collect the whole bundle. If
everyone passes, the money goes toward a round of celebratory beers.
I suppose every financial instrument designed by grad students must
culminate, one way or another, in beer. It’s psychological insurance of the best
kind.
There’s nothing more American than a gaudy competition. The $10,000 half-
court shot. The $30,000 dice roll. The $1 million hole in one. When rolling out
such promotional gimmicks, there’s just a teeny-tiny risk.
Someone might win.
As it turns out, this market is a dream for insurance companies. As merchants
of probability, their solvency depends on accurate computation: 50-to-1 payouts
on a 100-to-1 event will keep you in the black; 100-to-1 payouts on a 50-to-1
event will bankrupt you. The actuarial intricacies of home, life, and health
insurance make it easy for an insurer to miscalculate.
But prize giveaways? No problem!
Take the hole in one. Amateur golfers succeed roughly once in every 12,500
attempts. So, for a $10,000 prize, the average golfer costs $0.80. Charge $2.81
per golfer (one company’s advertised premium), and you’ll turn a tidy profit.
The same company insured a $1 million hole-in-one prize for just $300. It’s a
fine deal on both ends: the charity offloads risk, and the expected-value cost to
the insurer is just $80.
The company also insured an NBA fan’s $16,000 half-court shot for $800.
Unless the fan’s odds of success exceed 1 in 50—about what NBA players
manage in game situations—it’s a square deal. Or consider insuring Harley-
Davidson against a $30,000 prize given out to a customer who rolls six lettered
dice to spell out H-A-R-L-E-Y. The expected value is just $0.64 per contestant.
The company charged $1.50. Actuarial math has never been easier.
That’s not to call this type of insurance riskless. In 2007, the Boston-area
retailer Jordan’s Furniture ran a glitzy promotion: buy any sofa, table, bed, or
mattress in April or May, and if the Red Sox go on to win the World Series in
October, then the price will be refunded. The sales approached 30,000 items—a
total of perhaps $20 million.
Then the Red Sox won. The Sox-loving folks at Jordan’s were thrilled;
they’d bought insurance.
The folks at the insurance company? Less delighted.
“CHANGE OF HEART” WEDDING INSURANCE
Please join me now for a thought experiment. A stranger hands you a coin. You
flip it and hide the result. What’s the probability of heads?
INSURANCE-COMPANY INSURANCE
Knowledge gaps scare insurers, but they don’t jolt them awake in a cold sweat.
Their true wide-eyed nightmare is a single whispered word: dependence.
Think back to the Chinese traders, who scattered their merchandise across a
hundred boats. What if shipwrecks don’t come one by one? What if on 99% of
days, no boats sink, and on 1% of days, a storm sinks them all? Then insurance
is useless. We’re not facing individual risks that can be mitigated by
redistribution. We’re all holding the same awful ticket to the same terrible
lottery. Trading pieces of fate does nothing. When we go down, we go down
together.
That’s dependence. For insurers, it’s doom. Hurricane Katrina, for example,
exacted its $41 billion in damage all at once, far exceeding the year’s $2 billion
or so in premiums. In such cases, the insurance company faces the same risk as
the people it’s insuring. The eggs remain all in one basket.
The solution is to scale up: insure the insurance company. This is called
“reinsurance.” Insurers from different regions trade assets, in effect swapping
customers. This lets a company in one location assemble a risk portfolio that’s
spread far and wide.
When it’s not enough to scatter your livelihood across a hundred boats, then
scatter it across a hundred rivers.
HEALTH INSURANCE
I’ve saved for last the weirdest risk-reducing product of all: health insurance.
What’s so strange about it? First, the complexity. Deductibles, coverage
limits, preexisting conditions, co-pays, variable premiums… It’s not quite brain
surgery, but it is the financing of brain surgery, which is perhaps just as
daunting. And second, there’s the vexing question of whether, as our medical
knowledge and predictive powers grow, health insurance can continue to
function at all.
Here’s a simple model of healthcare. We each flip 10 coins. If you flip all
heads, then you contract the dreaded “ten-headed disease,” which will require
$500,000 in medical bills.
Each of us faces a roughly 1 in 1000 risk of disaster. But if we all chip in
$800 for insurance, then it works out. Per thousand people, the insurance
company collects in $800,000 and pays out just $500,000. They turn a profit; we
hedge against financial ruin. Everybody goes home happy.
But what if we know more? What if, before deciding whether to purchase the
plan, we can peek at the first five coins?
Now, from every 1000 people, roughly 970 of us glimpse a tail and sigh with
relief. We’re safe; no insurance necessary. But the remaining 30 people grow
nervous. One of them probably has the dreaded disease. Between them, those 30
can expect a collective $500,000 in expenses. Even if they redistribute the costs
equally, each remains on the hook for thousands and thousands of dollars.
Insurance no longer provides peace of mind; it drains it.
When we looked at the coins, we reduced our uncertainty, and without
uncertainty, insurance collapses. If we know in advance who will suffer—whose
boat will sink, whose employees will win the lottery, who will suffer that NFL-
dream-crushing injury—then insurance becomes impossible. You can only pool
risk among people who share it.
In our medical system, this problem is unfolding year by year. Genetic testing
and improved statistics threaten the basic logic of insurance.
I see no easy solution. If you individualize the rates, then some folks pay
pennies while others face premiums almost as steep as the medical bills
themselves. But charging everyone the same amount turns a project of mutual
benefit, with each of us hedging against risk, into a project of collective charity,
with some folks subsidizing others. That’s a harder sell. It’s one reason why
American healthcare remains so contentious.
As a teacher, I’m inclined to think of all knowledge as a gift. But insurance
complicates that story. Our ignorance of fate can force us to collaborate against
it. We have built a democracy from our uncertainty—and new knowledge
threatens that balance, as sure as flood or fire.
Chapter 15
HOW TO BREAK THE ECONOMY WITH A SINGLE PAIR
OF DICE
In September 2008, I began my final year of the great free-pizza scavenger hunt
called “college.” Knowing, if not quite believing, that after college only wage
earners could acquire pizza, I decided to attend the annual career fair. Booths of
employers would fill the gymnasium to offer free giveaways and (secondarily)
job applications.
But when I arrived, I found a half-empty gymnasium, a ghost town. In abrupt
unison, the investment banks had all decided that maybe now wasn’t such a good
time to hire.
We knew why. In the prior month, the global financial system had frozen,
tossed up a blue screen of death, and was now refusing to reboot. Wall Street
had devolved into the final scene of a Shakespearean tragedy: century-old
institutions lay heaped in the dirt, swords sticking out of them, wheezing death
monologues. Journalists batted around terms like “worst,” “recession,” and
“since the Great Depression,” often consecutively. Even the pizza crusts tasted
of anxiety.
In this chapter, we come to our final lesson on probability, perhaps the
hardest. Many would-be probabilists fall for the seductive idea of independence,
choosing to imagine our world as an aggregation of isolated events. But if
probability is to confront the world’s uncertainty, it must face the world’s
interconnectedness: its narrative threads, its causal chains.
A simple illustration: What’s the difference between rolling two dice and
doubling a single die?
Well, in each case, the sums range from a minimum of 2 (snake eyes) to a
maximum of 12 (double sixes).
With two independent dice, extreme values can unfold only a few ways. (For
example, only two combinations will yield a 3). Middle values can unfold in
many ways; for example, six different combinations will yield a 7. Thus, middle
values are more likely.
What about rolling just one die and doubling it? Now the “second” roll
depends entirely on the first; we’ve got a single event disguised as two. And so
the extreme outcomes are just as likely as the middle ones.
The difference is stark. Independence irons out extremes; dependence amplifies
them.
We can push it further. Let’s move from two dice to a million. Now, the
outcomes range from a minimum of 1 million (all ones) to a maximum of 6
million (all sixes).
What if each die acts alone, independent of the other 999,999? Then we find
ourselves in the long-run world of stable tendencies, where the exciting 6’s and
the disappointing 1’s arrive in equal proportions. The overwhelming likelihood
is that our total for all the dice will land smack in the middle, far from both
extremes. There’s a 99.9999995% probability we’ll land between 3.49 million
and 3.51 million. Getting the minimum of 1 million is all but impossible, its
probability less than 1 in a googol googol googol googol. (I could write the word
“googol” another 7000 times, but you get the idea.)
But what if we’re not rolling separate dice? What if we’re rolling a single die
and multiplying it by a million, reflecting it through a hall of mirrors? Well, then
we remain in the chaotic world of random chance. The later “rolls” depend
wholly on the first, providing no balance. Getting the minimum is no longer an
absurd when-pigs-fly proposition. Its probability is 1 in 6.
If you answered A, then you work for Wall Street. (Hey, nice suit! Is that
Italian?) If you answered B, then I’m honored that you’re reading my book,
Senator Sanders. And if you answered C, then you’re already familiar with a key
theme of this chapter: that the basic function of the financial sector is to decide
what things are worth. Stocks, bonds, futures, Rainbow Shout® contracts,
standard Parisian barrier options, single-use monocular default swaps… whether
you’re buying, selling, or Googling to see which of those I made up, you want to
know what these things are worth. Your livelihood depends on it.
The problem, of course, is that pricing ain’t easy.
Take bonds. These are pieces of debt, promises that you’ll get paid back.
Consider somebody who borrowed money to buy a house and promises to pay
back $100,000 in five years.
What’s this IOU worth to you?
Well, we begin with Pricing Challenge #1: putting a value on time. In the
carpe diem logic of finance, a dollar today is worth more than a dollar tomorrow.
First, there’s inflation (which gradually lowers the value of a dollar), and second,
there’s opportunity cost (i.e., a wisely invested dollar today will be worth more
by next year). As a crude estimate, we’ll say a dollar today is worth $1.07 next
year. Carry that forward, multiplying year after year, and you’ll find that $1
today is equivalent to $1.40 in five years.
Who knows what Magritte meant by it? This is the man who once painted a
bird grabbing a lady’s shoe and titled it God Is No Saint. But I think he’s
challenging our idea of the home as an emblem of security. Instead—and I don’t
mean to spook you—the house is something chancy and precarious, an
existential risk. Your house will likely be the largest investment of your lifetime,
costing several times your annual salary, incurring a debt that will last a
generation. The house is an icon not of stability but of probability.
Seven decades after Magritte’s visual pun, Wall Street faced its own little
problème de la maison: pricing CDOs. The problem was to identify the
relationship between the various mortgages. Last I checked, you and I are
separate people. So if I default, perhaps that has no bearing on you. On the other
hand, we share the same economy. We can escape a big recession no more than
we can escape summer pop songs. So if I default, perhaps that suggests you’re in
danger too. In Wall Street terms, the question was whether defaults pose an
idiosyncratic risk or a systematic one.
Are houses like thousands of separate dice, each roll independent of the
others? Or are they like a single die, multiplied thousands of times by a hall of
mirrors?
Imagine a CDO (tiny by real-world standards) built from 1000 of the
mortgages we explored above. Since we priced one at $50,000, the whole bundle
should be worth $50 million.
Now, if the mortgages are independent, Wall Street can sleep easy. Sure, our
investment might come in $1 million below what we expected, but it would take
absurd misfortune to lose $2 million, and a loss of $5 million is unthinkable, its
probability less than 1 in 1 billion. Independence stabilizes, all but eliminating
the chance of catastrophic loss.
Meanwhile, if all the houses represent a single roll of the dice, then Wall
Street shall suffer night sweats and scream dreams. Dangers unthinkable a
moment ago are now very, very thinkable. In this deal, there’s a horrifying 1-in-
3 chance we lose nearly half of our investment, and a bone-chilling 1-in-10
chance that we lose everything.
Of course, neither vision quite matches reality. We are neither hive-minded
drones living in perfect sync, nor rugged individualists unaffected by our
neighbor’s weather. Instead, our lives are somewhere in between, our futures
delicately intertwined. It seems obvious that one default heightens the likelihood
of another—but by how much, and under what conditions? These are the subtlest
challenges that probabilistic models can face.
Finally, if you know your statistical vocabulary, alarm bells start ringing at
the term “Gaussian.” In mathematics, the word arises whenever you add together
lots of independent events. But that’s the whole problem: these events weren’t
independent.
In all these ways, Wall Street overlooked the risk that Magritte saw, leading
to a calamity as surreal as anything the painter ever envisioned. And still, we
haven’t discussed the worst of it. The mispriced CDOs amounted to a few
trillion dollars—enough to damage the economy but not enough to explain the
body blow it suffered in September 2008. What left the economy collapsed on
the mat, dizzy and spitting teeth, wondering in bewilderment where that sucker
punch had come from? It was a probabilistic failure on an even grander scale.
A quick recap. CDOs were meant to achieve the stability of a million dice;
instead, they embodied the fragility of a single die. The CDSs doubled down
until the gamble loomed large enough to imperil the entire world economy. It all
prompts a natural question:
How could Wall Street have been so stupid?
You harvest the greatest minds from the fanciest universities, buy them
million-dollar supercomputers, pay them cosmic salaries, work them 90 hours a
week… and then walk in to find them screaming and jamming forks in the wall
sockets?
I wish I could write off these “independence vs. dependence” errors as a one-
time aberration, a peculiar circumstance of CDOs and CDSs. But if wishes were
houses, the CDOs might have stood a chance. The grim fact is that this error
runs to the core of financial markets.
5. ASHES, ASHES, WE ALL FALL DOWN
At the risk of being dubbed a neoliberal shill, I will state my belief that markets
work pretty well. Heck, I’ll go further: they work really well.
For example, this planet happens to harbor delicious fruit spheres called
“apples.” How should we allocate them? If farmers grow more than consumers
want, we’ll find piles of Red Delicious rotting in the streets. If consumers want
more than farmers grow, we’ll witness strangers clawing at one another to snag
the last McIntosh during a shortage. Yet somehow, against all odds, we manage
to grow just about the right number of apples.
The trick? Prices. Although we think of prices as determining our behavior
(“too expensive, so I won’t buy”), the reverse is just as true. Each of our
individual choices exerts a tiny influence on the price. If enough people refuse to
buy, the price will drop. If enough refuse to sell, the price will rise. The price
functions as an aggregate of all our independent judgments and decisions.
And so, like other aggregates of independent events, prices tend to yield
balanced, stable, reasonable outcomes. Aristotle called it “the wisdom of the
crowd.” Adam Smith called it “the invisible hand.” I call it “the independent
dice all over again, except this time, the dice are us.”
In theory, what works for apples ought to work for CDOs. Some folks will
overvalue them. Others will undervalue them. But in the end, a market full of
independent investors will drive the price to a stable equilibrium.
There’s just one problem: all too often, investors behave less like millions of
independent dice, and more like a single die multiplied by millions.
Take the 1987 stock crash. On October 19, prices nose-dived, plummeting
more than 20%. It came without warning: no market-shaking news, no high-
profile bankruptcy, no “Oh jeez, I’m in over my head” speech from the Fed
chair. Markets just collapsed. Only later did the postmortem identify a peculiar
trigger: lots of Wall Street firms had relied on the same basic theory of portfolio
management. Many even employed the same software. As markets slipped, they
sold the same assets in unison, sending prices into a downward spiral.
Portfolio management’s whole purpose is to bring safety through
diversification. But if everyone diversifies in exactly the same way, then the
resulting market ain’t so diverse.
If investors exercise their own judgment, then day-to-day price changes
should follow a bell-shaped normal distribution: up a little on some days, down a
little on others, but almost never leaping too fast or far. Alas, reality disagrees.
Market movements, with their occasional massive drops, follow more of a
power-law distribution. That’s the same mathematical model we use for
earthquakes, terrorist attacks, and other massive disruptions to highly sensitive
systems.
The market isn’t random like lots of dice added together. It’s random like an
avalanche.
Leading up to the financial meltdown of 2008, many banks relied on the
same limited set of models (such as the Gaussian copula). Rather than bringing
fresh insight, they herded around a single strategy. Even ratings agencies—
whose purpose and duty is to offer independent analysis—just parroted the
banks’ own claims. The referees became cheerleaders.
Why, back in September 2008, did I arrive to find a somber half-full
gymnasium? Why, in other words, did the financial system melt down?
Well, it’s complicated. Incompetence played its part, as it does in most
failures. (Just watch me bake sometime.) So did misaligned incentives, blind
optimism, naked greed, dizzying complexity, governmental dysfunction, and
interest rates. (Again, see my baking.) This brief chapter has told only a sliver of
the story, focused around a specific theme: the dangerous assumption of
independence, when dependence actually reigns.
Mortgages default together. CDSs pay out together. And actors in the
marketplace herd together around similar pricing strategies.
You want to break the economy with a single pair of dice? Honestly, it’s
easy. Just convince yourself that you’re rolling a million pairs, and then bet your
fortune on the roll.
IV
STATISTICS
Okay, let’s get this out of our system. Statistics are lies. Not to be trusted. All of
history’s wittiest people have said so. Haven’t they?
FAMOUS QUOTE I FOUND BY GOOGLING: “There are three kinds of lies: lies,
damned lies, and statistics,” said Mark Twain.
REALITY I FOUND BY DEEPER GOOGLING: People today often misattribute this to
Twain, which is perhaps fair, since Twain himself misattributed it to
Benjamin Disraeli. The correct attribution is unknown.
FAMOUS QUOTE I FOUND BY GOOGLING: “Do not trust any statistics you did not
fake yourself,” said Winston Churchill. (Or perhaps: “I only believe in
statistics that I doctored myself.”)
REALITY I FOUND BY DEEPER GOOGLING: A lie meant to slander Churchill. It may
have originated with Nazi propagandist Joseph Goebbels.
FAMOUS QUOTE I FOUND BY GOOGLING: “There are two kinds of statistics, the kind
you look up and the kind you make up,” said Rex Stout.
REALITY I FOUND BY DEEPER GOOGLING: Rex Stout was a novelist. He didn’t say
this; one of his characters did.
FAMOUS QUOTE I FOUND BY GOOGLING: “Politicians use statistics in the same way
that a drunk uses lampposts—for support rather than illumination,” said
Andrew Lang.
REALITY I FOUND BY DEEPER GOOGLING: Okay, this one is real, and it’s great.
1. THE MEAN
How It Works: Add up all the data you’ve got. Divide the total by the size
of the data set.
When to Use It: The mean (or “average,” as it’s often known) fills a basic
need in statistics: to capture the “central tendency” of a group. How tall is that
basketball team? How many ice-cream cones do you sell per day? How did the
class do on the test? If you’re trying to summarize a whole population with a
single value, then the mean is a sensible first port of call.
Why Not to Trust It: The mean considers only two pieces of information:
the total, and the number of people contributing to that total.
If you’ve ever split a hoard of pirate treasure, then you see the danger: There
are many ways to share a haul. How much does each individual contribute? Is it
balanced or grossly one-sided? If I eat a whole pizza and leave nothing for you,
is it fair to claim that our “average consumption” was half a pizza? You can tell
your dinner guests that the “average human” has one ovary and one testicle, but
will this not bring conversation to an awkward standstill? (I’ve tried; it does.)
Humans care about questions of allocation. The mean shrugs them off.
There’s a saving grace, which is that this feature makes the mean simple to
calculate. Say your test scores are 87, 88, and 96. (Yeah, you’re crushing this
class.) What’s your average? Don’t overheat your neurons with addition and
division; instead, reallocate. Take six points from your last test; give three to the
first and two to the second. Now your scores are 90, 90, and 90, with a single
point left over. Sharing that lonely point among the three tests, you arrive at a
mean of 90¹/³, with no brain strain required.
2. THE MEDIAN
How It Works: The median is the middlemost member of your data set. Half
your data lies below it, and half above.
When to Use It: The median, like the mean, captures the central tendency of
a population. The difference lies in its sensitivity to outliers—or rather, its
insensitivity.
Take household income. In the US, a wealthy family may earn dozens (even
hundreds) of times as much as a poor one. The mean, pretending that every
family has an equal share of the total earnings, is seduced by these exceptional
values, and is led away from the bulk of data points. It arrives at the figure of
$75,000.
The median resists the outliers’ gravitation. It names instead the nation’s
absolute middle household income, the perfect midpoint where half of families
are wealthier and half poorer. In the US, it’s near $58,000. Unlike the mean, this
offers a clean picture of the “typical” household.
Why Not to Trust It: Once you’ve found your median, you know that half
of the data lies above and half below. But how far away are those points—a
hair’s breadth from the median or a transcontinental flight? You’re looking only
at the middle slice of pie, never mind how big or small the other slices are. This
can lead you astray.
When a venture capitalist invests in new businesses, she expects most of
them to fail. The rare 1-in-10 home run compensates for all those little losses.
But the median misses this dynamic. “The typical outcome is negative,” it
squawks. “Abort mission!”
Meanwhile, an insurance company builds a careful portfolio, knowing that
the rare 1-in-1000 disaster can wipe out years of modest profits. But the median
overlooks the potential for calamity. “Hey, the typical outcome is positive,” it
cheers. “Never stop!”
This is why you’ll often find the median displayed alongside the mean. The
median reports on the typical value; the mean, on the total. Like two imperfect
witnesses, they tell a fuller story together than either can alone.
3. THE MODE
How It Works: It’s the most common value, the hippest, most fashionable
data point.
What if each value is unique, with no repetitions? In that case, you can group
the data into categories, and call the most common one “the modal category.”
When to Use It: The mode shines in conducting polls and in tabulating
nonnumerical data. If you want to summarize people’s favorite colors, you can’t
very well “total the colors up” to compute a mean. Or, if you’re running an
election, you’ll drive citizens mad if you line up the votes from “most liberal” to
“most conservative” and award the office to the recipient of the median ballot.
Why Not to Trust It: The median ignores the total. The mean ignores its
allocation. And the mode? Well, it ignores the total, its allocation, and just about
everything else.
The mode seeks a single most common value. But “common” does not mean
“representative.” The modal salary in the United States is zero—not because
most Americans are broke and jobless, but because wage earners are spread
across a spectrum from $1 to $100 million, whereas all wageless people share
the same number. The statistic doesn’t illuminate anything about the US. It’s
true in virtually every country, an artifact of how money works.
Turning to “modal categories” can only partway solve the problem. It places
a surprising power in the hands of the person presenting the data, who can
gerrymander the category boundaries to suit their agenda. Depending on how I
draw the lines, I can claim that the modal household in the US earns $10,000 to
$20,000 (going by increments of 10,000), or $20,000 to $40,000 (going by
increments of 20,000), or $38,000 to $92,000 (going by tax brackets).
Same data set, same statistic. And yet the portrait changes completely,
depending on the artist’s choice of frame.
4. THE PERCENTILE
How It Works: Recall that the median splits your data set right down the
middle. Well, the percentile is a median with a dimmer switch. The 50th
percentile is the median itself (half the data above and half below).
But you can pick other percentiles, too. The 90th lands near the top of your
data: Only 10% sits above while 90% sits below. Meanwhile, the 3rd resides
near the bottom of your data set: only 3% is below, with 97% above.
When to Use It: Percentiles are handy, flexible, and perfect for that favorite
human pastime: ranking things. That’s why standardized tests often deliver their
scores as percentiles. A raw score like “I got 72% of questions right” isn’t very
informative. Were these vicious snakebite questions or slow-pitch softballs?
However, “I’m at the 80th percentile” reveals right where you stand. You did
better than 80% of test takers and worse than 20%.
Why Not to Trust It: Percentiles share the median’s Achilles heel. They tell
you how much data lies above or below a certain point but not how far away it
is.
Take the finance industry, which employs percentiles to measure the
downside of an investment. You first imagine the spread of possible outcomes,
from triumph to cataclysm, and then pick a percentile (usually the 5th), which
you call “value at risk,” or VaR. It aims to capture the worst-case scenario. But
in fact, you’ll do worse exactly 5% of the time. VaR gives no clue of how much
worse, whether it’s pennies of extra losses, or billions.
You can better visualize the landscape of possibilities by checking more VaR
percentiles—for example, 3, 1, and 0.1—but by nature, a percentile cannot
detect the most violent and extreme losses, and so the true worst-case scenario
will always lurk just beyond sight.
5. PERCENTAGE CHANGE
How It Works: Don’t just report a change; first, divide it by the original
total.
When to Use It: Percentage change is all about putting things in perspective.
It frames gains and losses as a proportion of the whole.
Take a gain of $100. If I only had $200 to begin with, then this windfall
marks 50% growth, and I’m dancing a hallelujah Snoopy dance. But if I already
had $20,000, then my new income represents just 0.5% growth; I’ll content
myself with a single triumphant fist pump.
Such perspective taking is crucial when you’re watching a quantity grow
over time. If the Americans of 70 years ago had heard that our GDP increased by
$500 billion last year, they’d be awed. If they’d heard that it grew by 3%, they’d
feel right at home.
Why Not to Trust It: Hey, I’m all about perspective. But percentage change,
in its effort to provide context, can annihilate it instead.
When I lived in the UK, the tasty £2-per-jar tomato sauce sometimes went on
sale for £1. Those days felt like jackpots: 50% savings! I’d lug home a dozen
jars, enough for months of ravioli. Then, later, I’d find myself buying plane
tickets to attend a wedding in the US. Delaying the purchase a week might incur
a price bump of 5%. “Ah well,” I’d say as I accepted the higher rate. “It’s only a
little more.”
You can see the problem: my instincts run penny-wise and pound-foolish.
The “huge” discount on tomato sauce saved me £12 while the “modest” increase
on plane tickets cost me £30. A dollar is a dollar is a dollar, whether it’s part of a
$20 grocery bill or a $200,000 mortgage. Big price drops on cheap products are
dwarfed by a small bump on a rare extravagance.
6. THE RANGE
How It Works: It’s the distance between the biggest and the smallest data
point.
When to Use It: The mean, median, and mode deal with “central tendency”:
they aim to collapse a diverse population down to a single representative value.
The “range” takes the opposite goal: not to sweep disagreement under the rug,
but to quantify and display it, to give a measure of the data’s “spread.”
The range’s merit lies in its simplicity. It conceives of a population as a
spectrum running from “smallest” to “largest,” and it tells us the width of that
spectrum. It’s a quick-and-dirty summary of variety.
Why Not to Trust It: The range cares only about the biggest and smallest
slice of cake. That omits an awful lot of crucial information: namely, the sizes of
all those middle slices. Are they close to the maximum? Close to the minimum?
Scattered evenly in between? The range neither knows nor cares to ask.
The larger the data set, the more dubious the range becomes, as it disregards
millions of intermediate values to inquire about the two most extreme outliers. If
you were an alien who learned that there’s a 7-foot range of adult human heights
(from the record low of less than 2 feet to the high of almost 9), you’d be quite
disappointed to visit Earth and meet all of us dull 5-to-6-footers.
If you take a square root at the end, you’ll get the “standard deviation.” It’s
the more intuitive measure, since variance has strange squared units. (What’s a
“square dollar”? No one knows.)
Since variance and standard deviation go together, that’s how I’ll discuss
them.
When to Use It: Like the range, variance and standard deviation quantify a
data set’s variety, but—and I say this with the impartiality of a loving parent—
they’re better. The range is a quick, makeshift measure of spread; the variance is
a load-bearing pillar of statistics. By taking a contribution from every member of
the data set, the variance achieves the sophistication of a symphony, in contrast
to the range’s two-note ditty.
The variance’s logic, though convoluted, makes sense on inspection. It’s all
about the data’s distance from the mean. “High variance” means that the data is
scattered far and wide; “low variance” means that it’s huddled close together.
Why Not to Trust It: Sure, the variance takes a contribution from every data
point. But you can’t tell who is contributing what.
In particular, a single outlier can give the variance a huge boost. Thanks to
the squaring step, a single large distance (e.g., 122 = 144) can yield a greater
contribution than a dozen small ones (e.g., 32 = 9; 12 such terms total to just 96).
Variance has another feature that stumps lots of folks. (It’s not a bad trait,
just a counterintuitive one.) Students tend to consider a data set with many
different values (e.g., 1, 2, 3, 4, 5, 6) as more “spread out” than a data set with
repeated values (e.g., 1, 1, 1, 6, 6, 6). But variance isn’t interested in “variety”; it
cares only about distance from the mean.
In variance’s eyes, the spread of the latter set (with its repeated far-from-the-
mean values) outweighs the spread of the former (with its nonrepeated but
closer-to-the-mean values).
8. THE CORRELATION COEFFICIENT
How It Works: A correlation measures the relationship between two
variables. A person’s height and weight, for example. Or a car’s price and its
number sold. Or a movie’s budget and box office receipts.
The scale runs from a maximum of 1 (“wow, they totally go together”) to a
midpoint of zero (“huh, there’s no connection here”) to a minimum of -1 (“hmm,
they go opposite directions”).
That’s the quick-hit summary, anyway. To see how the correlation coefficient
actually works, check the endnotes.
When to Use It: Are wealthier nations happier? Does “broken windows”
policing prevent crime? Does drinking red wine prolong life spans, or just dinner
parties? All these questions deal with connections between pairs of variables,
between imagined causes and conjectured effects. Ideally, you’d answer them by
running experiments. Give 100 people red wine, give 100 grape juice, and see
who lives longer. But such research is slow, expensive, and often unethical.
Think of that piteous wine-deprived control group.
Correlation lets us tackle the same question sideways. Find a bunch of
people, measure their wine intakes and life spans, and see if the wine drinkers
live longer. Admittedly, even a strong correlation can’t determine causation.
Maybe wine lengthens life. Maybe long life drives people to drink. Maybe both
are driven by some third variable (e.g., rich people both live longer and can
afford more wine). It’s impossible to know.
Even so, correlational studies offer a great starting point. They’re cheap and
fast, and they allow for large data sets. They can’t pinpoint causes, but they can
offer tantalizing clues.
Why Not to Trust It: The correlation coefficient is one of the most
aggressive statistical summaries of all. It collapses hundreds or thousands of data
points, each with two measured variables, into a single number between -1 and
1. Suffice it to say, that’s going to leave some things out, a fact illustrated by the
mathematical oddity known Anscombe’s quartet.
Let us step into Anscombe Academy for Witchcraft and Wizardry, where
students have spent weeks preparing for exams in four classes: Potions,
Transfiguration, Charms, and Defense Against the Dark Arts. For each exam, we
shall consider two variables: the hours each student spent studying and that
student’s score (out of 13) on the exam.
From the summary stats, you’d think the four tests were identical:
The Charms test follows an even more certain pattern: studying improves
your scores, but with diminishing marginal returns. When you reach 10 hours,
extra study time begins to harm your score (perhaps by costing you sleep).
However, correlation is built to detect linear relationships, and so it misses the
nature of this quadratic pattern, resulting in a coefficient of 0.816.
Finally, in Defense Against the Dark Arts, every pupil studied for eight
hours, meaning that study time can’t help you predict the final score. There’s
just one exception: a hardworking outlier whose 19 hours of study paid off with
the runaway best score. That single data point suffices to strengthen the
correlation from 0 all the way up to…
0.816.
Each test follows its own distinct logic, obeying its own unique pattern. But
the correlation coefficient misses the story.
Then again, that’s the nature of statistics. As I like to say:
Feel free to quote me. Or, if you favor the traditional method, just make up
your own saying and attribute it to me.
Chapter 17
THE LAST .400 HITTER
Ever since its birth, baseball has been a game of numbers. Today, Wikipedia
lists 122 types of baseball stats, from DICE to FIP to VORP, and I suspect these
only scratch the surface. Go ahead: type three random letters; I wager that
someone, somewhere, keeps a fastidious record of that statistic.
This chapter will explore a single statistic, from its humble origin to its
gradual demise. That statistic is BA: Boston accents. Sorry, BA: beer absorbed.
Okay, fine, it’s BA: batting average.
Once upon a time, batting average reigned supreme. These days, statisticians
treat it like a piece of kitsch, a relic of simpleminded days. Is it time to put the
batting average out to pasture? Or does the old veteran with the aching joints still
carry a spark of magic?
1. BOXED LIGHTNING
In 1856, an Englishman named Henry Chadwick, working as a cricket reporter
for the New York Times, happened to attend his first baseball game. It caught his
fancy. “In baseball all is lightning,” he gushed, as only a cricket fan could. Like
a sloth smitten with the flash and pizzazz of a tortoise, Chadwick soon devoted
his life to this American pastime. He served on rules committees, wrote the
sport’s first book, and edited its first annual guide. But what earned Chadwick
his nickname as “the Father of Baseball” was something more elemental: the
stats.
Chadwick invented the “box score,” a table for tracking the game’s key
events. By scanning columns of numbers—runs, hits, putouts, and so on—you
could almost watch the innings unfold. Box scores weren’t about long-term
predictive power or statistical validity. They were about credit and blame, heroes
and villains, a story told in numbers. They summarized weather conditions and
highlighted key events, aiming (in a time before radio, action photography, or
MLB.com) to help you glimpse the action. It was SportsCenter for the 1870s.
Chadwick’s idea for “batting average” came from cricket, where there are
only two bases and you score a run every time you move from one to the other.
Cricketers keep batting until they make an out, by which time a good player will
have scored dozens of runs. (The all-time record is 400.) Thus, cricket’s batting
average is defined as “runs scored per out made.” A great player can maintain an
average of 50, even 60.
This definition made little sense in baseball, where a single hit ends your at-
bat. So, like any good mathematician, Chadwick played with the rules, trying out
a few definitions before settling on the one used today.
Batting average aims to capture your rate of success with a simple fraction:
your number of successful hits divided by the number of hits plus the number of
outs you make. Chadwick called it the “one true criterion of skill at the bat.”
In theory, batting averages range from .000 (never hitting) to 1.000 (never
failing to hit). In practice, almost all players cluster between .200 and .350.
That’s not much of a gap. Baseball’s royals (hitting 30% of the time) and its
peasants (hitting 27.5%) differ by just one hit in every 40 attempts. The naked
eye can’t tell the difference. Even over a whole season, a “worse” hitter might
outshine a “better” one by sheer luck.
Baseball, perhaps unique among sports, deals you enough cards to have a real
shot. In every 162-game season, a team’s batters face roughly 24,000 pitches.
For soccer to offer the same wealth of discrete data points, you’d have to reset
the ball at center field every five seconds, all season long. Better yet, whereas
other team sports unfold in multiperson melees, each baseball player bats alone,
his data clean and separable from his teammates’.
That’s the glory of batting average. But every statistic leaves something out
—and in this case, what’s missing is something crucial.
The formula itself was scarcely grammatical. The equals sign doesn’t really
mean “equals,” nor is the minus sign actual subtraction. Still, the article gave a
penetrating critique of several “old baseball ideas”—batting average chief
among them. The attack (credited to Rickey but ghostwritten by Canadian
statistician Allan Roth) begins with two letters: BB, for “base on balls.” Or, in
common parlance, a walk.
As baseball came of age in the 1850s, a player batted until he either hit the
ball into play, or swung and missed three times. With patient enough batters, the
game flowed like cold molasses. Thus, in 1858, “called strikes” were born. If a
player let a juicy pitch sail past, then it was treated as equivalent to swinging and
missing. But now the pendulum swung too far; cautious pitchers refused to
throw anything hittable. The solution, introduced in 1863, was to also call
“balls”: pitches deemed too off target for the batter to hit. Enough such “balls”
would grant the batter a free walk to first base.
Walks stumped Chadwick. Cricket’s closest equivalent is a “wide,” generally
viewed as a mistake by the thrower. So batting average ignored walks, as if the
at-bats had never occurred. Walks weren’t deemed an official statistic until
1910.
Today, the most skilled and patient hitters walk 18% or 19% of the time;
their reckless, swing-happy peers, just 2% or 3%. Hence the first term in
Rickey’s equation: a convoluted expression for what we now call “on-base
percentage,” or OBP. It’s your rate of getting on base, via either hits or walks—
in other words, your rate of not making an out.
Which statistic better predicts the number of runs a team will score: BA or
OBP? Running the correlations for 2017, BA is pretty strong, with a coefficient
of 0.73. But OBP is outstanding, with a coefficient of 0.91.
Next, Rickey (i.e., Roth) highlighted another shortcoming of batting average.
Hits come in four flavors, from a single (one base) to a home run (four bases).
As with desserts or retweets, more is better—but batting average treated them all
the same. Hence, the second term in Rickey’s equation, which amounts to a
measure of “extra bases” beyond first.
Today, we prefer a related statistic: “slugging percentage,” or SLG. It
calculates the average number of bases achieved per at-bat, from a minimum of
0.000 to a theoretical maximum of 4.000 (a home run every time). In practice, no
hitter has ever topped 1.000 for a whole season.
Like batting average, SLG ignores walks, and like any stat, it collapses
meaningful distinctions. For example, to slug .800 over 15 at-bats, you need to
get 12 total bases (since 12/15 = 0.8). There are a lot of ways to do that, not all
of them equal in value:
Since OBP and SLG each shed light on a different corner of the game,
they’re often used in conjunction. One common shortcut is to add the two,
arriving at a statistic known as “on-base plus slugging,” or OPS. In the 2017
data, its correlation with runs scored is an eye-popping 0.935—better than either
OBP or SLG alone.
On the 50th anniversary of the Life article, the New York Times showed the
formula to New York Yankees general manager Brian Cashman. “Wow!” he
said. “The guy was generations ahead of his time.” His praise belies the damning
truth: Cashman had never heard of the Life piece. Even after its publication,
batting average would reign for decades more, while OBP and SLG huddled in
the shadows. The baseball dialogue in The Old Man and the Sea probably had
more impact on the game than Rickey’s manifesto.
What was baseball waiting for?
The final two games of the season fell back-to-back on September 28. The
evening before, Williams prowled 10 miles across the streets of Philadelphia,
unable to sleep. Prior to the first game, according to one sportswriter, he “sat on
the bench, biting his fingernails. His mammoth hands trembled.” He later
reported “shaking like a leaf when [he] went to bat the first time.”
But the 23-year-old persevered. That afternoon, he managed six hits in eight
chances, boosting his average to .4057. (The headline writers didn’t quibble
about calling this .406.) Nearly 80 years have elapsed, and no one has hit .400
since.
In 1856, Henry Chadwick glimpsed a dusty, bare-handed game, an excuse for
a summer afternoon. The Englishman gave it numbers. Numbers gave it reach. A
century and a half later, baseball is a corporate triumph, with team payrolls
running in the hundreds of millions. The 19th-century batting average struggles
to keep pace in a 21st-century game, like a boy trying to field a scorching line
drive with his naked hands.
Still, against all the odds, .400 retains its magic. In April and May, when the
season is young and the sample sizes are as small as spring buds, you can often
find a player or two flirting with the .400 line. Before long, they’ll fade. But for
a week or so there’s a breath of hope across the land, a sense that mythical
creatures like dragons and .400 hitters have not yet vanished. The .500 OBP and
the .800 SLG will never stir the heart in quite the same way. We love .400
neither for its predictive power nor for its mathematical elegance, but because it
carries an electric charge, because it tells a story in three decimal places—four, if
you’re feeling picky.
Maybe no one will ever hit .400 again. Maybe it will happen next year. For
his part, Williams shrugged it all off. “If I had known hitting .400 was going to
be such a big deal,” he said 50 years later, “I would have done it again.”
Chapter 18
BARBARIANS AT THE GATE OF SCIENCE
Or that after plotting two close-together points on graph paper, you feel
emotionally closer to your family members than after plotting two far-apart
points?
Or that adopting certain “power poses” can inhibit stress hormones while
boosting testosterone, prompting others to judge you as more confident and
impressive?
I’m not making these up. They’re real studies, performed by real scientists,
wearing real lab coats and/or jeans. They’re grounded in theory, tested by
experiment, and vetted by peers. The researchers obeyed the scientific method
and hid no rabbits up their sleeves.
Yet these three studies—and dozens more, in fields as far-flung as marketing
and medicine—are now in question. They may be wrong.
Across the sciences, we are living through an epoch of crisis. After decades
of the best science they could muster, many scholars now find their life’s work
hanging in the balance. The culprit isn’t dishonesty, or a lack of integrity, or too
many passages arguing against free will. The illness runs deeper, all the way
down to a single statistic at the heart of the research process. It’s a figure that
made modern science possible—and that now threatens its stability.
1. TO CATCH A FLUKE
Every science experiment asks a question. Are gravitational waves real? Do
millennials hate solvency? Can this new drug cure antivax paranoia? No matter
the question, there are two possible truths (“yes” and “no”) and, given the
inherent unreliability of evidence, two possible outcomes (“you get it right” and
“you get it wrong”). Thus, experimental results can fall into four categories:
Scientists want true positives. They are known as “discoveries” and can win
you things like Nobel Prizes, smooches from your romantic partner, and
continued funding.
True negatives are less fun. They’re like thinking you’d tidied the house and
done the laundry, only to realize that, nope, that was just in your head. You’d
rather know the truth, but you wish it were otherwise.
By contrast, false negatives are haunting. They’re like looking for your lost
keys in the right place but somehow not seeing them. You’ll never know how
close you were.
Last is the scariest category of all: false positives. They are, in a word,
“flukes,” falsehoods that, on a good hair day, pass for truths. They wreak havoc
on science, sitting undetected in the research literature for years and spawning
waste-of-time follow-ups. In science’s never-ending quest for truth, it’s
impossible to avoid false positives altogether—but it’s crucial to keep them to a
minimum.
That’s where the p-value comes in. Its whole purpose is to filter out flukes.
To illustrate, let’s run an experiment: does eating chocolate make people
happier? At random, we divide our eager subjects into two groups. Half eat
chocolate bars; half eat graham crackers. Then, all report their happiness on a
scale from 1 (agony) to 5 (bliss). We predict that the chocolate group will score
higher.
But there’s a danger. Even if chocolate makes no difference, one group is
bound to score a little higher than the other. For example, look what happened
when I generated five random samples from the same population:
2. How big the data set is. A two-person sample inspires little confidence.
Maybe I happened to give the chocolate to an enthusiastic lover of life,
and the graham cracker to an ungrateful nihilist. But in a randomly
divided sample of two thousand people, individual differences should
wash out. Even a smallish gap (3.08 vs. 3.01) is unlikely to happen by
fluke.
3. The variance within each group. When the scores are wild and high-
variance, it’s easy for two groups to get different results by fluke. But if
the scores are consistent and low-variance, then even a small difference is
hard to achieve by coincidence.
The p-value boils all of this information down into a single number between
zero and 1, a sort of “coincidence craziness score.” The lower the value, the
crazier it would be for these results to happen by coincidence alone. A p-value
near zero signals a coincidence so crazy that perhaps it isn’t a coincidence at all.
(For a slightly more technical discussion, see the endnotes.)
Some p-values are easy to interpret: 0.000001 marks a one-in-a-million fluke.
Such coincidences are so rare that the effect is almost certainly real: in this case,
that chocolate makes people happier.
Meanwhile, a p-value of 0.5 marks a 1-in-2 event. Such results happen…
well, half the time. They’re as common as weeds. So, in this case, it would seem
that chocolate makes no difference.
Between these clear-cut cases lies a disputed borderland. What about a p-
value of 0.1? What about 0.01? Do these numbers mark out lucky flukes or
results so extreme that they probably aren’t flukes at all? The lower the p-value,
the better; but how low is low enough?
There are other ways to hack a p-value. A 2011 survey found large fractions
of psychologists admitting, under the veil of anonymity, to various “questionable
research practices”:
Even the most innocent-looking of these dogs can bite. Like, say, gathering
more data when your initial results prove inconclusive. Seems harmless, right?
To gauge the power of this p-hack, I simulated a study I call “Who’s the
Better Coin Flipper?” It’s simple: two “people” (read: spreadsheet columns)
each flip 10 coins. Then we check if one of them achieved more heads. In 20
simulations, I achieved a significant result once. That’s exactly as you’d expect
for p = 0.05: 20 trials, one fluke. Reassuring.
Then, I allowed myself the freedom to keep going. Flip another coin. And
another. And another. Stop the study if the p-value dips below 0.05 (or if we
reach 1000 flips with no success).
The results transformed. Now, 12 out of 20 studies yielded a significant
result.
Such tricks aren’t good science, but they’re not quite fraud, either. In their
paper reporting the survey results, the three authors called such methods “the
steroids of scientific competition, artificially enhancing performance and
producing a kind of arms race in which researchers who strictly play by the rules
are at a competitive disadvantage.”
Is there any way to level the playing field?
CLICKS
Turning back to Mathews and Newsweek, I want to confront the natural
question: what kind of metric is the Challenge Index?
3. WINDOW OR SCOREBOARD?
In his 1998 introduction to the very first index, Mathews wrote:
Nearly every professional educator will tell you that ranking schools is
counterproductive, unscientific, hurtful, and wrong. Every likely criteria
you might use in such an evaluation is going to be narrow and distorted…
I accept all those arguments. Yet as a reporter and a parent, I think that
in some circumstances a ranking system, no matter how limited, can be
useful.
All of this casts his Challenge Index as a rough measure: better than nothing,
and even better for wearing its imperfections on its sleeve. An honest window.
But when you publish your statistic in a national news magazine under the
headline “America’s Best High Schools”—well, it starts feeling an awful lot like
a scoreboard.
“The list has taken on a life of its own,” wrote a National Research Council
committee in 2002. “It is now so important to be included among the top 100
schools on the list that some competitive high schools not included have posted
disclaimers on their Web sites indicating why this is the case.”
“The parents are making the most noise,” said a teacher in Milwaukee,
Wisconsin. “Your standing in the community will go up because you’re offering
more AP classes and you might end up on the Newsweek top 100.”
One hallmark of bad scoreboards is that they’re easy to game. In the case of
the Challenge Index, you can push students into AP classes. “Because your
index only considers the number of AP tests taken, and not the actual scores,”
wrote Mathews’s Washington Post colleague Valerie Strauss, “schools put as
many kids into the test pipeline as they could.”
Another problem lay in the fraction. For convenience’s sake, Mathews had
used “graduating seniors” as his denominator, rather than “all students.”
Assuming every student graduates in four years, the math is equivalent. But a
high dropout rate yields a perverse reward. If three students each take an AP
exam and two later drop out, then by Mathews’s math, the remaining graduate
has taken three APs.
Here, then, is one way to tell the tale of the Challenge Index. It began as a
good window. By considering exams taken rather than exams passed, it looked
beyond wealth and privilege to the deeper question of whether students were
being challenged. Flawless? No. Valuable? Yes.
Then it grew. No longer was it a lone journalist identifying the “most
challenging” schools. Now it was a prominent newsmagazine crowning the
“best.” This created perverse incentives and strange outcomes, turning the good
window into a bad scoreboard.
We could tie a bow on our story right there and retire to watch football and/or
study for AP exams. But that would miss the most interesting twist—and the real
nature of the game that Mathews is playing.
Critics call the list easy to game, but Mathews doesn’t mind. In fact, that’s
the whole point: the more students taking the exams, the better. Schools that
urge, cajole, and incentivize aren’t cheating; they’re doing right by their
students. He’s even happy with the title “best,” telling the New York Times that
the word is “a very elastic term in our society.”
As support for his view, Mathews likes to cite a 2002 study of more than
300,000 students in Texas. Focusing on those with low SAT scores, researchers
found that students scoring a 2 on an AP exam (a nonpassing grade) later
outperformed comparable peers who hadn’t attempted the AP at all. The effort
itself—even without a passing score—seemed to lay the groundwork for success
in college.
All of this flips the narrative on its head. Mathews seems to envision the
Challenge Index as a flawed window, as well as the scoreboard the country
needs.
For better or worse, the list’s impact is real. Mathews has always drawn a line
at 1.000—one AP exam per graduating student. In 1998, only 1% of schools
nationwide qualified. As of 2017, that number is 12%. In Washington DC, at the
epicenter of Mathews’s influence (after all, he writes for the Washington Post), it
is over 70%.
There is a beast afoot in the library of life, a chimera known as the Digital
Humanities. It has the body of a literary critic, the head of a statistician, and the
wild hair of Steven Pinker. Some hail it as a burst of light into a dark cavern.
Others scorn it like a slobbering dog with its teeth sunk into a first edition of
Madame Bovary. What does the creature do?
Simple: it turns books into data sets.
Nilesh mourned the loss for months while I wrestled with the guilt of having
betrayed two friends at once, both the word and the roommate. I couldn’t help it.
The ghost in my brain, the one that turns meanings into words, works by instinct
and thrives in shadow. Calling attention to a particular word choice spooked the
ghost. It retreated.
Presented with Blatt’s statistics, it happened again. I became adverb-
paranoid. Ever since, I’ve been writing like a restless fugitive, afraid those -lys
may slip into my prose like spiders crawling into my sleeping mouth. I recognize
this is a stilted, artificial approach to language, not to mention a naïve
correlation-equals-causation approach to statistics. But I can’t help it. It’s the
promise and the danger of the digital humanities all in a nutshell—with the
emphasis, in my case, on “nut.”
Taken as a collection of words, literature is a data set of extraordinary
richness. Then again, taken as a collection of words, literature is no longer
literature.
Statistics work by eliminating context. Their search for insight begins with
the annihilation of meaning. As a lover of stats, I’m drawn in. As a lover of
books, I flinch. Is there peace to be made between the rich contextuality of
literature and the cold analytical power of stats? Or are they, as I sometimes fear,
born enemies?
Upon inspection, less than half of the 1-grams from the year 1900 turned out
to be actual words (as opposed to numbers, typos, abbreviations, and the like),
while more than 2/3 of the 1-grams from the year 2000 were words.
Extrapolating from hand-counted samples, the authors estimated the total
number of English words in each year:
Then, cross-referencing these 1-grams with two popular dictionaries, they
found that lexicographers are struggling to keep pace with the language’s
growth. In particular, the dictionaries miss most of the rarer 1-grams:
In my own reading, I don’t run into many of those rare nondictionary words.
That’s because… well… such words are rare. But the language is populated by
vast, silent ranks of one-in-a-hundred-million obscurities. Overall, the authors
estimated that “52% of the English lexicon—the majority of the words used in
English books—consists of lexical ‘dark matter’ undocumented in standard
references.” The dictionaries only scratch the surface, missing out on gems like
“slenthem.”
For these researchers, spelunking through the lexicon was a mere warm-up.
The authors went on to examine the evolution of grammar, the trajectories of
fame, the fingerprints of censorship, and the shifting patterns of historical
memory. All this they accomplished in just a dozen pages, mainly by tracking
the frequencies of well-chosen 1-grams.
The paper dropped jaws beyond mine. The journal Science, sensing its
magnitude, made it free to nonsubscribers. “A New Window on Culture,”
declared the New York Times.
Literary scholars tend to study an exclusive “canon,” an elite handful of
authors who enjoy deep, focused analysis. Morrison. Joyce. The cat who sat on
Joyce’s keyboard and produced Finnegans Wake. But this paper gestured toward
another model: an inclusive “corpus,” in which a far-flung population of books,
from the celebrated to the obscure, can all share attention. Statistics could serve
to topple the literary oligarchy and to instate democracy.
Now, there’s no reason both approaches can’t live side by side. Close reading
and statistics. Canon and corpus. Still, phrases like “precise measurement” point
to a conflict. Can the meanings of literature be made “precise”? Can they be
rendered “measurable”? Or will these powerful new tools lead us away from the
hard-to-quantify depths of art, in search of nails that our hammer can strike?
Since the Apply Magic Sauce team guards the secret of its technique, I began
snooping around to figure out how the algorithm might operate. Did it diagram
my sentences? Sniff out the latent patriarchy of my sentiments? Did it infiltrate
my thoughts, the way I imagined Virginia Woolf could have, elevating book
reading into a form of soul reading?
Nope. Chances are that it just looked at word frequencies.
In a 2001 paper titled “Automatically Categorizing Written Texts by Author
Gender,” three researchers managed to achieve 80% accuracy in distinguishing
male from female writers, just by counting the occurrences of a handful of
simple words. A later paper, titled “Gender, Genre, and Writing Style in Formal
Written Texts,” laid out these differences in plain terms. First, men use more
noun determiners (“an,” “the,” “some,” “most”…). Second, women use more
pronouns (“me,” “himself,” “our,” “they”…).
Even the frequency of the single innocuous word “you” gives a clue about
the author’s gender:
The system’s accuracy is made more impressive by its utter simplicity. The
approach disregards all context, all meaning, to focus on a tiny sliver of word
choices. As Blatt points out, it would evaluate the sentence “This sentence is
written by a woman” as most likely written by a man.
If you expand your sights to include all words, and not just little grammatical
connectors, then the results take a turn toward the stereotypical. When a data
company called CrowdFlower trained an algorithm to infer the gender of the
owners of Twitter accounts, it spat out the following list of gender-predictive
words:
And in Nabokov’s Favorite Word Is Mauve, Ben Blatt finds the most gender-
indicative words in classic literature:
It seems that Apply Magic Sauce relies on these sorts of clues, too. When
mathematician Cathy O’Neil tried the algorithm on a man’s writing about
fashion, he scored 99% female. When she checked it on a woman’s writing
about mathematics, she scored 99% male. Three of O’Neil’s own pieces scored
99%, 94%, and 99% male. “That’s not an enormous amount of testing,” she
wrote, “but I’m willing to wager that this model represents a stereotype,
assigning the gender of the writer based on the subject they’ve chosen.”
The inaccuracies don’t quiet my creepy feeling. It seems my maleness so
pervades my thinking that an algorithm can detect it in two nonoverlapping
ways: my pronoun usage and my fondness for Euclid.
I know that, on some level, this vindicates Woolf. She saw that men and
women experience different worlds, and believed that the struggle to give
women a voice had to begin all the way down at the level of the sentence. Crude
statistics bear this out: women write about different topics, and in different ways,
than men do.
Still, I find it all a little depressing. If Woolf’s writing reveals her femaleness,
I like to think it’s embedded in her wisdom and her humor, not in her low
density of noun determiners. To hear Woolf distinguish masculine from
feminine prose feels like going to a trusted doctor. To have an algorithm do the
same feels like getting a pat-down at the airport.
Armed with their data, Fred and Dave could reduce each author to something
like a deck of cards, dealing out words at predictable rates. Then, by tallying
word frequencies in the disputed papers, they could infer from which deck the
text had been dealt.
It worked. “The odds are enormously high,” they concluded, “that Madison
wrote the 12 disputed papers.”
In the half century since, their techniques have become standard. They have
helped explore the authorship of ancient Greek prose, Elizabethan sonnets, and
speeches by Ronald Reagan. Ben Blatt applied the algorithm almost 30,000
times, using just 250 common words, to determine which of two authors had
written a given book. His success rate was 99.4%.
My intellect knows there’s nothing wrong with this. Still, my emotions rebel.
How can you understand a book by shredding it to bits?
In 2011, a team of authors at the Stanford Literary Lab attempted a tricky
leap: from fingerprinting authors to fingerprinting genres. They used two
methods: word-frequency analysis and a more sophisticated sentence-level tool
(called Docuscope). To their surprise, both methods converged on accurate genre
judgments.
Take this excerpt, from a passage the computer identified as the most
“Gothic” page in their entire 250-novel corpus:
He passed over loose stones through a sort of court, till he came to the archway;
here he stopped, for fear returned upon him. Resuming his courage, however, he
went on, still endeavouring to follow the way the figure had passed, and
suddenly found himself in an enclosed part of the ruin, whose appearance was
more wild and desolate than any he had yet seen. Seized with unconquerable
apprehension, he was retiring, when the low voice of a distressed person struck
his ear. His heart sunk at the sound, his limbs trembled, and he was utterly
unable to move. The sound which appeared to be the last groan of a dying
person, was repeated…
I find this to be two kinds of creepy. First, there’s the Gothic creepiness of
ruined archways and deathly groans. And second, there’s the creepiness of a
computer detecting the Gothic air without even glancing at the words “archway,”
“ruin,” or “last groan of a dying person.” It had tagged the passage on the basis
of pronouns (“he,” “him,” “his”), auxiliaries (“had,” “was”), and verb
constructions (“struck the,” “heard the”).
I’m unnerved. What does the algorithm know that I don’t?
To my relief, the authors offered a tentative answer. There is no single
element that distinguishes a writer or genre, no unique feature from which all
others follow. Rather, writing has many distinctive features, running from the
galactic structure of the novel down to the molecular structure of the syllable.
Statistical tendencies and larger meanings can coexist, living side by side in the
same sequence of words.
Most of the time, I read for architecture. Plot, theme, character. It’s the high-
level structure: the aspects visible to any passerby but impenetrable to statistics.
If I look closer, I can see the brickwork. Clauses, sentence constructions, the
design of a paragraph. It’s the microlevel structure that my high school English
teachers taught me to scrutinize. Computers can learn to do the same.
I know it’s just a metaphor, but metaphor is the language that the ghost in my
brain speaks. Heartened, I grabbed the first section of this book (“How to Think
Like a Mathematician”) and sampled the rate of -ly adverbs. It landed at 11 per
1000 words—around the same as Virginia Woolf, which I took as a friendly
omen. Then, unable to help myself, I purged the document of extraneous -lys
until the rate dipped below eight per 1000. That territory belongs to Ernest
Hemingway and Toni Morrison. I was cheating, and it felt great.
Can the new statistical techniques find harmony with the older, richer, more
human ways of understanding language? Yes; conceivably.
V
ON THE CUSP
The Fitbit-clad among us have a pretty good idea of how many steps they take:
3000 on a sedentary day; 12,000 on an active one; 40,000 if they’re fleeing all
day from a very slow bear. (Maybe just four or five if they’re fleeing from a fast
one.)
This kind of counting masks a truth we all know: Not all steps are created
equal.
Mathematicians distinguish between two kinds of variables. Continuous
variables can change by any increment, no matter how small. I can drink 1 liter
of diet soda, or 2 liters, or any tooth-dissolving quantity in between. A
skyscraper can loom 300 meters, or 300.1, or 300.0298517. Between any two
values, no matter how close together, you can always split the difference.
Discrete variables, by contrast, move by jumps. You can have one sibling, or
two, but not 1¼. When you buy a pencil, the store can charge 50 cents, or 51
cents, but not 50.43871 cents. With discrete variables, some differences cannot
be split.
Life is a funny mixture of continuous and discrete variables. Ice cream is a
continuous quantity (giving continual pleasure), and yet it is served in discrete
sizes. The quality of job interviews varies continuously, and yet any application
yields a discrete number of job offers (zero or one). The speed of traffic is
continuous; the speed limit is a single discrete barrier.
This conversion process can amplify tiny increments into huge changes. A
tiny burst of acceleration earns you a speeding ticket. A misplaced hiccup during
a job interview costs you the offer. A desire for just a little more ice cream
forces you, through no fault of your own, to order an 18-scoop bucket called “the
BellyBuster.”
That’s what happens in a world that turns “continuous” into “discrete.” For
every cusp in life, there is a critical turning point—an infinitesimal step with the
power to change everything.
Chapter 21
THE FINAL SPECK OF DIAMOND DUST
About 250 years ago, the economist Adam Smith asked a question you might
expect from a toddler: why do diamonds cost so much more than water?
Sometimes you can reveal a misconception just by uttering a phrase. “Soar like
a penguin.” Nope, they don’t fly. “Famous historical Belgians.” Sorry,
Belgium’s most celebrated native is the waffle. “Too full for dessert.” C’mon,
nothing justifies refusing cupcakes. However, my favorite error-in-a-single-
breath is one you’ve probably heard countless times:
“Bumped into the next tax bracket.”
Those six words encapsulate a genuine, widespread, and wholly misguided
fear: What if earning a little extra lifts me into a higher tax bracket? Could I
wind up losing money overall?
In this chapter, I’ll explain the elementary mathematics behind the income
tax, narrate a brief history of its role in American civic life, and identify which
Disney characters are its most zealous advocates. But first, I want to reassure
you that tax brackets do not go bump in the night. The dreaded “tax bracket
bump” is as fictional as a Belgian celebrity.
Our tale begins in 1861. As the American Civil War looms, the federal
government begins cooking up get-rich-quick schemes, but none of them hold
much promise. The longtime tariff on foreign goods has stopped bringing
adequate funds. A tax on consumer purchases might alienate voters by hitting
poor Americans harder. And a tax targeting rich people’s wealth (including
property, investments, and savings) would violate the Constitution’s ban on
“direct” taxes. What’s a cash-strapped Congress to do?
The only thing they could. In August, the federal government introduced a
temporary emergency tax on income. On any earnings above $800, you now
owed 3%.
The rationale? Money has a diminishing marginal utility. The more dollars
you’ve got, the less another dollar means. Thus, it causes less pain to tax
somebody’s thousandth dollar than to tax somebody else’s first. The resulting
systems, wherein higher earnings face a higher marginal tax rate, are called
“progressive.” (A “regressive” tax draws a larger percentage from those with
lower incomes; a “flat” tax takes the same percentage from everybody.)
You can imagine our American forefathers, scratching their extravagant
beards, fretting whether a small raise might “bump” them from the freebie
bracket to the taxpayer bracket. After all, an income of $799.99 incurred no tax,
but an income of $800.01 did. Could a $0.02 raise really cost you $24 in tax?
Thankfully, no. Like its modern equivalent, the tax didn’t apply to all income
—just on the margin, to the last dollar earned. Whether you were a poor
Appalachian farmer or Cornelius Vanderbilt (the Bill Gates of the railroad age),
your first $800 was tax-free.
If you made $801, then you’d pay the 3% on the last dollar alone, for a total
liability of just $0.03.
If you made $900, then you’d pay the 3% on the last $100, for a total liability
of just $3.
And so on.
To visualize this, I like to imagine the government divides your money into
buckets. The initial bucket fits $800 and is labeled “Survival Money.” You fill
this first, and pay no tax on it.
The second bucket is labeled “Leisure Money.” Once the first bucket hits
capacity, the rest of your money goes here. Of this, the government takes 3%.
By 1865, the government had raised the rates and brought in a third bucket:
Then the war ended. The income tax lay dormant for a few decades, until a
flurry of activity at the end of the century. In 1893, a financial panic wracked the
country. In 1894, the income tax came riding in to save the day. Then in 1895,
the Supreme Court ruled that—ha ha, oops—that ban on “direct” taxes applies to
you too, income tax. You’re unconstitutional. How embarrassing.
It took two decades for a constitutional amendment to revive the tax. Even
then, it returned without flags and fanfare, preferring to sneak back in on tiptoes.
In 1913, the first year of the permanent income tax, only 2% of households paid
it. The marginal rates maxed out at a measly 7%, which didn’t even kick in until
$500,000 (the inflation-adjusted equivalent of $11 million today).
The rates have a pleasing ABC elegance, climbing from 1% to 7% in a tidy
linear progression, with cutoffs at nice round numbers.
But why these in particular? No reason, honestly.
“Individual judgments will naturally differ,” President Woodrow Wilson
wrote, “with regard to the burden it is fair to lay upon incomes which run above
the usual levels.” No airtight mathematical formula can dictate the “right” rates.
Rate setting is a subjective process, full of guesswork, politicking, and value
judgments.
Of course, arbitrary legal lines are not unique to taxation. In the US, the day
before your 21st birthday, you can’t buy alcohol; the day after, you can. In
theory, the law could ease you in gradually—allow you to buy beer at 19, wine
at 20, and hard liquor at 21—but society has chosen a bright, simple line over a
hard-to-enforce gradient.
The architects of that 1913 income tax went the other way. I find it funny that
they broke such a modest tax into such tiny increments. Their seven brackets
match the number we have today, except ours cover a range of 27%, and theirs
just 6%. It seems that the government didn’t trust the system’s marginal nature,
and thus sought to avoid big “jumps” between rates. Perhaps they were thinking
in terms of this graph:
When you focus on the rates themselves, every bracket transition looks sharp
and jagged. So lawmakers endeavored to smooth the system, to make it appear
safe and gradual.
Wasted energy, if you ask me. Savvy taxpayers don’t care about the abstract
properties of the tax scheme. They care what they’re paying—both overall and
on the margin—and by design, the income tax ensures that earning an extra
dollar will never alter the taxes already paid. The relevant graph is this one:
Whereas the first graph shows stark jumps, the second shows a single
continuous line. It reflects the shifting rates via its changing slope: nice and
shallow for a low tax rate, harsh and steep for a high one. This better captures
how people actually experience bracket transitions: not as abrupt leaps
(discontinuities, to a mathematician) but as moments where one slope gives way
to another (nondifferentiable points, in calculus-speak).
Evidently, nobody told Congress, because they took their cliff wariness to a
comical extreme in the 1918 tax scheme:
Looking at this mess, you’ll first notice that the rates have hulked out. In just
five years, the marginal rate on America’s wealthiest has grown elevenfold. I
guess that’s what happens when your country becomes embroiled in something
called “the war to end all wars” (although the change would long outlast the
conflict; the top rate hasn’t dipped below 24% since).
Even more striking—to me, at least—is the sheer number of tax brackets
instituted at that time. The US had more tax brackets than states, 56 to 48.
It reminds me of a favorite project from when I taught precalculus in
California. Every year, I’d ask 11th graders to design their own income tax
system. One industrious fellow named J. J. decided to take the “gradual
transition” idea to its natural conclusion, designing a system where the marginal
rates varied continuously, with no jumps whatsoever.
Imagine you want your marginal rates to start at zero and eventually reach
50% (on income above $1 million). You could do this with two brackets:
Or you could take the World War I–era approach, and break this into 50
brackets:
But why stop there? Make it 1000 brackets.
Or better yet, a million brackets.
Heck, at the extreme, you could just connect your endpoints with a straight
line.
This final graph embodies a tax system where every fraction of a penny is
taxed at an infinitesimally higher rate than the penny before. There is no such
thing as “brackets”; each atom of income faces its own unique rate. Everywhere
is a transition, and so, in some paradoxical sense, nowhere is.
Under such a system, the graph of total tax paid will look something like this:
The slope grows gradually. No nondifferentiable points here. It’s as smooth
as a German car accelerating on the Autobahn.
Look at the income tax through US history, and you’ll witness a similar
acceleration. Over the first half of the 20th century, the tax grew from
“unconstitutional proposal” to “tentative experiment” to “wartime necessity” to
“the government’s primary funding mechanism.” In 1939, at the dawn of World
War II, fewer than 4 million Americans paid income tax. By 1945, more than 40
million did. Revenue showed a similar growth, from $2.2 billion when the war
began to $25.1 billion by its end. Today, its various state and federal forms
amount to $4 trillion per year—almost a quarter of the country’s economy.
As the income tax surged in 1942, the government commissioned Walt
Disney to create a short film to inspire Americans to pay their taxes. The
secretary of treasury demanded a fresh-faced custom-animated protagonist, but
Walt Disney insisted on using the company’s biggest star at the time. Thus, more
than 60 million Americans got to see on their movie screens a display of
patriotic, taxpaying fervor (“Oh boy! Taxes to beat the Axis!”) from none other
than Donald Duck.
The cartoon must have worked, because two years later, the top marginal tax
rate achieved its historical high-water mark: 94%. Other countries pushed even
higher. In the 1960s, the UK taxed top earners at a marginal rate of 96%,
prompting the Beatles to begin their greatest album with the song “Taxman” and
these scathing lyrics:
The most extraordinary tale of high marginal tax rates comes from—guess
where?—Sweden. Rather than tell the tale myself, I shall defer to children’s
book author Astrid Lindgren, who in 1976 published a satire of her experiences,
in the evening tabloid Expressen. It tells of a woman named Pomperipossa,
living in a land called Monismania. Some folks in this land complained about the
“oppressive taxes” that funded the welfare state. But Pomperipossa, despite her
83% marginal tax rate, did not. Instead, she contented herself with keeping 17%,
and, “filled with joy,” she “kept skipping down the road of life.”
Now, Pomperipossa (a stand-in for author Lindgren) wrote children’s books.
In the eyes of the government, this made her a “small business owner,”
subjecting her to “social employer fees.” But Pomperipossa did not understand
the implications of this until a friend pointed them out:
“Are you aware of the fact that your marginal tax rate this year is
102%?”
“You are talking nonsense,” Pomperipossa said. “That many percent
does not exist.”
For she was not particularly familiar with Higher Mathematics.
Of the first $150,000 she earned, she would get to keep $42,000. But above
that lay only heartbreak. Additional earnings would do worse than just vanish;
they would take some of her earlier income along with them. Each additional
$100,000 pretax would reduce her income by $2000 posttax.
She said to herself: “My dear old woman… There are decimal points and
all those things, surely you have counted wrong, there must be 50,000 left
for you.” She started over, but the result did not change one bit… She
now understood that it was something dirty and shameful to write books,
since it was punished so severely.
The story ends with Pomperipossa, her earnings all sapped by taxes, going on
welfare. The final line: “And she never, ever wrote any books again.”
The figures in the story come directly from Lindgren’s life, and make
Sweden’s top rate today—a mere 67%—look as quaint as a butter churn.
“Pomperipossa in Monismania” caught fire in Sweden, igniting a fierce
debate that helped deal the Social Democratic Party its first electoral defeat in 40
years. Lindgren, a longtime Social Democratic supporter, kept voting for the
party in spite of her grievances.
Back in the US, although the arguments surrounding the income tax are now
a century old, they’re as vibrant (and acrimonious) as ever.
My students’ projects captured pretty much every side of the debate. Some
tweaked the rates upward in the name of redistribution, or downward in the
name of promoting economic growth. One student, who is as clever as he is
cantankerous, designed a regressive system, whereby higher income levels faced
a lower marginal tax rate. He argued that the poor needed an “incentive” to earn
money. Maybe he was sincere, maybe he was satirizing Republican politics, or
maybe he was just trolling me. (He got an A.) Still others sought a radical Robin
Hood system of economic justice and chose top rates close to 100%. A few
students chose rates above 100%, wishing to inflict Pomperipossa’s fate on the
ultrarich. Still other students wanted to step outside the income tax framework
altogether, to envision a whole new regime.
In short, I saw inventiveness everywhere and consensus nowhere. I guess
that’s what they call “America.”
Chapter 23
ONE STATE, TWO STATE, RED STATE, BLUE STATE
But by 1832, every state except South Carolina had tossed its electors into the
democratic fray, choosing to let the people decide.
What changed?
The Philly 55 had envisioned an enlightened democracy of conscience-driven
statesmen, unencumbered by those nasty, rivalrous things called “political
parties.” Then they went out and formed things that looked an awful lot like
political parties.
Hypocrisy aside, this had advantages. Now, you didn’t need to study the
conscience of every single candidate. You could just familiarize yourself with
the party platforms and vote for the team you prefer. No need for an
intermediary.
And so the country moved from human “electors” to abstract “electoral
votes,” the layer of statesmen replaced by a layer of mathematics.
Of course, as with many elements of the Constitution, this system encoded an
advantage for slavery. How did it work? Well, recall that a state got one elector
for each congressman. That included senators (two per state) and House
members (varying in proportion to population; in 1804, they ranged from one to
22 per state).
The Senate half of the equation gave a boost to small states, by treating little
Rhode Island (fewer than 100,000 people back in the day) on par with big
Massachusetts (more than 400,000). But the real plot twist unfolded over in the
House. The Philly 55 had debated whether, when apportioning House members,
the tally of people represented ought to include slaves. If yes, then the slave-
owning South would get extra representatives; if not, the North stood to benefit.
They compromised on counting them partway. Every five slaves would
register as three people. Hence the most infamous fraction in legal history: 3/5.
The Electoral College imported this proslavery compromise into presidential
elections, like a virus downloaded from an email attachment. Here’s the elector
count from 1800, compared to a hypothetical system where slaves went
uncounted:
Perhaps a 10-elector swing doesn’t sound like much. It was. Of the country’s
first 36 years, it spent 32 led by slave-owning Virginians. The only interruption
came from John Adams of Massachusetts, who lost a close reelection bid in
1800… So close, in fact, that 10 electors would have swung it.
In theory, if all states went proportional, neither party would benefit. It would
be like both sides in a duel laying down their guns. That’s the upper-left box.
But it’s not a stable equilibrium. Once your foe has dropped her weapon, you
might as well grab yours back. If every state does the same, then we soon land in
the lower-right box—which is, of course, where 96% of the country sits today.
Listening to a description of the Electoral College, you’d think that state
borders matter. That to share a driver’s license is to share a special kinship. But
that’s not how state legislators act. By choosing the all-or-nothing system, they
do what’s best for the party, even if it means marginalizing the state.
For example, here’s how the last 10 elections have turned out in Texas:
Going winner-take-all amounts to posting a big sign at the border that reads,
“Don’t Worry What My Voters Think; We Always Go Republican!” Under
winner-take-all, 55% is as good as 85%, and 45% is no better than 15%. Going
winner-take-all means the election is over before it begins, and so neither party
has any reason to tailor its policies to win over your citizens. Why waste
resources where they will make no difference?
On the other hand, here’s how those elections look if you allocate electors
proportionally:
Now, modest shifts in the vote leave a real impact on the elector count. If
Texas wants its voters to have a stake in the game, then it ought to go
proportional, so that an extra Texan vote actually boosts a candidate’s chances of
winning. It would give campaigns reason to campaign.
Why doesn’t this happen? Because “maximizing the impact of a marginal
vote in my state” is not what motivates legislators. First and foremost, they are
not Texans, or Californians, or Kansans, or Floridians, or Vermonsters.
They are all Democrats. They are all Republicans.
After more than a century in which the Electoral College always aligned with
the nationwide popular vote, we’ve seen two divergences in the last five
elections. In 2000, the Democrats won the popular vote by 0.5%, while the
Republicans won the Electoral College by five votes. Paper-thin margins on both
fronts. In 2016, the disparity widened: the Democrats won the nationwide vote
by 2.1%, while the Republicans won the Electoral College by 74.
Is the Electoral College now tilted to favor Republicans?
Nate Silver, the modern-day statistical prophet of the Electoral College, has a
beautiful method for answering this question. It allows us to determine the
Electoral College advantage, not just in exceptional years like 2000 and 2016,
but in any election.
The procedure (which I’ll apply to 2012) goes something like this:
You eye the title of this chapter with justified skepticism. “History?” you ask.
“What do you know of history, mathematician?”
I mumble a few incoherencies about walruses, tax law, and some wigs in
Philadelphia; your pity grows.
“Historians find causal patterns in the sprawl of the past,” you explain. “Your
tidy formulas and quaint quantitative models have no place in this messy human
realm.”
I hunch my shoulders and begin to sketch a graph, but you shush me.
“Run home, mathematician!” you say. “Go, before you embarrass yourself!”
Alas, I lost my chance to avoid embarrassment the day I blogged my first
stick figure, and so, in halting tones, I begin to tell my tale.
1. A NOR’EASTER FROM A ROUNDING ERROR
In the winter of 1961, two surprises hit the East Coast, more or less
simultaneously.
First, in Washington, DC, 8 inches of snow fell on the eve of President
Kennedy’s inauguration. Anxious Southern drivers, possibly interpreting the
snow as a signal of Armageddon, abandoned their cars by the thousands.
Apocalyptic traffic jams ensued. The US Army Corps of Engineers managed to
clear a path for the inaugural parade only by recruiting hundreds of dump trucks
and flamethrowers to the cause.
It was, in short, chaos.
Second, back in Kennedy’s home state of Massachusetts, a researcher named
Edward Lorenz found something funny. Since the year before, he’d been
developing a computer model of the weather. To begin, you entered some initial
conditions. Then, the computer ran them through a set of equations. In the end, it
delivered a printout of the resulting weather. It could use these results as the
initial conditions for the next day, iterating the process to create months’ worth
of weather from a single starting point.
One day, Lorenz wanted to re-create an earlier weather sequence. One of his
technicians retyped the inputs, rounding them a little (e.g., 0.506127 to 0.506).
The tiny errors—smaller than what weather instruments could even detect—
should have faded into the background. And yet, within weeks of simulated
weather, the new sequence had diverged completely. A tiny adjustment had
created a whole new chain of events.
It was, in short, chaos.
This moment marked the birth of a new experimental style of mathematics,
an interdisciplinary insurgency that soon became known as “chaos theory.” This
field explored a variety of dynamic systems (gathering storms, turbulent fluids,
populations in flux) with a strange set of shared features. They tended to follow
simple and rigid laws. They were deterministic, with no room for chance or
probability. And yet, because of the subtle interdependence of their parts, they
defied prediction. These systems could amplify small changes into tremendous
cascades, a tiny upstream ripple into an enormous downstream wave.
Lorenz and the nation’s capital were both stunned by the unpredictability of
the weather. But the link between the events runs deeper than this. Forget the
chaos of the snowstorm, and consider instead the fact that John F. Kennedy was
being inaugurated at all.
Three months earlier, he had defeated Richard Nixon in one of the closest
elections in US history. He won the nationwide vote by just 0.17% and carried
the Electoral College thanks to narrow margins in Illinois (9,000 votes) and
Texas (46,000 votes). Half a century later, historians still debate whether voter
fraud might have tipped Kennedy over the edge. (Verdict: Probably not, but who
knows?) It is not hard to imagine a nearby parallel universe where Nixon eked
out a win.
But it is very hard to imagine what would have happened next.
The Bay of Pigs invasion, the Cuban missile crisis, Kennedy’s assassination,
LBJ’s ascension to the presidency, the Civil Rights Act, the Great Society, the
Vietnam War, the Watergate break-in, Billy Joel’s timeless hit “We Didn’t Start
the Fire”… All this and more hinged upon decisions made in the White House.
A 0.2% swing in November of 1960 could have altered the path of world history,
like a rounding error that births a nor’easter.
Ever since I was old enough to notice the world changing, I’ve been
wondering how to conceptualize those changes. Civilization charts a path that’s
unknowable, unforeseeable, and unimaginable except after the fact. How can we
make sense of a system where enormous and untold consequences can issue
from a single indiscernible step?
Now, does history strike you more as a single pendulum or a double one?
History kicks right; a tyrant topples. It swings left; a war begins. It pauses as
if resting, and a startup scheme hatched in a California garage remakes the world
in its emoji-drenched image. Human civilization is an interconnected system,
hypersensitive to change, with brief intervals of stability and flurries of wild
activity, deterministic yet wholly unpredictable.
In mathematics, an “aperiodic” system may repeat itself, but it does so
without consistency. We’re told that those who do not learn their history are
doomed to repeat it, but perhaps the situation is worse than that. Maybe, no
matter how many books we read and Ken Burns documentaries we watch, we
are all doomed to repeat history—and still, in spite of ourselves, to be caught off
guard, recognizing the repetitions only in hindsight.
3. THE GAME OF LIFE
“Come now, mathematician,” you cajole. “People just aren’t that complicated.”
I frown.
“Don’t take it personally,” you say. “But you’re not that hard to predict.
Economists can model your financial choices. Psychologists can describe your
cognitive shortcuts. Sociologists can characterize your sense of identity and pick
apart your choice of Tinder photo. Sure, their colleagues in physics and
chemistry may roll their eyes and make vomiting gestures, but the social
scientists can achieve a surprising degree of accuracy. Human behavior is
knowable, and history is the sum of human actions. So shouldn’t it be knowable,
too?”
That’s when I grab a computer and show you the Game of Life.
Like all fun things, it involves a grid. Each square—known as a “cell”—can
assume two states: “alive” or “dead.” In fairness, the word “game” may overstate
things, because Life unfolds from step to step according to four automatic,
immutable rules:
That’s it. To start the game, just bring a few cells to life. Then, watch as the
board changes, step by step, according to these rules. It’s like a snowball that
needs only a nudge to begin rolling downhill. No further input is required.
The big choices we make [in life] are practically random. The small
choices probably tell us more about who we are. Which field we go into
may depend on which high school teacher we happen to meet. Who we
marry may depend on who happens to be around at the right time of life.
On the other hand, the small decisions are very systematic. That I became
a psychologist is probably not very revealing. What kind of psychologist I
am may reflect deep traits.
Alternate histories tend to dwell on the “great men” and “important battles”
of conventional history, missing quieter possibilities that cut against the grain of
the dominant culture. The rules of plausibility (what stories we find credible and
compelling) don’t always mirror those of probability (what things actually came
close to happening).
True chaos is a narrative-destroying idea, a thought as anarchic as a bomb.
We’re not done. Switch to a 100-meter ruler, and the same process will
repeat. Curves and crinkles overlooked by the longer rulers now become
apparent, increasing the total length.
We can keep repeating this process. The closer you look, the longer the
coastline becomes—in theory, forever.
Suffice it to say, this is kind of weird. With most inquiries, looking closer
helps clarify the answer. Here, in a frightening inversion, looking closer causes
the problem to unravel. It never simplifies, never resolves.
This trend reaches its extreme with the Koch snowflake, a mathematical
object consisting of bumps upon bumps upon bumps. Although it occupies only
a small patch of the page, the length of its boundary is theoretically infinite.
In the graphic novel From Hell—a work of speculative history, about the
series of murders in London’s Whitechapel neighborhood in 1888—writer Alan
Moore brings in Koch’s snowflake as a metaphor for the nature of historical
scholarship. “Each new book,” he says in the afterword, “provides fresh details,
finer crenellations of the subject’s edge. Its area, however, can’t extend past the
initial circle: Autumn 1888, Whitechapel.”
In Moore’s telling, there is something bottomless about the study of history.
The further we zoom, the more we see. A finite stretch of time and space can
enclose infinite layers of detail, allow for endless chains of analysis. Chaos is
complexity all the way down, never pixelating, never terminating, never
resolving.
Is history chaotic like the Game of Life—simple on a small scale yet
unpredictable on a large one? Or is its unpredictability like the weather’s—with
wild swings on the small day-to-day scale yet averaging out in the long run to a
stable climate? Or is it perhaps like the Koch snowflake—with chaos on every
level, complexity at every scale? These metaphors compete in my mind, like
three PowerPoint presentations projected onto a single screen. I sometimes feel
I’m on the cusp of figuring it out—and then I check the news, and the world has
changed again, into yet another strange and unknowable shape.
ACKNOWLEDGMENTS
The elements of this book are a bit like the atoms in my body: only nominally
and temporarily “mine.” They’ve circulated for years, coming from too many
sources to credit or trace. The best I can do is gesture toward the whole
ecosystem that made this book possible.
For the book’s style, I thank all of the witty, kindhearted folks at the Yale
Record, with special hat tips to Davids Klumpp and Litt, and Michael Gerber
and Thornton.
For the book’s perspective, I thank my extraordinary colleagues at King
Edward’s School, with special affection for Tom, Ed, James, Caz, Richard, Nei
—ah, heck, for everybody. Teachers are a playful, critical, inclusive, curious,
and slightly batty tribe that I’m proud to call my people.
For the book’s sense of purpose, I thank my students and my teachers, who
have shaped in uncountable ( 1) ways my thinking about math and about the
world.
For the book’s errors (especially any omitted thank-yous), I apologize in
advance.
For the book’s existence, I thank: the dozens of gracious people who gave
feedback and advice (see the endnotes); Chank Diesel, for elegantly AutoTuning
my handwriting into something radio-ready; Mike Olivo, for educating me about
wookiees; Paul Kepple, for assembling so many bad drawings into a beautiful
whole; Elizabeth Johnson, for bringing peace to the ugly war between myself
and hyphens, and for knowing how many a’s are in “Daaaaaamn”; Betsy
Hulsebosch, Kara Thornton, and the rest of the Black Dog & Leventhal team;
Dado Derviskadic and Steve Troha, for being the first to envision this book—
eons before I could—and for helping me to get there; and Becky Koh, for her
wonderful editing, a job that seems to combine the trickiest elements of
Executive Producing and parenting.
Love and gratitude to my family: Jim, Jenna, Caroline, Lark, Farid, Justin,
Diane, Carl, my happy triangle Soraya, my potato wizard Skander, Peggy, Paul,
Kaya, and the whole Orlin, Hogan, and Williams clans. With fond memories of
Alden, Ros, Pauline, and, of course, Donna.
Finally: thanks, Taryn. You picked math and I’m glad I came along for the
ride. I love you even more than I troll you.
ENDNOTES
ASIDES, SOURCES, THANK-YOUS, MATHEMATICAL
DETAILS, AND JOKES THAT WERE TOO WEIRD AND
DISRUPTIVE FOR THE TEXT
Section I: How to Think Like a Mathematician
Chapter 1: Ultimate Tic-Tac-Toe
• Their board looked like this: This game’s origins are murky. It may have first
appeared in Games magazine in the late 1990s or early 2000s (although the folks
at Games hadn’t heard of it when I asked them). In 2009, a wooden board-game
version called Tic-Tac-Ku won an award from Mensa. Perhaps the game had
multiple independent discoveries, like dancing or calculus.
• with my students: When I first showed the game to my students at Oakland
Charter High School in 2012, they took to calling it Ultimate Tic-Tac-Toe. My
2013 blog post with this title seems to have marked an inflection point for the
game’s popularity: Wikipedia, a handful of academic papers, and a larger
handful of phone apps all refer to the game by that name. The obvious
conclusion: take pride, Matadors! You named this thing.
• “what does any of this have to do with math?”: I thank Mike Thornton, who
read an early draft of this chapter and asked exactly this question. Mike’s edits
are like Leonard Cohen’s songwriting or Hemingway’s prose: I always knew
they were good, but the older I get, the more I appreciate them.
• deeper insight into the nature of rectangles: The key idea is that thin
rectangles have disproportionately large perimeters, whereas squarish rectangles
have disproportionately large areas. So just pick a long thin rectangle (e.g., 10 by
1) and a squarish one (e.g., 3 by 4).
• Again, see the endnotes: If you require the solutions to use only whole
numbers, the problem is a lot of fun. Here’s my derivation of a formula that
generates an infinite family of solutions:
That gives you infinite solutions, but not all of them, because other choices
for d may still yield integer values for c. For example, this formula misses my
favorite solution: 1 by 33 and 11 by 6. My colleague Tim Cross, a dexterous
hand with Diophantine equations, showed me a nifty way to characterize all
possible integer solutions. In the misanthropic tradition of my field, I’ll leave
that as an “exercise for the reader.”
• a guaranteed winning strategy: The actual strategy is a little too complicated
to describe here, but you can see it implemented by some folks from Khan
Academy at https://www.khanacademy.org/computer-programming/in-tic-tac-
toe-ception-perfect/1681243068.
• a clever Brit locked himself in an attic: I recommend the full story: Simon
Singh, Fermat’s Last Theorem (London: Fourth Estate Limited, 1997).
• the great Italian mathematician Cardano: This quote comes from the only
dictionary I’ve ever read for pleasure: David Wells, The Penguin Dictionary of
Curious and Interesting Numbers (London: Penguin Books, 1997).
Chapter 3: What Does Math Look Like to Mathematicians?
• I zip through five chapters of a Twilight novel: I’m more of a Hunger Games
guy, to be honest.
• mathematicians employ strategies: Michael Pershan, the Most Curious and
Analytical Man in the World, articulated this “strategy” idea before I had gotten
there myself. My thanks for his help with this chapter.
• peeling the circle like an onion: There’s a nice animation at
www.geogebra.org/m/WFbyhq9d.
Chapter 4: How Science and Math See Each Other
• teachers like me: My own teacher for this chapter was David Klumpp, whose
feedback combines the erudition of someone like Carl Sagan with the gentle
humanity of someone like Carl Sagan. (David Klumpp is Carl Sagan.)
• algebra was perhaps the dullest: Israel Kleiner, “Emmy Noether and the
Advent of Abstract Algebra,” A History of Abstract Algebra (Boston:
Birkhäuser, 2007), 91–102, https://link.springer.comchapter10.1007/978-0-
8176-4685-1_6#page-2. I’ve done some real violence to Kleiner’s argument; the
key point is that the 19th century brought big advances in analysis and geometry,
while algebra remained in a more concrete and primordial state.
• a mathematician named Emmy Noether: Joaquin Navarro, Women in Maths:
From Hypatia to Emmy Noether. Everything is Mathematical. (Spain: R.B.A.
Coleccionables, S.A., 2013).
• “She often lapsed”: Professor Grace Shover Quinn, as quoted in: Marlow
Anderson, Victor Katz, and Robin Wilson, Who Gave You the Epsilon? And
Other Tales of Mathematical History (Washington, DC: Mathematical
Association of America, 2009).
• Sylvia Serfaty: All the Serfaty quotes come from her interview with: Siobhan
Roberts, “In Mathematics, ‘You Cannot Be Lied To,’” Quanta Magazine,
February 21, 2017, https://www.quantamagazine.org/sylvia-serfaty-on-
mathematical-truth-and-frustration-20170221. I recommend Roberts’s writing on
math the way I recommend R.E.M.’s best albums: zealously.
• Grothendieck: Colin McLarty, “The Rising Sea: Grothendieck on Simplicity
and Generality,” May 24, 2003,
http://www.landsburg.com/grothendieck/mclarty1.pdf.
• Gaussian correlation inequality: Natalie Wolchover, “A Long-Sought Proof,
Found and Almost Lost,” Quanta Magazine, March 28, 2017,
https://www.quantamagazine.org/statistician-proves-gaussian-correlation-
inequality-20170328. The story is a great read, even after you’ve encountered
my spoilers of it.
• a beloved physics teacher: Farhad Riahi, 1939–2011.
• A student named Vianney: Whereas Corey is a pseudonym, Vianney is not. I
figure she deserves the victory lap. I’ve fudged the dialogue to fill gaps of
memory, but this is pretty much how it went down.
• Ngô B u Châu: The quotes come from a press conference Châu gave at the
2016 Heidelberg Laureate Forum. My enormous thanks to A+ human being
Wylder Green and the HLF team for making my attendance there possible.
• the Great Pyramid of Giza: I should credit here the essential source for this
chapter: Mario Salvadori, Why Buildings Stand Up (New York: W. W. Norton,
1980). It’s a stately, upstanding book, without which this chapter would have
toppled like an erroneous proof. My thanks also to Will Wong, architect of
thought and of intramural sports victories, for his help with this chapter.
• a peculiar loop of rope: I learned about the Egyptian rope pullers from: Kitty
Ferguson, Pythagoras: His Lives and the Legacy of a Rational Universe
(London: Icon Books, 2010).
• frustums: Truncated pyramids with trapezoids for lateral faces. A word worth
knowing.
• less than 0.1% of its internal volume: Wikipedia lists dimensions for
three passageways (descending, ascending, horizontal) and three rooms (king’s
chamber, queen’s chamber, grand gallery). These volumes totaled to 1340 m3,
which in a structure of 2,600,000 m3 amounts to roughly 0.05%. I rounded this
up to 0.1%, then (with more thanks to Wikipedia) calculated the volume of the
Empire State Building (2,800,000 m3). Taking 0.1% of this, you get 2800 m3.
Dividing by the area of a single floor (roughly 7400 m2), you get a height of 38
centimeters, which is about 15 inches. I rounded up to 2 feet. Then, just as I
completed this manuscript, a hidden chamber was discovered! Still, my rounding
should more than cover the possible error.
• crucial supporting character: Pun 100% intended.
• the beam’s effect: This discussion is pilfered from Salvadori’s Why Buildings
Stand Up, and no doubt loses much in the retelling.
• Architects are no fools: Ted Mosby notwithstanding.
• hence, the term “I-beam”: I borrow again from Salvadori. As Will Wong
points out, a more traditional presentation would foreground the desirable
properties (stress distribution, torque prevention, etc.) created by the cross-
section’s I shape.
• The Pratt truss: My knowledge of trusses derives from that greatest of human
creations, Wikipedia. For more, check out the “Truss” and “Truss Bridge” pages.
Chapter 7: Irrational Paper
• When I moved to England: Thanks to Caroline Gillow and James Butler, two
people with hearts so wide they make the Atlantic look piddling, for their help
and encouragement on this chapter, and for making that “I moved to England”
experience as wonderful as it was.
• My 12-year-old student Aadam: The same day Aadam coined “disintegers,”
an 11-year-old student named Harry responded to my greeting of “Hello,
algebraists!” with this reply: “Why not call us alge-zebras?” I tell ya, teaching is
a good gig.
• all exactly the same proportion: Another delight is that each sheet of A1 has
an area of precisely 1 m2, and each step down the series cuts the area in half.
Thus, eight sheets of A4 will give you precisely a square meter (except not
precisely, because… irrationals).
Chapter 8: The Square-Cube Fables
#1: Why are big hot-air balloons more cost-effective? Because the canvas
required depends on the surface area (2D), while the lift provided depends
on the volume of helium (3D).
#2: Why do huge dino-birds like turkeys take so long to cook? Because heat
absorption grows with surface area (2D), but heat required grows with
volume (3D).
#3: Why is dry wheat perfectly safe while wheat dust is an explosive hazard?
Because exothermic chemical reactions occur on the surface of a
substance, and tiny dust particles are far more surface-heavy than intact
wheat stalks.
• “paradox of the dark night sky”: I first learned about Olber’s paradox from
Pieter van Dokkum’s, astronomy course Galaxies and Cosmology. If you want
someone to credit (or blame) for this chapter’s eclecticism, look no further than
my liberal arts education.
• brighten you to death: I give John Cowan the final word here: “What if there
are Dark Things (planets, dust, etc.) between us and the stars? Wouldn’t that
make some stars invisible and eliminate the paradox? No, because the stars
behind the Dark Things would over time heat them up to the same temperature
as the stars, eliminating all darkness.”
• Edgar Allan Poe prose poem: E. A. Poe, Eureka: A Prose Poem (New York:
G. P. Putnam, 1848).
Chapter 9: The Game of Dice
• these Roman emperors: The historical facts in this chapter come from three
sources, listed here in rough order of how much I plundered from them:
• Deborah J. Bennett, Randomness (Cambridge, MA: Harvard University
Press, 1998).
• “Rollin’ Bones: The History of Dice,” Neatorama, August 18, 2014.
Reprinted from the book Uncle John’s Unsinkable Bathroom Reader,
http://www.neatorama.com/2014/08/18/Rollin-Bones-The-History-of-
Dice/.
• Martin Gardner, “Dice,” in Mathematical Magic Show (Washington, DC:
Mathematical Association of America, 1989), 251–62.
• does not make a fair die: The snub disphenoid has two cousins, where every
face is an equilateral triangle yet the die is unfair: (1) the triaugmented triangular
prism (shown to me by Laurence Rackham), and (2) the gyroelongated square
bipyramid (shown to me by Tim Cross and Peter Ollis). If you prefer
quadrilateral faces, there’s (3) the pseudo-deltoidal icositetrahedron (shown to
me by Alexandre Muñiz).
• why aren’t they more popular?: To be fair, there are examples. Ancients in
the Indus Valley threw triangular long dice molded from clay. Their
contemporaries in India rolled rectangular ones, carved from ivory.
• demands a red carpet: This perhaps explains why all the modern long dice
I’ve seen are either (a) tiny, or (b) twisted so as to keep the long faces equivalent
while generating a more satisfying and contained tumble. The Dice Lab, whose
work helped to inspire this chapter, sells some nice instances of the latter.
• mathematical principle: continuity: For more on this line of argument, see:
Persi Diaconis and Joseph B. Keller, “Fair Dice,” American Mathematical
Monthly 96, no. 4 (April 1989): 337–39,
http://statweb.stanford.edu/~cgates/PERSI/papers/fairdice.pdf.
• this approach bugs people: Nevertheless, several civilizations have done
similar. Ancient Inuits rolled chair-shaped ivory dice, on which they counted
only three of the six sides. The Papago Indians threw bison bones, counting only
two of the four sides.
• people try to cheat: For ingenious and evil details, see: John Scarne, Scarne
on Dice, 8th ed. (Chatsworth, CA: Wilshire Book Company, 1992).
• Renumber the faces: A related approach, unsubtle yet elegant, comes from the
gangster Big Jule in the musical Guys and Dolls. His lucky dice are blank, with
no spots on them at all. But don’t worry: Big Jule remembers where all the spots
used to be.
• nothing appears to be wrong: Well… almost. A true expert can recognize a
tap, because from certain angles the faces do not follow the usual configuration,
but rather a mirror image. For this reason and others, taps con artists always
move their cheat dice in and out of play quickly, to avoid detection.
• to roll that d100: As Ralph Morrison, knower of all things trivial and
nontrivial, pointed out to me, the Dice Lab sells a gorgeous 120-sided die. For
more: Siobhan Roberts, “The Dice You Never Knew You Needed,” Elements,
New Yorker, April 26, 2016, https://www.newyorker.com/tech elementsthe-dice-
you-never-knew-you-needed.
Chapter 10: An Oral History of the Death Star
• the team responsible: In reality, I drew heavily from: Ryder Windham, Chris
Reiff, and Chris Trevas, Death Star Owner’s Technical Manual (London:
Haynes, 2013). I wrote this chapter in part to bring a smile to the face of Neil
Shepherd, but don’t tell him I said that.
• architects, engineers, and grand moffs: If you enjoy this kind of thing, I
recommend: Trent Moore, “Death Star Architect Defends Exhaust Ports in
Hilarious Star Wars ‘Open Letter,’” SyFy Wire, February 14, 2014,
http://www.syfy.com/syfywire/death-star-architect-defends-exhaust-ports-
hilarious-star-wars-open-letter%E2%80%99.
• face of a dianoga: My thanks go to Gregor Nazarian, who makes my heart
soar like a John Williams theme, and who helped greatly to improve this chapter.
(I’ve tagged him here because he suggested I use the dianoga, not because his
face resembles one.)
• at a diameter of about 400 kilometers: The rock and ice figures come from
Megan Whewell of the National Space Centre, as quoted by: Jonathan
O’Callaghan, “What Is the Minimum Size a Celestial Body Can Become a
Sphere?” Space Answers, October 20, 2012,
https://www.spaceanswers.com/deep-space/what-is-the-minimum-size-a-
celestial-body-can-become-a-sphere/. The figure for “imperial steel” is my own
blind extrapolation.
• only 140 kilometers across: My research turned up figures ranging from 120
kilometers to 150 kilometers.
• 2.1 million people on the Death Star: I searched and searched to find a more
realistic figure, but the sources all seem to agree. Star Wars fans have debated
this question (https://scifi.stackexchange.com/questions/36238/what-is-the-total-
number-of-people-killed-on-the-2-death-stars-when-they-explode) and tend to
arrive at figures between 1 and 2 million. Windham et al. give roughly 1.2
million people and 400,000 droids. Wikipedia gives 1.7 million people and
400,000 droids. To prove the robustness of my point, I’ve taken the highest
numbers I could find.
• exploiting a thermal vent: I’m not sure that I’ve successfully shoehorned this
narrative into the new continuity established by Rogue One. No doubt certain
stubborn and admirable purists will be upset by the liberties here taken. On the
other hand, I’ve given these characters from “a galaxy far, far away” access to
2010 census data on West Virginia, so maybe redistributing the architectural
credit isn’t my worst crime against canon.
• I was pressed into teaching 10th-grade biology: Witness this typical scene
from April 2010:
KISA (with bright-eyed curiosity): What exactly happens in the endoplasmic
reticulum?
ME: There’s no way of knowing. It is an unsolvable enigma, beyond human
imagination.
TIM (in a bored monotone): The textbook says that’s where proteins are
folded.
ME: Well, obviously, Tim. I meant aside from that.
• The riddles of resemblance strike at the heart of biology: For a take that’s
more sophisticated than mine but still plenty readable, see: Razib Khan, “Why
Siblings Differ Differently,” Gene Expression, Discover, February 3, 2011,
http://blogs.discovermagazine.com/gnxp/2011/02/why-siblings-differ-
differently/#.Wk7hKGinHOi.
• unfolding in very few ways: This same logic underlies the concept of entropy,
the universe’s tendency toward disorder.
Take a collection of bricks. There are precious few ways to form a building,
and many dull, indistinguishable ways to form a pile of rubble. Over time,
random changes will accumulate, almost all of them making your arrangement
more rubble-y, and almost none making it more building-y. Thus, time favors
rubble over bricks.
Similarly, there are very few ways for the particles of food dye to gather on
one side of a water glass; it’d be like the molecules all flipping heads. But there
are many, many ways for those particles to scatter more or less evenly through
the liquid; each such scattering is like a different combination of heads and tails.
That’s why random processes lead in erratic but inexorable ways toward greater
entropy, an even mixing of the universe’s ingredients. The cosmic preference for
disorder is, at its heart, combinatorial.
• between 16 and 30 heads: This has a probability around 96%, so for one in 25
readers, my prophecy will prove false. That said, if as many as 25 readers
actually go flip 46 coins, then the math-book-reading audience is even more
hard-core than I suspected.
• Blaine Bettinger: The graph here comes straight from him. See more at:
https://thegeneticgenealogist.com/wp-content/uploads201606/Shared-cM-
Project-Version-2-UPDATED-1.pdf.
The x-axis is labeled “centimorgans,” which are (to me, anyway) science’s
most confusing unit. A centimorgan is a length of chromosome that has a 1%
chance of being broken up by a chromosomal crossover in any given generation.
With close relatives, you’ll share lots; with distant relatives, very few. Thus,
“centimorgans shared” is a measure of genetic proximity.
So far, so good. But because crossing over happens with different
probabilities all over the genome, a centimorgan is not a constant length. Where
crossing over is common, centimorgans are short; where it’s rare, they’re long.
On top of that, different DNA-sequencing companies divide the human genome
into different numbers of centimorgans. And did I mention that 100
centimorgans do not make a morgan?
A further puzzle: when I converted the centimorgans in this graph to
percentages, I found the distribution centered not around 50%, but 75%. Why?
My wife, Taryn, explained: It’s because commercial DNA kits can’t distinguish
between having both chromosomes in common and having just one. Each counts
as “matching.” By the logic of coin flips, two siblings will tend to have 50% of
their DNA single-matching, 25% double-matching, and 25% not matching at all.
Thus, 75% will match singly or doubly. Hence, the graph centered on 75%.
• roughly twice per chromosome: The number I found was 1.6. It’s higher for
women than for men. In any case, my tripling is a gross underestimate of
possible genomes, because crossing over can happen (in theory) at any point
along the DNA sequence, which allows for myriad additional possibilities. For
more, see Ron Milo and Rob Phillips, “What Is the Rate of Recombination?”
Cell Biology by the Numbers, http://book.bionumbers.org/what-is-the-rate-of-
recombination/.
Chapter 13: What Does Probability Mean in Your Profession?
• Actual Probability vs. How People Treat It: Adapted from Kahneman,
Thinking Fast and Slow, 315.
• Donald Trump defeated Hillary Clinton: For more on this outcome, see
99.997% of all content published on social media since.
• your job was pretty boring: Michael Lewis, Liar’s Poker: Rising Through the
Wreckage on Wall Street (New York: W. W. Norton, 1989).
• You’re increasing the odds: Nate Silver, The Signal and the Noise: Why So
Many Predictions Fail—but Some Don’t (New York: Penguin Books, 2012),
135–37.
Chapter 14: Weird Insurance
• insurance is a strange and silly thing: For this chapter, I drew tremendous
help from Shizhou Chen, my former 10th-grade student and current intellectual
superior. The original draft had a more presumptuous title and a less self-aware
intro; Shizhou cut me down to size. “Quirky and funny examples,” she wrote,
“but not exactly the full picture of what insurance is about.” Touché, Shizhou.
• China five millennia ago: Emmett J. Vaughan, Risk Management (Hoboken,
NJ: John Wiley & Sons, 1996), 5.
• “Iran, 400 BCE”: Mohammad Sadegh Nazmi Afshar, “Insurance in Ancient
Iran,” Gardeshgary, Quarterly Magazine 4, no. 12 (Spring 2002): 14–16,
https://web.archive.org/web/20080404093756/; http://www.iran-
law.com/article.php3?id_article=61.
• National Lottery Contingency Scheme: “Lottery Insurance,” This Is Money,
July 17, 1999, http://www.thisismoney.co.uk/money/mortgageshome/article-
1580017/Lottery-insurance.html.
• What might surprise you is the margin: Shizhou makes a good point here: In
niche insurance markets like this, there’s less competition, and thus higher
markups than in big markets like dental or home insurance.
• if you can’t insure ’em, join ’em: Shizhou makes this astute addendum:
“Small business owner? Yup. Large firm? Totally illegal. The accounting
department would rather put $1 million under ‘insurance’ than $5 under
‘lottery.’”
• “multiple-birth insurance”: Laura Harding and Julian Knight, “A Comfort
Blanket to Cling to in Case You’re Carrying Twins,” Independent, April 12,
2008, http://www.independent.co.uk/money/insurance/a-comfort-blanket-to-
cling-to-in-case-youre-carrying-twins-808328.html. The quote from David Kuo
also comes from this piece.
• financial and psychological: For similar analysis, see: “Insurance: A Tax on
People Who Are Bad at Math?” Mr. Money Mustache, June 2, 2011,
https://www.mrmoneymustache.com/2011/06/02/insurance-a-tax-on-people-
who-are-bad-at-math/. As the Mustache man writes: “Insurance of all types—
car, house, jewelry, health, life—is a crazy field swayed by lots of marketing,
fear, and doubt.”
• alien-abduction policy: Check it out at http://www.ufo2001.com. My source:
Vicki Haddock, “Don’t Sweat Alien Threat,” San Francisco Examiner, October
18, 1998, http://www.sfgate.com/news/article/Don-t-sweat-alien-threat-
3063424.php.
• selling abduction coverage in earnest: Teresa Hunter, “Do You Really Need
Alien Insurance?” Telegraph, June 28, 2000,
http://www.telegraph.co.uk/finance/4456101/Do-you-really-need-alien-
insurance.html.
• the following criteria: These are my own invention. Shizhou lent me her notes
from an undergraduate module, and the list of “characteristics of an insurable
risk” differed slightly:
1. “Potential loss is significant enough that people would exchange
premium for coverage.”
2. “Loss and its economic value is well-defined and out of the
policyholder’s control.”
3. “Covered losses reasonably independent among policyholders.”
• Hole-in-One Prize Insurance: Two such companies are Insurevents
(http://www.insurevents.com/prize.htm) and National Hole-in-One
(http://holeinoneinsurance.co.uk).
• Jordan’s Furniture ran a glitzy promotion: Scott Mayerowitz, “After Sox
Win, Sofas Are Free,” ABC News, October 29, 2007,
http://abcnews.go.com/Business/PersonalFinance/story?id=3771803&page=1.
• the company Wedsure: See its policies at http://www.wedsure.com.
• reimbursing for weddings canceled: Haddock, “Don’t Sweat Alien Threat.”
• began offering to reimburse third-party: Amy Sohn, “You’ve Canceled the
Wedding, Now the Aftermath,” New York Times, May 19, 2016,
https://www.nytimes.com201605/22/fashion/weddings /canceled-weddings-
what-to-do.html.
• Become so expert in the risk: This helps broaden your business, too, by
allowing you to serve as a sort of risk consultant. “Special expertise is one of the
reasons firms buy insurance,” Shizhou told me. “When filming movies, there’s
always an insurance inspector to make sure actors are safe. Without them,
movies would have more mad and pointless explosions than what we see today.”
• “reinsurance”: Olufemi Ayankoya, “The Relevance of Mathematics in
Insurance Industry,” paper presented February 2015.
• insure against career-ending injury: “Loss-of-Value White Paper: Insurance
Programs to Protect Future Earnings,” NCAA.org,
http://www.ncaa.org/about/resources/insurance/loss-value-white-paper.
• purchasing an additional plan: Andy Staples, “Man Coverage: How Loss-of-
Value Policies Work and Why They’re Becoming More Common,”
SportsIllustrated.com, January 18, 2016, https: //www.si.com/college-
football201601/18/why-loss-value-insurance-policies-becoming-more-common.
• Jake Butt: A great name, right? Will Brinson, “2017 NFL Draft: Jake Butt
Goes to Broncos, Reportedly Gets $500K Insurance Payday,” CBS Sports, April
29, 2017, https://www.cbssports.com/nfl/news/2017-nfl-draft-jake-butt-goes-to-
broncos-reportedly-gets-500k-insurance-payday/.
• health insurance: You could fill a bookshelf with this stuff. I recommend the
work of Vox journalist Sarah Kliff, https://www.vox.com/authors/sarah-kliff.
Shizhou recommends the two-part This American Life episode: “More Is Less”
(#391), and “Someone Else’s Money” (#392),
http://hw3.thisamericanlife.org/archive/favorites/topical.
Chapter 15: How to Break the Economy with a Single Pair of Dice
• What is the basic activity of Wall Street banks?: Far and away the most
essential source for this chapter was: David Orrell and Paul Wilmott, The Money
Formula: Dodgy Finance, Pseudo Science, and How Mathematicians Took Over
the Markets (Hoboken, NJ: John Wiley & Sons, 2017).
• make educated guesses: Okay, in my case, it’s more like “make up random
probabilities.” Wall Street employs two more serious approaches. First, they
anchor on historical data. And second, they look at the market price for similar
bonds and use this to infer what the probability of default must be. This latter
method can create spooky dependencies and feedback loops: Instead of
exercising your own judgment, you echo the perceived wisdom of the market.
Put a pin in that idea.
• collateralized debt obligations: Michael Lewis, The Big Short: Inside the
Doomsday Machine (New York: W. W. Norton, 2010). Or, if you prefer
listening to reading: “The Giant Pool of Money,” This American Life, episode
#355, May 9, 2008A. This episode launched the essential Planet Money podcast.
• surrealist painter René Magritte: I saw these sketches in Belgium’s lovely
Musée Magritte on June 4, 2017. I recommend a visit if you’re in Brussels and
would like some surrealism to go with your frites.
• an idiosyncratic risk or a systematic one: These are the standard finance
terms, brought to my attention by Jessica Jeffers, to whom I owe a huge thank-
you for her help with this chapter. I am only 37% joking when I say that Jess is
my choice for Fed chair.
• the notorious “Gaussian copula”: Another essential source for this chapter,
and this discussion in particular, is: Felix Salmon, “Recipe for Disaster: The
Formula That Killed Wall Street.” Wired, February 23, 2009,
https://www.wired.com/2009/02 wp-quant.
• CDS. It stood for “credit default swap”: Also, “complete damn stupidity.”
• How could Wall Street be so stupid?: See another essential source for this
chapter: Keith Hennessey, Douglas Holtz-Eakin, and Bill Thomas, “Dissenting
Statement,” Financial Crisis Inquiry Commission, January 2011, https://fcic-
static.law.stanford.edu/cdn_media/fcic-reports/fcic_final reporthennessey_holtz-
eakin_thomas_dissent.pdf.
• “the wisdom of the crowd”: James Surowiecki, The Wisdom of Crowds: Why
the Many Are Smarter than the Few and How Collective Wisdom Shapes
Business, Economies, Societies, and Nations (New York: Anchor Books, 2004).
• 1987 stock crash: Orrell and Wilmott, 54.
• citizens that H. G. Wells envisioned: Thanks to Richard Bridges for (1) his
help with this chapter; and (2) being a Platonist, a pragmatist, a teacher, and a
brilliant mind, and for proving by construction that all those things can coexist.
• Take household income: All data from Wikipedia. For you, dear reader, only
the best.
• the tasty £2-per-jar tomato sauce: Loyd Grossman’s. They make a solid jar
of tikka masala, too.
• take a square root at the end: My students tend to find it peculiar (and
needlessly complicated) to square the distances, average them, and then take the
square root. Why not just average the distances and leave it at that? Well, you’re
welcome to do this; the result is called the “mean absolute deviation,” and it
serves much the same role as standard deviation. However, it lacks some nice
theoretical properties. Variances can be added and multiplied with ease; this
makes them vital for building statistical models.
• how the correlation coefficient actually works: Okay, buckle up! Things are
about to get hectical, which is my portmanteau of “hectic” and “technical.” To
begin, take a scatter plot—say, height against weight. Represent each person as a
dot.
Now, find the average height and average weight in the population.
Next, take an individual person. How far is that person from the average
height? How far from the average weight? Count “above average” as positive
and “below average” as negative.
Then—and this is the crucial step—multiply these two values.
If the person is above average in both, you’ll get a positive result. Same if
they’re below average in both (because the product of two negatives is positive).
But if they’re above average in one and below average in the other, then you’ll
get a negative result (because the product of a negative and a positive is
negative).
You do this for each and every person, then average together all the products.
The result is something called “covariance” (a cousin of the regular variance).
You’re almost done! As a last step, divide this number, to yield a final result
between -1 and 1.
(Divide by what? Well, consider the shortcoming of covariance: If people’s
weights and heights are both very spread out, then “distance from the average”
will typically be a large number. In other words, covariance is larger for volatile
variables and lower for stable ones, regardless of the relationship between them.
How do we fix this problem? Just divide by the variances themselves.)
Whew! Now for the easy part: interpreting the values.
A positive value (e.g., 0.8) suggests that people above average on one
variable (e.g., height) are usually above average on the other (i.e., weight). A
negative value (e.g., -0.8) suggests the opposite: it’s a population dominated by
tall, light people and short, heavy ones. Finally, a value close to zero suggests
there’s no meaningful relationship at all.
Chapter 17: The Last .400 Hitter
• an Englishman named Henry Chadwick: As quoted in: “Henry Chadwick,”
National Baseball Hall of Fame, http://baseballhall.org/hof/chadwick-henry.
“Every action,” he said of baseball, “is as swift as a seabird’s flight.” As my
friend Ben Miller points out, this raises a question: African swallow, or
European?
• all-time record is 400: This record belongs to West Indies batsman Brian
Lara, who in 2004 scored exactly 400 runs against England without making an
out. I’m grateful for the strange statistical rhyme with this chapter’s title.
• “one true criterion”: As quoted in: Michael Lewis, Moneyball: The Art of
Winning an Unfair Game (New York: W. W. Norton, 2003), 70. It will surprise
no one when I confess that this chapter owes an unpayable debt to Moneyball,
and if you find the story of baseball statistics tolerable (never mind interesting),
then you’ll enjoy the book.
• you’d have to reset the ball at center field every five seconds: I took the
English Premier League’s 38-game season. Calling each game 90 minutes, plus
a generous 10 minutes for stoppage time, that’s 3800 minutes. Twelve data
points per minute (i.e., one per 5 seconds) would give 45,600 data points—still
less than a baseball season’s 48,000, but close enough.
• the first edition of Ernest Hemingway’s: Ernest Hemingway, The Old Man
and the Sea, Life, September 1, 1952. Above the title, it read: “The editors of
Life proudly present for the first time and in full a great new book by a great
American writer.”
• “Goodby to Some Old Baseball Ideas”: Branch Rickey, “Goodby to Some
Old Baseball Ideas,” Life, August 2, 1954. The subhead reads, “‘The Brain’ of
the Game Unveils Formula That Statistically Disproves Cherished Myths and
Demonstrates What Really Wins.”
• in 1858, “called strikes” were born: E. Miklich, “Evolution of 19th Century
Baseball Rules,” 19cBase ball.com, http://www.19cbaseball.com/rules.html.
• Enough such “balls”: It took a long time to settle on how many balls
constituted a walk. At first it took three, then nine, then eight, then six, then
seven, then five, until in 1889 the number finally settled at its current value of
four.
• Walks weren’t deemed an official statistic until 1910: Even then, writers
didn’t exactly embrace them. Just listen to sportswriter Francis Richter:
• Source: Bill James, The New Bill James Historical Baseball Abstract (New
York: Free Press, 2001), 104.
• hitters walk 18% or 19% of the time: In 2017, for example, the league leader
was Joey Votto, who walked 134 times in 707 plate appearances. That’s 19% of
the time.
• just 2% or 3%: In 2017, Alcides Escobar walked 15 times in 629 plate
appearances, for a rate of 2.4%. Tim Anderson outdid him by walking 13 times
in 606 plate appearances, or 2.1%.
• OPS: As a math teacher, I’ve always hated this junkyard stat, because I lose
life force every time someone adds two fractions with different denominators. I
always wished they’d calculate a new statistic, “bases per plate appearance”—
just like SLG, except you count walks on par with singles. Researching this
chapter, I realized my folly: Although cleaner in concept, this new statistic
would actually be less predictive in practice. For the 2017 numbers, its
correlation with a team’s run scoring is 0.873. That’s worse than OBP.
• On the 50th anniversary of the Life article: Alan Schwarz, “Looking Beyond
Batting Average,” New York Times, August 1, 2004,
http://www.nytimes.com/2004/08/01/sports/keeping-score-looking-beyond-
batting-average.html.
• The baseball dialogue in The Old Man and the Sea: For example, “‘I would
like to take the great DiMaggio fishing,’ the old man said. ‘They say his father
was a fisherman.’”
• writer Bill James: Scott Gray, The Mind of Bill James: How a Complete
Outsider Changed Baseball (New York: Three Rivers Press, 2006). The book
has some choice Bill James aphorisms, including: “There will always be people
who are ahead of the curve, and people who are behind the curve. But
knowledge moves the curve.” Also: “When you add hard, solid facts to a
discussion, it changes that discussion in far-reaching ways.”
• batting average is a sideshow: And yet, through the 1970s, official MLB
statistics listed team offenses in order of batting average, not runs scored. “It
should be obvious,” quipped James, “that the purpose of an offense is not to
compile a high batting average.”
• salaries soared: Michael Haupert, “MLB’s Annual Salary Leaders Since
1874,” Outside the Lines (Fall 2012), Society for American Baseball Research,
http://sabr.org/research/mlbs-annual-salary-leaders-1874-2012. I cheat big-time
in the graph by not adjusting for inflation; for example, Joe DiMaggio’s $90,000
salary in 1951 would be more like $800,000 in 2017 dollars. Even so, it’s hard to
argue with the impact of free agency.
• “Baseball is the only field of endeavor”: Pete Palmer, The 2006 ESPN
Baseball Encyclopedia, (New York: Sterling, 2006), 5.
• a heartbreaking .39955: Bill Nowlin, “The Day Ted Williams Became the
Last .400 Hitter in Baseball,” The National Pastime (2013), Society for
American Baseball Research, https://sabr.org/research /day-ted-williams-
became-last-400-hitter-baseball.
• no one has hit .400 since: Ben Miller (a gentleman, a culinary hero, and an
incurable Sox fan) points out that Williams’s success went far beyond that
number. In 1941, he “had an OBP of .553, which stood as the single-season
record for more than 60 years… Also, Ted Williams’s career OBP of .482 is the
highest ever. Oh, also, Ted Williams’s career batting average of .344 is 6th-
highest all-time, and you have to go all the way to Tony Gwynn at #17 to find
the next all-time BA leader who played after 1940.” No matter how you slice it,
the Splendid Splinter was good at his job.
• “If I had known hitting .400”: Bill Pennington, “Ted Williams’s .406 Is More
Than a Number,” New York Times, September 17, 2011,
http://www.nytimes.com/2011/09/18/sports/baseball/ted-williamss-406-average-
is-more-than-a-number.html.
Chapter 18: Barbarians at the Gate of Science
• In 1982, Jay Mathews was the Los Angeles bureau chief: Jay Mathews,
“Jaime Escalante Didn’t Just Stand and Deliver. He Changed U.S. Schools
Forever,” Washington Post, April 4, 2010, http://www.washingtonpost.com/wp-
dyn/content/article/2010/04/02/AR2010040201518.html.
• “sheer mockery”: “Mail Call,” Newsweek, June 15, 2003,
http://www.newsweek.com/mail-call-137691.
• “with the hope that people will argue”: Jay Mathews, “Behind the Rankings:
How We Build the List,” Newsweek, June 7, 2009,
http://www.newsweek.com/behind-rankings-how-we-build-list-80725.
• heed the tale of British ambulances: Tim Harford, Messy: How to Be
Creative and Resilient in a Tidy-Minded World (London: Little, Brown, 2016),
171–73.
• Teacher’s Value Added: This one is real: Cathy O’Neil, Weapons of Math
Destruction: How Big Data Increases Inequality (New York: Broadway Books,
2016), 135–40.
• “Nearly every professional educator”: Jay Mathews, “The Challenge Index:
Why We Rank America’s High Schools,” Washington Post, May 18, 2008,
http://www.washingtonpost.com/wp-
dyn/content/article/2008/05/15/AR2008051502741.html.
• “passer rating”: I’ve been watching football since I was seven, and have
never understood passer rating. I figured it was time to give it a shot.
The formula took a few minutes to unpack, but once I did, I saw that it isn’t
all that complicated. First, it assigns points (1 per yard, 20 per completion, 80
per touchdown, and -100 per interception). Second, it computes the number of
points per passing attempt. And third, it tacks on a meaningless
addition/multiplication flourish. Or, in equation form:
We could leave it there… except this formula allows for negative scores (if
you throw too many interceptions) and has an unattainable maximum (whereas
the actual maximum of 158¹/³ has been achieved in games by more than 60
quarterbacks). To fix this, you need to cap the points contributed by each stat,
and for that, it’s easier to present the formula as it is in the main text.
Why does passer rating feel so confusing? As with most scary formulas, two
factors come into play: (1) it is confusing, being a strangely weighted average of
four variables, each subject to arbitrary cutoffs; and (2) on top of that, most
sources present it in a needlessly counterintuitive and opaque way. Look on
Wikipedia if you want to see what I mean.
• “One of its strengths is the narrowness”: Mathews, “Behind the Rankings.”
• “The list has taken on a life of its own”: J. P. Gollub et al., eds., “Uses,
Misuses, and Unintended Consequences of AP and IB,” Learning and
Understanding: Improving Advanced Study of Mathematics and Science in U.S.
High Schools, National Research Council (Washington, DC: National Academy
Press, 2002), 187, https://www.nap.edu/read/10129 chapter12#187.
• “The parents are making the most noise”: Steve Farkas and Ann Duffett,
“Growing Pains in the Advanced Placement Program: Do Tough Trade-Offs Lie
Ahead?” Thomas B. Fordham Institute (2009),
http://www.edexcellencemedia.net/publications/2009/200904_growingpainsintheadvancedpla
According to the survey, more than one in five AP teachers nationwide
believed that the Challenge Index had had “some impact” on their school’s AP
offerings. In suburbs and cities, the number was closer to one in three.
Meanwhile, just 17% viewed the list as “a good idea.”
• “Because your index only considers the number”: Valerie Strauss,
“Challenging Jay’s Challenge Index,” Washington Post, February 1, 2010,
http://voices.washingtonpost.com/answer-sheet/high-school/challenged-by-jays-
challenge-i.html.
In 2006, the Challenge Index ranked Eastside High in Gainesville, Florida, as
the #6 school in the US. But only 13% of Eastside’s nearly 600 African
American students were reading at grade level. Schools #21 and #38 showed
similar incongruities. Critics saw this as a sign that unprepared students were
being shoveled into college-level courses in an effort to catch Newsweek’s eye.
Source: Michael Winerip, “Odd Math for ‘Best High Schools’ List,” New York
Times, May 17, 2006,
http://www.nytimes.com/2006/05/17/education/17education.html.
• Who, exactly, is the index for?: John Tierney, “Why High-School Rankings
Are Meaningless—and Harmful,” Atlantic, May 28, 2013,
https://www.theatlantic.com/national/archive/2013/05/why-high-school-
rankings-are-meaningless-and-harmful 276122.
• “We cannot resist looking at ranked lists”: Mathews, “Behind the
Rankings.”
• “we are all tribal primates”: Jay Mathews, “I Goofed. But as Usual, a Smart
Educator Saved Me,” Washington Post, June 25, 2017,
https://www.washingtonpost.com/local/education/i-goofed-but-as-usual-a-smart-
educator-saved-me/2017/06/25 /7c6a05d6-582e-11e7-a204-
ad706461fa4f_story.html.
• “a very elastic term in our society”: Winerip, “Odd Math for ‘Best High
Schools’ List.”
• Mathews likes to cite a 2002 study: Jay Mathews, “America’s Most
Challenging High Schools: A 30-Year Project That Keeps Growing,”
Washington Post, May 3, 2017, https://www.wash
ingtonpost.com/local/education/jays-americas-mo st-challenging-high-schools-
main-column/2017/05/03 /eebf0288-2617-11e7-a1b3-faff0034e2de_story.html.
• seemed to lay the groundwork for success in college: Not all scholars agree.
The 2010 book AP: A Critical Examination of the Advanced Placement Program
found a consensus building among researchers that the AP expansion had begun
to see diminishing returns. Coeditor Philip Sadler said, “AP course work does
not magically bestow advantages on underprepared students who might be better
served by a course not aimed at garnering college credit.” Source: Rebecca R.
Hersher, “The Problematic Growth of AP Testing,” Harvard Gazette, September
3, 2010, https://news.harvard.edu/gazette/story/2010/09 philip-sadler.
• As of 2017, that number is 12%: Mathews, “America’s Most Challenging
High Schools.”
• “the view that schools with lots of rich kids are good”: Mathews, “The
Challenge Index.”
• how you feel about his vision: I have a humble suggestion for Mathews:
Instead of counting “exams taken,” count “scores of at least 2.” In my
experience (and in the Texas study that Mathews likes to cite) a 2 shows some
intellectual spark, a signal of growth. But I’m not convinced the typical 1-scorer
has gained from the class. Requiring students to score above the minimum
would eliminate the incentive to throw wholly unprepared kids into the test.
Chapter 20: The Book Shredders
• Ben Blatt’s delightful book: Ben Blatt, Nabokov’s Favorite Word Is Mauve
(New York: Simon & Schuster, 2017).
• the greatest novels: I computed the “greatness score” from data on Goodreads,
where readers give ratings from one to five stars. First, I calculated the total
number of stars awarded to each book. Faulkner’s ranged from under 1500 (for
Pylon) to over 500,000 (for The Sound and the Fury). Then I took a logarithm,
which collapses this exponential scale down to a linear one. The correlation
between adverbs and “greatness” was -0.825. Equivalent analyses for
Hemingway and Steinbeck yielded coefficients of -0.325 and -0.433: substantial,
but hard to discern on a graph. The adverb data comes from Blatt, and the
method is a variation on his. (He uses number of ratings instead of number of
stars; it yields pretty much the same result.)
• a blockbuster research article: Jean-Baptiste Michel et al., “Quantitative
Analysis of Culture Using Millions of Digitized Books,” Science 331, no. 6014
(2011): 176–82,
http://science.sciencemag.org/content/early/2010/12/15/science.1199644. They
wrote:
The corpus cannot be read by a human. If you tried to read only the
entries from the year 2000 alone, at the reasonable pace of 200
words/minute, without interruptions for food or sleep, it would take eighty
years. The sequence of letters is one thousand times longer than the
human genome; if you wrote it out in a straight line, it would reach to the
moon and back 10 times over.
• “New Window on Culture”: Patricia Cohen, “In 500 Billion Words, New
Window on Culture,” New York Times, December 16, 2010,
http://www.nytimes.com/2010/12/17/books/17words.html.
• “precise measurement”: Michel and Aiden wrote (emphasis mine):
The weight, the pace, the stride of a man’s mind are too unlike [a
woman’s] for her to lift anything substantial from him… Perhaps the first
thing she would find, setting pen to paper, was that there was no common
sentence ready for her to use.
Though she enjoys that masculine style (“swift but not slovenly, expressive
but not precious”), she added, “It was a sentence that was unsuited for a
woman’s use.”
Charlotte Brontë, with all her splendid gift for prose, stumbled and fell
with that clumsy weapon in her hands… Jane Austen looked at it and
laughed at it and devised a perfectly natural, shapely sentence proper for
her own use and never departed from it. Thus, with less genius for writing
than Charlotte Brontë, she got infinitely more said.
They took copies of each essay and dissected them, cutting the words
apart and arranging them (by hand) in alphabetical order. At one point
Mosteller and Wallace wrote, “during this operation a deep breath
created a storm of confetti and a permanent enemy.”
Nothing is more useful than water: but it will purchase scarce any thing;
scarce any thing can be had in exchange for it. A diamond, on the
contrary, has scarce any value in use; but a very great quantity of other
goods may frequently be had in exchange for it.
From: Adam Smith, An Inquiry into the Nature and Causes of the Wealth of
Nations (1776), book I, chapter IV, paragraph 13, accessed through the online
Library of Economics and Liberty:
http://www.econlib.org/library/Smith/smWN.html.
• Classical economics lasted a century: My go-to source for this chapter was
this wonderful book: Agnar Sandmo, Economics Evolving: A History of
Economic Thought (Princeton, NJ: Princeton University Press, 2011).
• This idea long predates marginalism: Campbell McConnell, Stanley Brue,
and Sean Flynn, Economics: Principles, Problems, and Policies, 19th ed. (New
York: McGraw-Hill Irwin, 2011). The “earth’s fertility” quote comes from
section 7.1, Law of Diminishing Returns,
http://highered.mheducation.com/sites/0073511447/student_view0/chapter7/origin_of_the_ide
• The earth’s fertility resembles a spring: As Mike Thornton points out—and I
humbly thank him for his help with this chapter—this analogy elides some
subtleties. Heterogeneous land can be perfect for growing various crops that
prefer different conditions, and farmers can take steps (such as rotating crops) in
order to improve the quality of the soil.
• This function of utility is peculiar: William Stanley Jevons, “Brief Account
of a General Mathematical Theory of Political Economy,” Journal of the Royal
Statistical Society, London 29 (June 1866): 282–87,
https://socialsciences.mcmaster.ca/econ ugcm3ll3/jevons/mathem.txt.
• “A true theory of economy”: Jevons again. Quotable guy. Writing this
chapter made me a fan.
• In school, I learned that economics has mirror symmetry: For a refresher, I
drew from the lecture notes of Michal Brzezinski, an assistant professor of
economics at the University of Warsaw,
http://coin.wne.uw.edu.pl/mbrzezinski/teaching/HEeng/Slides/marginalism.pdf.
• self-described “standup economist”: Yoram Bauman. I quote from his video
Mankiw’s Ten Principles of Economics, Translated, which cracked me up when
I first saw it in college. Check out his site at http://standupeconomist.com/.
• “The proper mathematical analogy”: Sandmo, Economics Evolving, 96.
• “persist in using everyday language”: Sandmo, 194.
• “the only work by an economist”: The quote is from Wikipedia. Shhh, don’t
tell anyone.
• renewed empiricism: Jevons embodied this trend. He predicted that Britain
would soon exhaust its coal resources, and argued that the ups and downs of the
business cycle stemmed from cooler temperatures caused by sunspots. Okay, so
he was wrong about both, but we know that thanks to methods that he himself
ushered in.
Walras, by contrast, was a bit of an anti-empiricist. In his view, you must first
filter out the messy details of reality to arrive at pure quantitative concepts.
Second, you must operate on and reason about these mathematical abstractions.
Third, almost as an afterthought, you look back to practical applications. “The
return to reality,” Walras wrote, “should not take place until the science is
completed.” To Walras, “science” was something that happens far away from
reality.
If you meet an economist who still thinks this way today, just follow these
simple steps: (1) raise your arms; (2) begin bellowing; (3) if the economist
continues to approach, box him on the ears. Remember, these economists are as
frightened of us as we are of them.
On July 25, Madison raised the point again—although this time, he endorsed
direct election as his favorite method, saying: “As an individual from the S.
States he was willing to make the sacrifice.” Read the details for yourself in
“Notes on the Debates in the Federal Convention”:
http://avalon.law.yale.edu/subject_menus/debcont.asp.
Further evidence that the slavery advantage was a feature, not a bug: in his
first inaugural address in 1833, Southerner Andrew Jackson proposed replacing
the Electoral College with a direct election—but only if the irregularities in vote
weighting were preserved. Source: Amar, 347.
• Here’s the elector count from 1800: I drew census data from Wikipedia.
Then, I redistributed the number of representatives in the House in proportion to
the free population of each state.
• bucket-filled with red and blue: Koslig informed me that the “Democrat =
blue, Republican = red” color scheme was once the opposite. It shifted in the
1990s and was cemented in 2000. For more, you can read a fun Smithsonian
piece: Jodi Enda, “When Republicans Were Blue and Democrats Were Red,”
Smithsonian.com, October 31, 2012, https://www.smithsonianmag.com/history
when-republicans-were-blue-and-democrats-were-red-104176297.
• 60% of the vote gets six electors: Even this is not straightforward as it sounds.
Take the 2016 election, where the Democrats’ 46.44% rounds up to 50% (for
five electors), and the Republicans’ 44.92% rounds down to 40% (for four
electors). That’s only nine. Who gets the last elector? You could give it to the
Libertarians, but their 3.84% is pretty far from 10%. It’s probably more logical
to give the last elector to the party that was closest to earning it outright—in this
case, the Republicans, who were just 0.08% away.
• where 96% of the country sits today: As Koslig points out, this dynamic is
different in states that favor one party in presidential elections and the other for
state government. In recent years, a few such states have considered joining
Nebraska and Maine in their way of allocating electors, as a way for the state-
government-ruling party to seize some electors from its rival.
• a beautiful method for answering this question: Nate Silver, “Will the
Electoral College Doom the Democrats Again?” FiveThirtyEight, November 14,
2016, https://fivethirtyeight.com/features/will-the-electoral-college-doom-the-
democrats-again/.
• no correlation between which party has the Electoral College advantage:
“The major exception,” Silver points out,
was in the first half of the 20th century, when Republicans persistently
had an Electoral College advantage because Democrats racked up huge
margins in the South, yielding a lot of wasted votes.… The question is
whether Democrats are reentering something akin to the “Solid South”
era, except with their votes concentrated in more urban coastal states….
He knew now that historians did not study history. No human mind could
encompass the history of Hain: three million years of it…. there had been
uncountable kings, empires, inventions, billions of lives lived in millions
of countries, monarchies, democracies, oligarchies, anarchies, ages of
chaos and ages of order, pantheon upon pantheon of gods, infinite wars
and times of peace, incessant discoveries and forgettings, innumerable
horrors and triumphs, an endless repetition of unceasing novelty. What is
the use trying to describe the flowing of a river at any one moment, and
then at the next moment, and then at the next, and the next, and the next?
You wear out. You say: There is a great river, and it flows through this
land, and we have named it History.
To receive special offers, bonus content, and news about our latest ebooks and
apps, sign up for our newsletters.
Sign Up
Or visit us at hachettebookgroup.com/newsletters