Making Space - Jennifer M. Groh
Making Space - Jennifer M. Groh
SPACE
The Belknap Press of Harvard University Press Cambridge, Massachusetts | London, England | 2014
JENNIFER M. GROH
MAKING
SPACE
How the Brain Knows Where Things Are
CONTENTS
1.
2.
3.
4.
5.
6.
7.
8.
9.
10.
Notes
Credits
Acknowledgments
Index
Color illustrations follow page 86
1
7
51
69
107
143
161
177
189
203
219
225
231
237
MAKING
SPACE
because their shapes differthe cat does not have a handle. Indeed, we
see these things as distinct from each other precisely because they have
boundaries in space. You can see where the cat ends and the rug begins
because the pattern of light changes from one place (the cat) to another
(the rug). If we could not detect that spatial boundary, we would not be
able to tell that the cat and rug are separate entities. Youd be surprised
when a piece of the rug got up and started meowing.
Your understanding of location and boundaries allows you to handle
physical objects and move about in the world. You can reach out and
pick up that coffee mug, or the cat, and when you do, your hands form
an appropriate shape to grasp the object. You can walk around the
coffee table without hitting your shins (usually).
This book is about the magnificent computational power devoted by
the brain to helping you accomplish these ordinary feats of perception
and behavior. It is about how your eyes are like radios and how your
ears tell time, and how you really do have rocks in your head, very
small ones that help you tell which way is up.
It is also about how you know where you are and how you get from
one place to another, and why you might take a shortcut to come back.
Its about why its harder to remember how to get somewhere if youve
always been the passenger than if you have driven that way yourself.
Its about why, when you set off to do something and then forget what
you are up to, going back to the place where you started might help you
remember.
In this book, I will describe how our senses measure physical
energysuch as light, sound, and pressure on the skinand what we
know about how the brain evaluates those measurements to make inferences about the locations of the objects and events occurring around
M A K I N G SPAC E
T H I N K I N G A B O U T S PAC E
your sense of your position in the world. Your brains different sensory
and motor systems all work in concert to produce this sense. Not only
does your memory and knowledge help you know where you are, but
where you are triggers memories and knowledge. So when you have
forgotten whats on your grocery list, going back to the kitchen will
help you remember what you were planning to cook.
The style of approach in this book stems from my background as
a computational and experimental neuroscientist. I view the brain as
something that can be studied and understood from the perspective
of How is it built? What can it do? As an experimental neuroscientist, I (and my students) conduct experiments to measure the activity
of neurons in the brain in response to different sensory stimuli and
during different kinds of behavior. As a computational neuroscientist, I
imagine ways that different neurons might be arranged to figure things
outto make deductions about sensory events or to signal the muscles
to move in response to those events, for example. These form the basis
of models of how the brain computes.
This view of the brain as a biological device for computing gives
the book what I hope is a demystifying flavor. That is, I attempt to
explain exactly what we know and how we know it. We understand
much about how individual neurons work and how the activity of
populations of neurons reflect the sensory events in the environment.
The way neurons respond in turn gives insight into how we perceive.
For example, neurons in the visual pathway seem to be optimized for
locating the boundaries of objects. We think this property helps you to
distinguish the cat against the backdrop of the rug.
You know you really understand how a system works when you can
build one that can do the same thing. At present, this is the holy grail
M A K I N G SPAC E
of brain science. For example, engineers have built cameras with much
higher resolution than the human eye and supercomputers capable of
storing and processing vast amounts of data with astonishing speed.
But even though computers can now routinely forecast the weather and
beat humans at chess, they cannot reliably recognize that your mother
is the same person with and without her sunglasses ona task that is
effortless for the biological computer in your head. The ability of the
brain to conclude that two things are the same despite what may be
very little similarity in the raw datasuch as the spatial pattern of light
and dark and coloris remarkable, and we do not know how the brain
does it. The how would you build it approach can help us identify
these gaps in our understanding.
More broadly, the computational tenor of this book stems from one
of my original interests in neuroscience. To me, the most fascinating
problem in all of science is the question of how neural firing patterns
create thought. I have always wanted to know how our rich mental
lives of thinking, imagining, reasoning, worrying, wanting, planning,
remembering, and deciding are embodied in electrical pulses generated by microscopic cells in the brain. This book will not provide the
answer. But it will provide an empowering analogy: perhaps, if we can
understand how the brain processes information about things we can
measure, like sensory stimuli, we may gain insights into how the brain
processes things we cannot measure, like ideas. Perhaps the neural
mechanisms for reasoning about the tangible and reasoning about the
intangible are similar. Perhaps the brains systems for thinking about
space are also the brains systems for simply thinking.
Nine-tenths of the brain doesnt seem like a crazy estimate after
all...
T H I N K I N G A B O U T S PAC E
and neuroscience, but this has not always been the case. Philosophers,
mathematicians, artists, and astronomers, such as Democritus, Plato,
Euclid, Alhazen, Leonardo da Vinci, Ren Descartes, and Johannes
Kepler, laid the groundwork for the modern era of visual science. These
scholars employed a mixture of imagination, observation, logic, and,
eventually, measurements to deduce the essential features of how the
eye might work.
When philosophers in ancient Greece first began debating how
vision works, they did not have the benefit of our modern view of the
universe as composed of energy and matter. Instead, they thought the
world and everything in it was made up of different mixtures of the
four so-called elements: earth, fire, air, and water. Around 450 BCE,
Democritus and other philosophers began advancing the theory that
all substances were composed of minute particles they termed atoms.
These atoms were the smallest units of a substance and could not be
divided further. They gave a substance its character, so atoms of water
were slippery, atoms of air were lightweight and transparent, and so
forth. For these early Greeksand for centuries afterthe burning
mystery was What, exactly do the eyes see? What is the physical entity
they detect?
Democritus extrapolated from budding atomic theory to propose a
mechanism for vision: he suggested that some kind of tiny material
emanated from objects in the visual scene and entered the eye. This
material was supposed to leach or be thrown off the surfaces of objects
in the scene, as fire gives off smoke and heat or snakes shed their skin.1
Unlike actual material, it would have to be emitted continually without causing the source to waste away. As this stuff entered the eye,
Democritus postulated, it somehow retained the essential character of
M A K I N G SPAC E
the sourceits size, shape, colorwhich could then be felt by the eye
as if by touch.
Democritus got the essential direction of vision correct: whatever
this vision dust was, it was supposed to travel from the object to the
eye, a theory known as intromission. But other Greek philosophers
reversed the process. Plato, for example, believed that some visual
power emerged from the eye instead of the other way around, like
superrays glowing from the eyes of a comic-book action hero. He too
drew a parallel between vision and touch, but emphasized vision as an
outward-reaching process called extramission.
The core of the debate between extramission and intromission
would not be settled for more than a thousand years (Figure 2.1). As the
Greco-Roman Empire declined, the works of the Greek philosophers
fell into obscurity. But the subsequent rise of the Islamic world reenergized such intellectual pursuits, and these works were rediscovered and
translated into Arabic.
Initially, the outreach view held sway. Muslim scholars such as Abu
Yusuf Yaqub ibn Ishaq al-Kindi and Hunain ibn Ishaq subscribed to the
extramission theory chiefly because of flaws in the original formulation of intromission. The sticking point was Democrituss conception
of vision as involving a kind of material particle. If such particles were
felt inside the eye and conveyed information on such features as the
size of the object, wouldnt these particles scale with the size of the
object? How could really large objects, such as mountains, enter small
eyes? And how could the same object enter the eyes of many observers
at once? Such issues required explanation.
Conceptually, these scholars were noticing a point that remains a
focus of modern visual science: seeing can feel like it is something we
T H E WAY S O F L I G H T
FIGURE 2.1 Early Greek scholars debated whether vision involves something entering the eye or leav-
ing it. Plato thought the eyes sent forth some kind of rays to feel the world. Democritus and later
Alhazen thought the opposite. Not until Keplers work in the first part of the seventeenth century was
the optical basis of how the eye forms an image of the visual scene fully understood. This 1685 drawing
illustrates the emerging concept of how this happens.
10
M A K I N G SPAC E
do to the world, rather than something the world does to us. We dont
passively sit back and have the entire scene float into our eyes. Instead,
we cast our gaze about, inspecting different objects as they capture our
interest. As we move our eyes, our view of the scene changes. We see
what we are looking at very clearly, and we are much less aware of the
parts of the visual scene that we are not looking at directly. These early
scholars wondered why moving the eyes should affect what we see if
vision is based on automatic emanations from the scene. Such observations seemed damning to the intromission theory and more consistent
with the eyes as directing rays outward to feel the scene.
The debate finally began to be resolved a few decades after the turn
of the first millennium. Sometime between 1028 and 1038 CE, the
intromission theory was resurrected and revised by Abu Ali al-Hasan
ibn al-Hasan ibn al-Haytham, known in the West as Alhazen.* Alhazen
argued against the Greek versions of both extramission and intromission and introduced his own theory. He rejected the arguments
of the extramissionists on logical grounds. Plato and other adherents
of extramission had suggested that something comes out of the eye,
merges with light, and returns to the eye. Alhazen argued that since
the critical step is the return to the eye, it was not necessary to postulate the initial emanation. This part was not provable, was not required
* Alhazen led an interesting life. He served for a time as a government official in either
Basra or Egypt, but then supposedly feigned insanity to escape his administrative
duties (in what may have been the first instance of a strategy now frequently employed
by some university faculty the world over to avoid committee work). While serving
house arrest for malingering, Alhazen was free to concentrate on his philosophical
studies, authoring more than 180 tracts on subjects as diverse as mathematics, astronomy, metaphysics, logic, and medicine, as well as optics.
T H E WAY S O F L I G H T
11
12
M A K I N G SPAC E
named for their sensitivity to light) line the retina at the back of the eye.
Photoreceptors achieve their light-sensing properties via special molecules known as photopigments. Photopigments are like paint in that they
reflect light of some wavelengths and absorb others. In paint pigments,
the wavelengths that are reflected give the paint its color. In photorececeptor pigments, absorption is limited to electromagnetic radiation
in a narrow range of the electromagnetic spectrum, involving wavelengths between about 400 and 700 nanometers. The signals of radios,
cell phones, microwave ovens, or police radar guns have wavelengths
outside this range. Such signals are invisible to our photopigment molecules and, by extension, to us. With different photopigment molecules,
other species can see electromagnetic radiation slightly outside this
range, specifically down to around 300 nanometers (ultraviolet light)
for some species of birds and bees.
Photopigment molecules act as a kind of microscopic light switch, a
switch that light can flip. This ability stems from the shape of photopigment molecules and how their shape changes when they absorb light. To
set the stage for visualizing this for a complicated molecule like a photo
pigment, Ill first describe what we know about the shape of a simple
molecule, water, to illustrate some of the principles.
The shape of a molecule is governed by what atoms it is made of
and how they are connected. Water molecules are made up of one
oxygen atom and two hydrogen atoms (Figure 2.2), connected in the
sequence hydrogen-oxygen-hydrogen. You can think of these atoms
as beads on a piece of bent wire. The bonds connecting these atoms
form at a particular angle determined by the elbow room needed
by the electrons composing the bond as well as those of other nearby
electrons.
T H E WAY S O F L I G H T
13
electrons
FIGURE 2.2 Schematic drawing of a water molecule, illustrating its three-dimensional shape. The oxygen atom
(O) is larger than the hydrogen atoms (H). It forms two
bonds with these hydrogen atoms by sharing an electron
with each one. The other four electrons of the oxygen
atom form pairs as if they were unconnected bonds. The
two bonds with hydrogen and the two pairs of electrons
can be thought of as sticking out from the oxygen like a
tetrahedron, creating the maximum space between each
spoke.
14
M A K I N G SPAC E
Once made, bonds tend not to break unless energy is applied. For
photopigments, light provides that energy. When light shines on the
photopigment molecule, one of these rigid double bonds breaks for an
instant. The two carbon atoms are still held together, but only by a
single bond. For that brief moment, the bond can rotate. The molecule
is free to spin around the axis of the remaining single bond. The second
bond then reforms, but now the whole molecule is in a different configuration (Figure 2.3).
FIGURE 2.3The light-sensitive molecule
In the dark:
H
C
Additional carbon
and hydrogen atoms
C
C
In light:
C
C
C
H
T H E WAY S O F L I G H T
15
Think of the two carbon atoms as you and a dance partner, facing
each other, clasping each others hands. If you let go of one hand, you
can spin. When you grab hold again, perhaps you are now back to front,
and what was once on your left is now on your right. Thats the essence
of the kind of rotation that is possible with a single bond.
Now imagine you are a four-handed dancing alien, using two hands
to grip your four-handed partner and the other two to grasp two more
chains of dancers-cum-atoms. When you release one hand and spin,
you fling these chains of atoms around too, like a microscopic game of
crack the whip. Before you spin, the chain on your left was adjacent to
your partners right-side chain and vice versa. Now, after youve spun,
theyve switched places. The spatial positions occupied by you and all
the other atoms in the molecule determine its shape. So, if you move
them around, the whole molecule changes shape.
This partial breakage of the bond and the ensuing rotation and reformation of the double bond is the only part of the process of seeing that
is directly triggered by light, and it takes only one photon of light to
accomplish this for a given molecule of photopigment. The change in
the shape of an individual molecule of photopigment ultimately affects
large numbers of other molecules, amplifying the effects of an individual photon. It does so by triggering a change in a catalyst, a general term
for any kind of substance that facilitates a chemical reaction without
being altered or consumed by it. Proteins that serve this function are
more commonly called enzymes.
Because catalysts such as enzymes are not consumed by the chemical
reactions they facilitate, they can be used over and over. You can think of
them as the machines in a factory. Imagine a machine in a soda-bottling
plant. The bottle-capping machine must first be turned on, but once
16
M A K I N G SPAC E
operational, it can cap many bottles. At the end of the day, thousands of
bottles will have been capped, but the machine itself will be unchanged.
The catalyst affected by the change in shape of the photopigment
is a protein molecule called opsin. Like all proteins, opsin is built from
components known as amino acids. The sequence of amino acids in a
protein (and thus the shape of that protein) is specified by the DNA in
your genes. The ability of a protein to help a reaction depends on its
conformationjust like the bottle-capping machine with its specially
shaped clamps for the bottle and cap. The parts of protein molecules
that grasp onto other molecules are called active sites. These active sites
must be unblocked before the protein can do its work.
Thats the case with opsin when it is connected to the photopigment in the dark. With the photopigment in its original shape, access
to opsins active site is blocked. When light changes the shape of the
photopigment molecule, it no longer fits into its slot linked to the opsin
molecule. Like the proverbial square peg in a round hole, the photopigments new, more awkward shape causes its connection to opsin to be
severed. The photopigment molecules role in detecting light is done,
but opsins work is just beginning. Opsin, no longer encumbered by the
photopigment molecule, also changes shape. Its active site is now accessible, and it is able to catalyze the reaction it is suited for. Opsin catalyzes
a reaction that, in turn, activates another protein catalyst, which activates another, and so forth. At each step, more and more molecules are
involved, magnifying the impact of the initial photon absorption.*
* If youd like the details: the reaction opsin catalyzes involves another protein, transducin, which is normally bound to a molecule known as GDP (guanosine di-phosphate). With the help of opsin, the transducin swaps out the GDP molecule for a
T H E WAY S O F L I G H T
17
different molecule, GTP (guanosine triphosphate). This change causes the transducin
itself to change shape and break apart into several pieces. One of the fragments of
the transducin molecule then binds to a third protein, phosphodiesterase. This has
the effect of changing the shape of phosphodiesterase, exposing an active site in this
molecule. The newly activated phosphodiesterase catalyzes another reaction, this
time involving cGMP (a circular molecule called cyclic guanosine monophosphate),
which is converted into something else (an unwrapped version, called 5-guanosine
monophosphate).
18
M A K I N G SPAC E
positive
0- Rest
-
negative
Not rest.
Neuron responding.
Rod photoreceptors
Time
FIGURE 2.4 If you compare the voltage inside versus outside a neuron, you will find that
the inside is usually negative (has more negatively charged ions or fewer positively
charged ions) compared to the outside. The voltage difference across the neurons
membrane when nothing is happening, i.e., when there are no sensory stimuli or no
incoming signals from other neurons, is called the resting membrane potential. When
sensory events or other inputs cause that voltage difference to change, the neuron
is said to be responding. Light causes the resting membrane potential of photoreceptors to dip and become slightly more negative. Neurons can also respond to their
inputs with positive changes in their resting membrane potential.
electrical charge) into and out of the neuron. Known as ion channels,
these pores let different kinds of ions flow through the membrane. For
example, some ion channels permit the passage of sodium ions, and
others allow potassium or chloride to pour through. Because such ions
have an electrical charge, when they move into or out of the neuron,
they change the voltage across its membrane.
Ion channels can be either opened or closed like a gate. The relevant ion channel in photoreceptors is a sodium channel whose gate is
regulated by a messenger molecule. When the messenger molecule is
plentiful, most of these ion channel gates are open, and sodium gushes
into the cell. When scarce, the gate slams shut, staunching the sodium
T H E WAY S O F L I G H T
19
20
M A K I N G SPAC E
T H E WAY S O F L I G H T
21
do this much only because they usually have multiple pigment spot
ocelli on different sides of the body surface, allowing them to compare
light levels in different directions).
In vertebrates, however, photoreceptors are not external but internal, lining the back surface of the spherical eye. The shape of the eye
and its component partsspecifically the pupil, the lens, and the eyes
fluidsregulate the path light takes to reach the retinal photoreceptor
sheet. Collectively, these structures control the spatial pattern of light
reaching the retina, producing a replica of the pattern of light in the
environment with its spatial relations intact.
Deducing how the eye forms such an image took centuries, again
beginning in ancient Greece, this time with Euclids work on geometry,
and culminating years later in Renaissance Europe. Then and there,
a breakthrough by the astronomer Johannes Kepler revealed that the
pupil joins forces with the crystalline and aqueous contents of the eye
to both prune and bend light into a sharp and clear reflection of the
visual world. This works because light travels in (reasonably) straight
lines. Lights straightness makes spatial vision possible, but only if you
have some way of telling where the lines are coming from.
Euclid was the first to emphasize the straight lines of vision. This
had not been immediately obvious. As will be discussed in Chapter 5,
we dont hear in straight linessound can curve around obstacles
and both sound and light can reflect off of appropriate surfaces. So, the
concept of vision involving straight lines was not as readily apparent as
it might seem to us today. Euclid pointed out that where and how large
an object will appear depends on the angle formed by these visual
rayslines that connect the objects boundary with the eye (Figure
2.5). When an object is large, the angle traced out by rays connecting
22
M A K I N G SPAC E
FIGURE 2.5 Rays of light emanating from different parts of an object arrive at the eye at different directions, forming an angle. Although Euclid did not know what light is (or which direction it travels), he
nevertheless pointed out the geometric relationship between objects and the eye. Large objects
involve large angles at the eye, and smaller objects involve smaller ones.
one side of the eye to the other side will be large. When the object is
small, that angle will be more acute.
But how does the eye determine where rays of light are coming
from? Light scatters, so the light at any given point is a composite of
light arriving from many other locations. Consider, for example, a
Renoir painting of a group of people having lunch (Plate 2). Reflected
light spreads in all directions from the surface of each objectthe
glasses, the plates, the wine bottles, or the people. Now think about
what this means from the perspective of some particular vantage point.
If you were to measure the light arriving at some specific location in the
scenelets say the shoulder of the woman on the leftit would turn
out that light from all the different positions in the scene arrives there
at once. Some of the light is reflected from the top of the wine bottle,
some from the water glass, some from the fruit bowl, and some from
all the other locations with an unobstructed straight-line path to that
location. Its a blur. If you look at her shoulder, you see no indication
that the light there came from any of these sources. In short, although
T H E WAY S O F L I G H T
23
the light arrives on a straight-line path, there are many such paths, and
they are pooled all together. No image is formed.
What is needed for spatial vision is a way to organize the rays of
light that originate from different places. If the ray of light reflecting
from the wine bottle and the ray of light from the fruit bowl could be
made to maintain their spatial relationship to each other when they
reach a set of light sensors, then it would be possible to see these two
light sources as separate and distinct. There must be a one-to-one correspondence between location of origin and location on the set of light
sensors.
One way to sort the rays of light is to filter them through a very small
opening. The image-forming capacities of pinhole-size openings have
been known since around the fourth or fifth century BCE. A device
known as the camera obscura was apparently invented independently
in ancient Greece and China and was described by both Aristotle and
the Chinese philosopher Mozi. A camera obscura is a dark box or room
with a tiny aperture through which light is admitted. The light passes
through the opening and projects onto the far wall, forming a clear
image of the external scene (Figure 2.6).
Because of the small size of the opening, only one ray of light from
each location can get through to the rear wall. As always, in the scene
outside the chamber, the rays of light are reflecting in all directions from
all positions in the scene, but the wall blocks most rays of light from
entering the chamber. Follow the path of light from, say, both the top
and bottom of the tower in the scene. At the outside surface of the wall,
rays of light from both locations are mixed together. But a ray of light
from the top of the tower can pass through the opening on a downward
path and hit the lower part of the inner wall behind the opening. A ray
24
M A K I N G SPAC E
FIGURE 2.6 A camera obscura. When light travels through a small opening, the rays
of light reaching any particular location on the opposite wall originate from only one
location in the external scene. Light from other locations is blocked by the outer wall.
The resulting one-to-one correspondence between locations on the inner wall and
locations in the external scene means that an image of the external scene is formed.
of light from the bottom of the tower can pass through the opening on
an upward path and hit the upper part of the wall behind the opening.
On that rear wall, the light that hits it at any particular spot comes from
only one direction in the external scene. This forms an image. (Well
come back to the fact that the image is upside down and backward.)
In part, the eye uses the same pinhole technique. What light enters the
eye is regulated by the pupil, a small, transparent opening in the otherwise opaque front of the eye. Light passes through this opening to reach
the retina, located on the inner surface of the back of the eye. In some
simple creatures, thats enoughthe narrow pathway prevents light
from entering and reaching a given spot on the back of the retina unless
T H E WAY S O F L I G H T
25
26
M A K I N G SPAC E
this affects not only the whole theory of eclipses, but also most of all ...
the measurement of the celestial bodies.4 Any good scientist needs to
know how his or her instruments work and to what extent they can be
trusted. For astronomers of that era, their instruments were their own
eyes, and Kepler argued that it was important to know whether seeing
should be believing.
The particular issue Kepler sought to explain concerned a mysterious
discrepancy in the estimates of the size of the moon obtained during an
ordinary full moon versus during a solar eclipse. Kepler, Brahe, and
other astronomers used a camera obscura to make such measurements.
The image of the moon was projected onto the back wall, where its
diameter could be measured and converted to units of angle based on
the distance between the aperture and the rear surface (Figures 2.7 and
2.8). The puzzle that emerged with this method was that the diameter
of the moon when recorded during a solar eclipse appeared smaller than
the diameter observed at other times.
Blur
Sun
Measured
diameter of
sun with large
aperture
Sun
Measured
diameter of
sun with small
aperture
FIGURE 2.7 Images are slightly blurred when they pass through an aperture. The amount of the blur
depends on the size of the aperture. In the drawing on the left, light from the lower edge of the sun
can pass through the aperture on a set of paths ranging from those that graze the lower edge of the
aperture to those grazing the top edge. The spatial extent of this set of alternate paths creates blur at
the edge of the projected image. Since all of this light is very bright (in the case of the sun), the edges
of the projected image of the sun will not necessarily look blurry to our eyes. Instead, the apparent
diameter of the sun will include the blurry portion. With a smaller aperture (right), the blurry portion will
be smaller, and the apparent diameter of the sun will be correspondingly smaller as well.
T H E WAY S O F L I G H T
27
Moon
Measured
diameter of
moon by itself
Moon
Measured
diameter of
moon during
eclipse
FIGURE 2.8 When you measure the diameter of the moon using a pinhole camera, it will be slightly too
large by an amount determined by the diameter of the pinhole. During an eclipse, however, the moon
reflects no light of its own; it is entirely in shadow. It will appear to be smaller than when seen by itself,
because the blur from the rays of light from the sun nibble in around the edges of the shadow cast by
the moon.
The problem stems from the different effects of blur for bright versus
dim objects. When you are looking at the image of a very bright but
otherwise homogeneous object projected onto the rear screen of a
pinhole camera obscura, the effect of the aperture diameter is to make
the image appear too large; you judge the edge of the object to be the
outer edge of the blur (Figure 2.7).5 Brahe was aware of this and applied
a correction for aperture diameter to his measurements of the size of
the sun and the moon. But a puzzling thing happened when he viewed
and measured the images of the sun and the moon during a solar eclipse
(Figure 2.8). Remember that a solar eclipse involves the moon coming
between the earth and the sun. You can see the outer ring of the sun
behind the moon, which appears like a dark disk in front of the sun.
When viewed through the pinhole camera, the sun appeared too large,
as usual, but the moon appeared too small.
We now know that blur occurs for both the suns image and the
moons image, but since the moon is dark in this circumstance, the suns
blur essentially nibbles in from the actual circumference of the moon,
making it seem smaller. But this was by no means clear to astronomers
28
M A K I N G SPAC E
of that era. Instead, they were inclined to apply the same subtractive
correction to the moon that they did to the sun, thus estimating the
moons size as smaller than it actually is.
Kepler solved the mystery of the blurry pinholes using simple tools
at his disposal: a spool of thread and a book. He set the book on a high
place, representing the sun or moon, and positioned a barrier with a
small aperture between it and the floor. He then attached a piece of
thread, modeling rays of light, to each corner of the book, pulled them
through the aperture, and marked with chalk the place where each
piece of thread reached the floor when pulled taut. The chalk marks
traced out the outline of the book (but inverted and reversed, as we
have already noted). This provided a convincing model of the principle
of how light passes through apertures and forms images. His accurate
geometrical drawings allowed him to sort out the details of how aperture size and distance from the screen affected image size and blur.
Kepler and other scientists of the day realized that the pupil of the
human eye is large enough to produce a considerable amount of blur.
A pupil diameter of 2.5 mm (millimeters) produces about six degrees
of blur on the surface of the retina, which is about 25 mm behind the
pupil. That corresponds to about 2 inches of uncertainty about the location of a visual stimulus located 20 inches in front of the eye (or about 5
cm of uncertainty for a stimulus 50 cm away). With this amount of blur,
we would be unable to distinguish one column of print from another in
the newspaper. Yet we seem to see much more clearly than this. In fact,
we are able to distinguish differences in the position of visual stimuli
that are about three hundred times smaller than that. So, Kepler realized that there must be more to image formation in vision than just the
aperture provided by the pupil.
T H E WAY S O F L I G H T
29
passes between substances of different densities, like air and water. How
sharply depends on the lights trajectory, ranging from not at all for light that
hits the surface perpendicularly to quite
steeply for light that arrives obliquely.
Early scholars of vision postulated that
we see only the unrefracted light that
arrives head on to the air-fluid boundary at the entrance of the eye.
Light
Less dense medium (e.g., air)
More dense medium (e.g., water)
Refracted light
The other piece of the puzzle has to do with the bending of light,
or refraction (Figure 2.9). Light only travels in straight lines when it is
traveling through a medium of constant density. It bends as it passes
from air into substances that are more (or less) dense. (Thats why your
legs may appear crooked when you look down at them in waist-deep
water. Your legs are the same as they ever were, but the light is not traveling on a straight line from them to your eye. The rays of light bend
when they leave the water and enter the air. Your brain cant fathom
that light is not traveling straight, and so assumes that it must be your
legs that are crooked.)
The eye contains several substances that differ in density from air
the cornea, the aqueous and vitreous humors, and the lenswhich
cause light to bend (Figure 2.10). The space between the cornea and
the crystalline lens is fluid filled (the aqueous humor). Behind the lens,
inflating the globe of the eye like a basketball, is the jelly-like vitreous
humor. Light is refracted or bent by each of these substances.
Scholars wondered if the way light bends as it entered the eye might
somehow solve the pupil-blur problem, but before Kepler, they were on
the wrong track. Most proposed that somehow only unrefracted light
30
M A K I N G SPAC E
Ciliary muscle
Anterior chamber
(aqueous humor)
Fovea
Lens
Cornea
Pupil
Vitreous
humor
Iris
Choroid
Sclera
Retina
Optic disc
Optic nerve
FIGURE 2.10 The anatomy of the human eye. The pupil is the opening in the iris, the
part that gives your eye its color. The pupil acts like a pinhole. The cornea, lens, and
aqueous and vitreous humors are all clear, but more dense than air, causing refraction. The cornea and humors actually contribute more of the refraction than the lens
and can therefore be thought of as being part of a cumulative lens complex. The lens
proper is commonly viewed as the key player primarily because it provides the only
adjustable component of refraction. The shape of the lens is controlled by tiny ciliary
muscles inside the eye. These muscles can either flatten the lens or plump it up, altering its focal length to bring an image into focus on the retinas photoreceptors at the
back of the eye. Directly opposite the lens and pupil is an area of the retina where the
photoreceptors are particularly densely packed, known as the fovea. Signals are sent
from the retina to the brain via the optic nerve. (Well return to the fovea and the optic
nerve in Chapter 4.) The whole thing is encased in a tough outer coat (the sclera) with
a blood-vessel-filled lining (the choroid).
was seen by the eye, because this light is a little brighter than light that
has been bent. But Kepler realized this effect was small, and he sought
an alternative explanation.
Kepler argued that light arriving at the curved surface of the eye
would be bent in such a way that light originating from the same point
T H E WAY S O F L I G H T
31
FIGURE 2.11 When light passes through a lens, it bends according to its angle of approach. When it hits perpendicularly, it
goes straight through (bottom dashed line or top solid line). But
when it arrives at a steeper angle, it is bent more. At just the right
distance behind the lens, known as the focal distance, all the lines
from a particular position in the world converge on a common
position. In the drawing above, the dashed lines indicating the
paths of light originating from the bottom of the tree come back
together in one spot, as do the solid lines from the top of the tree.
In this fashion, an upside-down and backward image is formed.
FIGURE 2.12Ren Descartes, writing a few decades after the
32
M A K I N G SPAC E
T H E WAY S O F L I G H T
33
34
M A K I N G SPAC E
T H E WAY S O F L I G H T
35
Convergence angle of
the eyes for a far target
Convergence angle
for a near target
FIGURE 2.13When you look at something far away, the angle between your
two eyes is only slightly converged.
When you look at something close, the
eyes converge more sharply. Our ability
to adjust the relative angle of our eyes
allows us to aim both eyes at the same
visual stimulus. We use both our knowledge of our eyes convergence angle
and the relative position of a given
objects image in each eye to estimate
its distance.
nearer or farther from the convergence distance of the two eyes will
project images that do not match up in the two eyes.
Consider the scene in Figure 2.14. Suppose your eyes are looking at
the coffee cup and are turned inward slightly so that the cup is centered
for each eye. The image of the cup lands at the middle of the retina. The
pear is at the same distance as the coffee cup and a little off to one side.
Its image lands off to one side of the retina. Because it is at the same
distance as the coffee cup, the image is displaced by about the same
amount in each eye.
Now consider the vase of flowers. The vase is behind the coffee cup
and on the line of sight for the left eye, so its location on the retina
is approximately the same as for the cup, which is actually slightly
blocking its view. But in the right eye, the vases image lies adjacent to
that of the cup. We can use the cups image as a marker for two sites
on the retina that correspond or are equivalent to each other in the
two eyes. The positioning of the image of the cup and the image of the
vase in the right eye depends on how far away the vase is, relative to
the distance of fixation. (If the eyes shifted to both look at the vase, the
image of the vase would then be centered in both eyes.) This ability to
36
M A K I N G SPAC E
FIGURE 2.14 Each eye has a different vantage point on the scene. Differences in the
relative positions of images in each eye can provide information about distance. If
the left- and right-eye versions of two objects occupy the same relative positions,
the brain can infer that the two objects are about the same distance away. In this
illustration, the coffee cup and pear are the same distance away, and their images
are separated by the same amount of space on the retina of each eye. If the left- and
right-eye versions are different, that indicates that one object is farther away than the
other. Here the image of the vase occupies a different position relative to the coffee
cup and pear in the left eye versus the right, due to its greater distance.
T H E WAY S O F L I G H T
37
38
M A K I N G SPAC E
T H E WAY S O F L I G H T
39
and be visible primarily by one eye or the other, some method of showing each eye a slightly different view is employed.
One such method is to show the two superimposed but offset images
in slightly different colors. When viewed through glasses that also have
panes of different colors in front of each eye, the image seen by each
eye is slightly different, just as it would be when viewing a real three-dimensional scene. To see how this works, look at Plate 3. There are two
main objects in the scene, the capitol dome and a statue in front of it.
The capitol is farther away than the statue. The capitol is shown in
the usual manner, with just one clear image (in black and white). The
statue consists of two overlaid versions, one bluish and offset to the
right and the other reddish and offset to the left. If you view this image
through glasses with one red pane and one blue pane, the eye viewing
through the red pane will see primarily the blue image, and the eye
viewing through the blue pane will see primarily the red image. Both
of those images will appear to be a similar, dark color, because blue
when viewed through a red lens looks dark and vice versa. Also, both
eyes will see the capital dome, because it is black and white, and they
will see it at the same place, because it is a single image. However, the
relative locations of the capitol dome and the statue will be different
for each eye, and that will make the statue appear to be at a different
distance than the capitol dome. If you put the glasses on correctly, with
the red lens in front of the left eye and blue in front of the right eye, the
statue will appear nearer than the capitol. Reverse them, and the statue
should, in principle, appear to be farther away (but it probably wont,
for reasons well come back to below).
Two other methods of showing the two eyes different images involve
glasses with shutters and polarization. With the shutter method, the
40
M A K I N G SPAC E
T H E WAY S O F L I G H T
41
Vertically polarized
output
Vertical filter
Horizontally polarized
output
Horizontal filter
42
M A K I N G SPAC E
T H E WAY S O F L I G H T
43
dome that are not blocked from view. When your brain is confronted
with a conflict between the in-front-of versus behind clues from the
scene and the stereovision cues from each eye, it will usually opt in
favor of the former.
The surrealist painter Ren Magritte illustrates how powerful occlusion can be (Plate 4). This painting is startling because Magritte pits
occlusion against other clues to distance, making us see depth differently in different parts of the scene. If you look at approximately the
upper left quarter of the painting, to the left and above the rider, you
get a clear impression from occlusion of which trees are in front of the
horse and rider and which are behind. But what is in front and what is
behind in that part of the scene doesnt carry over to other parts. Look
at the small tree that obscures our view of the riders right shoulder.
When you follow the tree trunk down, you will see that where it meets
the ground it is partially blocked from view by a larger tree. And if you
follow that larger tree back up, youll see that it in turn seems to be
behind the horse and rider.
Magrittes odd tricks also rely on contrasting occlusion with haze,
another clue to three-dimensional space. For distant objects, the intervening air produces blur and a slight bluing of the colors. More distant
mountain ranges appear more fuzzy and bluer than closer ones. Here,
Magritte blurs and blues some of the foliage to make it appear to be in
the background. He then places a strip of that supposed background
foliage in front of the horse and rider, confusing us again about how far
away they are.
Another clue to distance is relative size. When an object is farther
away, it will cast a smaller image on the retina than when the same
object is closer. How big something appears gives a sense of how far
44
M A K I N G SPAC E
away it is. We tend not to be aware that we use size as a proxy for
distance. In the photograph in Figure 2.17, your brain readily concludes
that the children are in the foreground and the tractor is farther away,
without noticing that that impression is based on how much of the
expanse of your retina is occupied by the images of the children in
comparison with the tractor.
Your expectations of how big the object should be affects your interpretation of distance. Once, I was walking in the hills in California
and came across a tarantula in the trail. I was doubly startled, not just
because it was a tarantula, but also because I initially thought it was
much closer than it actually was. A spider of that retinal image size
must be very near, according to my brain. It took an instant longer
to realize that it wasnt that close, it was just very big compared to
other spiders I was accustomed to seeing. (By that point I had jumped
back about five feet to reconsider whether I really needed to take that
particular path.)
T H E WAY S O F L I G H T
45
46
M A K I N G SPAC E
FIGURE 2.19M.C. Eschers famous drawing Waterfall uses linear perspective and
T H E WAY S O F L I G H T
47
48
M A K I N G SPAC E
T H E WAY S O F L I G H T
49
something very difficult with a small margin for error like walking a
tightrope, we tend not to think about where our body parts are.
As we saw in the case of vision, this requires sensors capable of
measuring what we want to know, in this case, the state of the muscles
and joints of your body. These sensors signal the brain regarding how
your muscles are contracting and what forces are at play in your limbs
and joints. But the measurements taken by these sensors have only an
indirect and ambiguous relationship to the position of the body itself.
Your brain has to infer the configuration of your skeleton in space from
the signals it receives from these sensors.
Two kinds of sensors measure and report your bodys position to the
brain. One kind is located inside your muscles and looks a little bit like
a spool or spindle of tissue with nerve fibers coiled around it. This type
of neuron is called the muscle spindle receptor, and it measures how long
the muscle is. Muscle spindle receptors run alongside individual fibers
of the muscle, stretching as they do (Figure 3.1). Messages from muscle
spindle receptors provide the brain with insights into a joints position
from how much the attached muscles are stretched.
For example, the muscles in your upper arm (your biceps and triceps)
are responsible for moving your forearm and changing the angle of your
elbow. When the biceps muscle contracts, it shortens, bending your
elbow and pulling your forearm closer to your upper arm. As the angle
between your forearm and your upper arm becomes smaller, tension
on the muscle spindle receptors in the biceps eases, causing them to
slacken. On the back of your arm the triceps muscle lengthens as you
bend your elbow, and its muscle spindle receptors become taut.
If this were all there is to it, you would be able to determine the angle
of your elbow simply by comparing the degree of stretch in the muscle
52
M A K I N G SPAC E
Golgi
tendon
organ
Muscle
spindle
receptor
FIGURE 3.1 The brain deduces the position of joints by interpreting signals
from sensory receptors sensitive to
stretch in the muscles and tendons.
Muscle spindle receptors are found
inside the muscle, strung in parallel
with the main fibers that provide the
contraction strength of the muscle.
They detect how long the muscle is.
Golgi tendon organs are located at the
junction between the muscle fibers and
tendons. They detect how much force
the muscle is applying to the tendon.
spindles on each side of your upper arm. If biceps muscle spindles are
long, your elbow is straight and your arm fully extended, and if triceps
muscle spindles are long, your arm is fully flexed and your elbow is bent
all the way. But thats not the whole story.
When I dropped the glass of iced tea, the problem wasnt solely one
of hand position but also of pressure. My hand was properly shaped to
grip the glass, but I did not apply enough force to the glass to keep it
from slipping from my fingers. Force is measured by the other kind of
body sensor, Golgi tendon organs (which, contrary to the name, are
neurons and not organs.)
Golgi tendon organs are connected to the muscle on one side and to
the tendon and bone on the other side (Figure 3.1). Like muscle spindle
receptors, these neurons are sensitive to stretch, but they measure it
in a different place. They are connected in series with, or in line with,
the muscle, rather than alongside it like the muscle spindles. When the
muscle contracts, the Golgi tendon organ stretches.
Both muscle spindle receptors and Golgi tendon organs are necessary because the way muscles contract (and limbs move) depends on
53
Tendon
Muscle
Muscle
spindle
receptor
Golgi
tendon
organ
how much weight theyre carrying. When your hand is empty, a certain
amount of force in your biceps muscle will hold your elbow at a ninetydegree angle (see, for example, Figure 3.2a). Either the length of the
muscle (detected by the muscle spindle receptor) or the force applied by
the muscle to the tendon (detected by the Golgi tendon organ) could be
used to infer joint angle.
But when you are holding something heavy, more tension is
required to hold the arm in the same position. To generate this force,
the muscle contracts harder, shortening more and pulling on the
tendon more in compensation (Figure 3.2b). For the same joint angle
with a heavy load, the muscle spindle will signal that the muscle is
54
M A K I N G SPAC E
shorter, and the Golgi tendon organ will signal that there is more
force in comparison to the same joint position when you are not
carrying anything. The reports from the two kinds of sensory receptors are somehow merged in the brain to infer whether the elbow is
sharply bent or you are merely holding something heavy. A similar
process occurs in your fingers, hands, legs, feet, toes, neckall the
joints of your body.
Stretching affects electricity in these receptors. As we saw for vision,
the story involves ions, little holes in the membrane through which
those ions can pass, and a chain reaction. But how it works is a little
different. Instead of altering the electricity in a neuron indirectly
through a series of biochemical reactions, stretching a muscle spindle
or Golgi tendon organ receptor probably literally pulls open some of
the pores in the neurons membrane. We dont know for sure, but it is
easy to imagine that tugging on the membrane could create or enlarge
openings in it, just the way pulling on a sponge will expand its holes
(Figure 3.3).1 When the gates open wide, ions can flow, leading to electrical changes in the neuron. The mechanical state of the membrane
thus controls its electrical state.
Electrical signaling in neurons has some quirks that affect body position sensing. We saw already that the brain has to combine information
FIGURE 3.3Stretching a neuron is
Stretch
outside (+)
+
ion channel
membrane
inside ()
55
from muscle spindles and Golgi tendon organs because neither on its
own provides unambiguous information about body position. But it must
also combine information from both sides of the limb (from both sets of
muscles and tendons) to infer the position of your elbow. One reason is
that your brain doesnt really do negative numbers, making it tricky to
code the full range of possible limb positions using only the sensors on
one side of the limb.
To understand why brings us back to the particular ions involved
in electrical signaling in these neurons and a kind of signal called the
action potential. Action potentials, also called spikes, form the currency
of electrical information in many neurons in the brain and nervous
system. An action potential consists of a very rapid pulse, lasting about
1 to 2 ms (milliseconds) (Figure 3.4). The voltage shoots up from a negative value (more electrons inside the neuron) to a positive one (more
electrons outside), and then comes back down again, dipping slightly
below the original electrical potential before coming to rest at the original, prespike voltage.
Action potentials are caused by the opening and then closing of
several different ion channels in the membrane of the neuron. First, an
ion channel that allows positively charged ions to flow into the neuron
opens, then that one closes and another that allows positively charged
ions to flow out opens. The voltage across the membrane swings from
its normal negative value briefly to a positive value and then back again,
as positive ions flow first into, and then out of, the neuron.
The reason that opening of ion channels can cause ions to flow first
into, and then out of, the neuron is that these two sets of channels
permit different kinds of ions to flow. The initial inward flow involves
sodium ions and the later outward flow involves potassium ions. Which
56
M A K I N G SPAC E
(+) out
(+) in
Sodium
m
Potassiu
+50
70
0
Time (ms)
direction these ions flow depends on two factors, their relative concentrations inside versus outside the neuron and the electrical potential
across the membrane.
When a particular ion is more densely concentrated in one area than
another, it will tend to leak from the area where there is more of it to
the area where there is less of it, if given the chance. When you pour
milk into your coffee, the milk and coffee initially occupy separate
spaces in your cup. The milk particles are more densely concentrated in
the dollop of milk, and the coffee particles are packed in around it. But
as you stir, the milk and coffee particles each diffuse outward until they
are evenly mixed. Sodium is more densely packed outside the neuron,
and potassium is more densely packed inside it. The higher density of
sodium outside a neuron tends to make sodium flow or diffuse inward
when sodium ion channels open. And the opposite is true of potassium,
57
58
M A K I N G SPAC E
feedback cyclemore sodium flows in, making the voltage inside still
more positive, opening more channels, and allowing yet more sodium
to flow in.
This positive feedback cycle is soon interrupted by the triggering
of the potassium channels. As the membrane potential gets still more
positive, the second gate, the potassium channel, also changes shape
and opens up. Once open, the potassium can flow freely, and it floods
out of the neuron. This counterbalances the sodium flow and restores
the membrane potential back to its original level. (In fact, the sodium
flow actually stops, because the sodium channels get wedged or stuck
in a closed position when the voltage goes very positive, and they cant
reopen until a return to a negative voltage level reboots them.)
These self-generating, self-restoring action potentials are triggered
when something nudges the neuron into a little more electrically positive territory. In the case of the stretch-sensitive Golgi tendon organ
and muscle spindle receptors, the initial trigger for a spike is the bump
in voltage caused by the muscles or tendons pulling on the receptor
neuron and opening some of its (probably sodium) ion channels. The
small electrical change induced by such mechanical forces is then
amplified by the chain reaction of voltage-sensitive sodium and potassium channels, which creates spikes.
Because spikes are caused by an automatic sequence of events, every
spike is about the same as every other spike (Figure 3.5). They have
about the same size, shape, and time course.2 So the spike, by itself, cant
tell you much about stretch. Strong stretch does not make the spike any
bigger. But stronger stretch causes more spikes. Weaker stretch causes
fewer spikes to occur. The amount of stretch is reflected in the number
of spikes that occur, not in their size. So, your brain monitors whether
59
FIGURE 3.5 Like Tolstoys happy families, all action potentials are (more or less) alike.
The shape, size, duration, and height of the peaks are all about the same. What varies
is how many there are in a period of time. So, the train of spikes in (a) might occur
when a muscle spindle or Golgi tendon organ neuron is pulled hard, and the train of
spikes in (b) might occur when it is pulled more weakly.
spikes are occurring and how rapidly they happen to determine how
much that sensor neuron is stretched.
This is where negative numbers come in. Neural firing patterns cant
consist of fewer than zero spikes. They can range from zero up to about
five hundred spikes per secondthats a 2-ms interval between spikes.
Closer together and they would run together; there wouldnt be enough
time for the ion channels to open and close again before the next spike.
But theres no such thing as a negative firing rate.
The brain could deal with this problem in a couple of ways. One
would be to tailor the available dynamic range of neural firing to the
range of possible stretches. A small number of spikes would indicate no
stretch, and a large number of spikes would indicate a strong stretch.
60
M A K I N G SPAC E
And, indeed, there are small muscle fibers that set the tension applied
to the muscle spindle receptors, like the tuning pegs that tighten the
strings on a guitar.
But the other way is to organize the system like a game of tug-ofwar. With some receptors stretched when the arm moves one way and
others stretched when the arm moves the other way, the brain is able
to compare the signals from a variety of receptors, some of which will
have high levels of activity and others low levels at any given moment
and for any given position. Whenever your biceps is contracted, your
triceps is stretched and vice versa. So there is always someone minding
the store, generating a decent, above-zero firing rate. Comparing which
side is firing more vigorously provides information about body position. This comparison serves the purpose that would otherwise require
a robust ability to traffic in numbers that could be either positive or
negative.
Still, these receptors can be fooledsubject to illusions in which
your brain is duped into believing something that is not actually true.
As we saw in Chapter 2 on vision, the brain can be deceived into seeing
depth where none exists if different views of the scene are presented to
each eye. Bamboozling a muscle spindle receptor involves a different
trick. For example, if you use a physical therapy vibrator, you can make
a muscles spindle receptors behave as if that muscle is being stretched.
The shaking induced by the vibration likely causes the spindle receptors to fire action potentials. Your brain interprets this signal as an indication that the limb must be moving when it is not.
Applied to your biceps muscle (on top of your upper arm), vibration
may create the sense that your elbow is straightening, as though your
triceps (on the opposite side) had contracted. This can have further
61
FIGURE 3.6 Vibrating the biceps muscle may make you feel like your elbow is unbending. If you touch
some other part of your body, such as your nose or the top of your head, it may feel like your nose or
the top of your head is bulging outward to keep in contact with your hand.
62
M A K I N G SPAC E
to one side. This can be done using eyeglasses or goggles with displacing
prisms in lieu of conventional lenses. Such prism lenses bend light before
it reaches your eyes, changing where things seem to be located in the
visual scene relative to your body. When you wear prism goggles that
shift light to the left by, say, about ten degrees, the visual scene wont
appear that different to you. The images are the same, just shifted over
to the side. But the relationship between where things appear visually
and how you have to move to touch or avoid them is now changed. The
mismatch will cause you to stumble around and bump into things.
As Stratton did, your body will soon adjust to this altered vision,
but it will do so limb by limb. If you try to play a game of darts, your
initial throws will miss, landing about ten degrees to the left of where
the target really is. Keep at it, though, and over the next dozen or so
throws, your darts will be more or less centered (at least as well as
they are normally). Once your throws are accurate again, switch arms
and throw again. Assuming youre sufficiently ambidextrous enough
to even attempt this, your throws should once again land about ten
degrees to the left. Youll have to teach that arm too. Your brain appears
to be careful not to generalize until the need is demonstrated.
These examples illustrate that sensing visual location and sensing
body position work in partnership with each other. This partnership
begins in infancy, when you first learned the relationship between your
eyes world and your bodys world. As a baby, you werent born knowing the relationship between your body and the things that you could
see or touch. Babies have to learn this body part by body part. Theres a
lot of flailing and spilled breakfast cereal involved in that process. And a
lot of practice, over a surprisingly long period of timemonths to learn
how to move the tongue to manipulate solid food, about a year to learn
63
to walk, several years to run, jump, throw and catch a ball, or pour a
glass of iced tea.
The final piece of your bodys sense of space concerns feeling things with
your skin.* This involves a variety of different sensor types specializing
for different kinds of touch stimuli, from vibration to strong pressure
to light stroking. Objects dont even have to touch your skin directly.
Take a pen and move its barrel along your forearm, just close enough
to brush the fine hairsyou should be able to feel it quite clearly. The
very hair on our bodies is part of our sense of touch. In this, we are a
little like cats, rodents, and a variety of other animals who are capable of exploring objects with their whiskers. We do the same with our
hair follicle receptors, although our hairs are softer, and we dont have
the ability to wiggle them around to probe our environment the way
bewhiskered animals do.
As you should now expect from our discussion of the eye and the
visual image, the skin and hair follicle receptors only respond if touch,
pressure, or vibration reaches them. These mechanical forces have
to cause the receptor to change its electrical signaling. The limited
distance over which touching can impact a receptors membrane
means that these neurons can provide information to the brain about
the location of a stimulus. They have receptive fields on the skin. If something touches your skin in the receptive field of a particular receptor,
the receptor responds; if not, it doesnt. Your brain monitors which
* Neuroscientists group the senses of touch, body position, and other types of information gleaned from receptors in the body under the heading of the somatosensory
systemfrom soma, the Greek word for body.
64
M A K I N G SPAC E
receptors are responding and learns from this where the stimulus is
located on the body surface. Touch signals are then combined with
information about body position to estimate where the touched object
is in space.
I once did an experiment that tested how this works.4 Subjects
(my graduate advisor and I!) sat at a desk touching two small posts
mounted on it, one for each hand. Each post could vibrate, and when
it did, we were supposed to look at itthat is, move our eyes from
looking at a visual target elsewhere to looking at the hand that was
touching the vibrating post (Figure 3.7). We did this in the dark, so
we couldnt see our hands; the eye movement had to be made based
on touch. We had no trouble when we just reached straight out with
our arms to touch the postsright hand on the right post, and left
on the left. But strange things happened when we crossed our hands
so that the left hand grasped the right post and vice versa. In this
FIGURE 3.7 When you feel something in your hand (such as a vibrating post), you
might make an eye movement to look at what you feel. The path of that eye movement is usually quite straight: the dots illustrating where the eyes are looking at any
given moment are spread along a line from the starting point to the ending point.
However, if your body is in some unusual position, such as having your hands crossed,
your brain is initially confused about where to look, and the eyes set off in the wrong
direction and then correct.
65
66
M A K I N G SPAC E
Although in this instance our sense of body position took some time
to catch up with reality, normally this sense is most accurate just after
we move, and it degrades the longer we stay still. Try thinking about
where your body is next time you wake up from a nap and before you
move. You may find that you cant quite tell at what angle your knee is
bent, or how high above your head your arm is crooked. But as soon as
you move, that uncertainty should go away.
The reason for this uncertainty is that the receptors that sense body
position respond most vigorously to a change in position, and their
activity levels tend to trail off with the passage of time. This is true not
just for your position sense but also for your sense of touch. Consider
your socks. Are you wearing any? How do you know? Because Ill bet
you cant feel them. Your clothes tend to disappear from your touch
sense, but any new stimulus, like a mosquito biting, will be responded
to quickly and vigorously.
Now that you know something about how several different kinds of
sensory informationvision, touch, and body positionare detected
by sensory organs and what kind of spatial information these organs
can provide, we are ready to move into the brain itself and find out more
about how this information is represented. How do neural signals
serve as a code for this sensory information? You may have noticed that
we havent yet talked about hearing and how we know where sounds
are located. That will come later.
67
70
M A K I N G SPAC E
Dendrites
Synapses
Synapses
Axon
Cell body
Information flow
71
72
M A K I N G SPAC E
out the contours of objects. And they begin doing this right in the
retina.
The retina contains a sequence of different kinds of neurons, arranged
in several layers beginning with the photoreceptors. Once light has
been converted to electrical signals by photoreceptors, these signals can
pass from neuron to neuron (from axon to synapse to dendrite to cell
body to axon) until they reach the last group of neurons in the retinal
chain. This path is where the process of demarcating the boundaries
of objects begins. The first step is to make a neuron that responds
better to an edge of some kind than to a uniform stimulus. Edges are
what distinguish one object from the next. So, emphasizing edges can
provide a clue to where one thing ends and another begins.
But what does it mean to emphasize an edge, and how does a neuron
equip itself to do this? Photoreceptors are not sensitive to edges per se;
their responses are dictated by the amount of light at a given location,
the amount of light in the receptive field. But the next few neurons
along the chain get input from photoreceptors with a combination of
different kinds of synapsesa mix of synapses that act in either positive
or negative ways. This creates neurons able to probe the visual scene
for areas of contrast, an essential component of an edge.
Like the tug-of-war between sensors on opposite sides of a limb,
synapses provide a means to incorporate both positive and negative
signaling. At a positive or excitatory synapse, the voltage in the recipient
neuron shifts upward (that is, it becomes less negative) in response to an
incoming signal. Excitatory synapses make the receiving neuron more
likely to fire an action potential. At a negative or inhibitory synapse, the
voltage may not actually change very much because there is a limit to
73
how negative the voltage can get, but the neuron becomes less capable
of responding to other excitatory inputs.* This reduces the number of
action potentials it generates.
The action of neurotransmitter receptor molecules determines
whether a synapse is excitatory or inhibitory. A given type of neurotransmitter may interact with many different receptor molecules, each of
which may have a different effect on the recipient neuron, depending
on what kind of ion channel it controls and whether it opens that channel or closes it. Complicated!
The pattern of positive and negative synapses that occur between
photoreceptors and later retinal neurons creates sensitivity to differences
in the level of light at different locations. Heres how it works.
First, lets return to a point we touched on briefly in Chapter 2: light
causes the electrical signal of the photoreceptor itself to become more
negative. So we can think of the raw effect of light as being an inhibitory effect. But when the photoreceptor transmits its signal to the
recipient neurons, it can do so using either an excitatory or inhibitory
synapse. The result is similar to multiplying by a positive versus by a
negative numberretaining the direction of the original signal via the
excitatory synapses, or inverting it via the inhibitory ones.
Consider a hypothetical photoreceptor, forming an excitatory
synapse with one neuron (A) and an inhibitory synapse with another
* This is due to the full mixture of ions, not just potassium and sodium but also
the negatively charged chloride ions. As discussed earlier, how they tend to diffuse
depends on the combination of their concentration gradients and their electrical
charges. The net result is that even when inhibited, the membrane potential does not
get much more negative than about 70 mV.
74
M A K I N G SPAC E
75
Best stimulus
(Figure 4.2). If the light spot has the right size and position to line up
with the receptive field of the photoreceptor in the center, and the darker
background covers the receptive fields of the photoreceptors in the outer
ring, then each of these two sources of inputs will provide the most positive (or least negative) effect that they can on the recipient neuron. The
surround photoreceptors generate a positive signal in response to the
dark stimulus, and this excitation is conveyed to the recipient neuron
via an excitatory synapse. The central area of photoreceptors is inhibited
by the light spot, which spares the recipient neuron from the otherwise
76
M A K I N G SPAC E
inhibitory signals that would be generated if this center were being stimulated by a dark stimulus. A polka-dot sensor!
Other recipient neurons have the opposite pattern of inputsexcitatory inputs in the center and inhibitory inputs in the surrounding
circleso that they respond best to dark polka dots on light backgrounds. The polka dots can also be different sizes; sensitivity to larger
polka dots occurs in neurons that receive input from larger clusters
of photoreceptors. Such assortments of connection patterns provide a
range of response patterns, but all are devoted to emphasizing regions
of the visual image where patterns of light vary.
By accentuating local variation in the visual image, polka-dot sensing is a baby step toward identifying the boundaries of whole objects.
However, even after pooling signals across varying numbers and
arrangements of photoreceptors, the retinas recipient neurons are
sensitive only to very small, mostly round regions of space. So the
brain must build on this initial pattern of contrasting connections. At
each successive stage of synapses, signals are combined from a broader
region of the visual world, highlighting contrasting parts of the visual
scene and in more interesting ways than just round patterns.
A central challenge the brain faces when building such complex
connection patterns is keeping all those patterns organized. This is
particularly true when the connections span long physical distances,
such as between the retina and the brain itself. The last layer of the
retinathe last to receive messages passed along from the photoreceptorsconsists of a group of neurons known as retinal ganglion cells.
Retinal ganglion cells have dendrites and cell bodies in the retina, but
their long and stringy axons travel from the back of the eye to the
77
78
M A K I N G SPAC E
that their receptive fields are near but not at exactly the same location
as the first set of neurons. Move farther, and the receptive fields will
shift farther as well. The same is true of the primary visual cortex
(Plate 5), the next stage of visual processing after the thalamus, which
also receives an orderly pattern of connections from the thalamus. The
pattern of light in the visual scene mapped onto the retina by the pupil
and lens is thus again mapped onto the thalamus and the visual cortex
by the pattern of neural connections that route the information to that
destination.
The orderly pattern of connections from the retina to the thalamus and the visual cortex is also useful for creating detectors for more
complicated features of the visual scene. In the visual cortex, there are
neurons whose receptive fields are stubby rather than round.2 They
respond best not to polka dots but to stripes. Some neurons have receptive fields matched to horizontal bars, others are matched to vertical
bars, and still others are best suited to diagonal bars, either this way \
or this way /. The bars can be fat or skinny, long or short. An example
of the response pattern of a neuron that responds best to vertical bars is
shown in Figure 4.3.
Together, the whole set of orientation-selective neurons in the thalamus and the visual cortex integrate what angle is it? together with
where is it? The answer to where is it? comes from the general location of a neurons receptive field. Is it up, down, left, right, or straight
ahead? And every single location in the visual scene is watched
by a different subgroup of neurons in the visual cortex. Within that
subgroup, some neurons respond better to horizontal angles, some
to vertical, some to diagonals, and some to angles in between. Also
within that subgroup, neurons with similar orientation preferences
79
Stimuli
Orientation-selective
responses
Stimulus
on
Stimulus
off
Voltage
Time
80
M A K I N G SPAC E
FIGURE 4.4 This painting by Chuck Close, Self Portrait, 2007, reminds me of the organ-
ization of the visual cortex. A coarse scale gives the overall shape: the face of the artist
in the painting or the map of the visual scene in the brain. At a fine scale, Closes use
of pixels containing small shapes, round or oblong in various orientations, calls to
mind the receptive fields of individual neurons and their selectivity for orientation at
particular locations.
81
But, as yet, it does not group sets of these high-contrast regions together.
Edges identified at one location in the visual scene are not yet linked
with those found elsewhere. The contours of your coffee mug and the
table it is resting on both elicit enhanced responses in corresponding
regions of the visual map, but each is emphasized equally, and there
is no indication that the contours associated with the mug are distinguished from those of the table.
Signs of grouping of contours based on the object with which they
are affiliated are found in a subset of neurons in the visual cortex.
These neurons respond differently to the same light-dark contour in
their receptive field, depending on what object the rest of that contour
encircles.
Thats a complicated thought that requires delving a little deeper
into what we mean by object. In the case of a coffee mug on a table,
the meaning of object is clear: a physically contiguous, tangible item
distinct from other nearby items. The coffee mug is distinct from a
background: an area that might well also have physical objects in it but
which may be farther away, partially obscured, and/or more amorphous. Backgrounds are less salient than objects.
Distinguishing an object from a background can be easy or hard.
The coffee mug stands out effortlessly not only from the table but also
from very background-like regions such as the floor or walls. But lets
consider some illusions that are more ambiguous, such as the illustration in Figure 4.5.
What do you see? You should see one of two things. Either you
should see silhouettes of two faces looking at each other, or you should
see a vase. Whichever one you didnt see at first glance, look again,
and see if you can perceive it now that you know what to look for. This
82
M A K I N G SPAC E
83
FIGURE 4.6 Some neurons in the visual cortex respond differently to the same pattern
of light and dark in their receptive fields, depending on whether the dark part forms
the object or the light part does. Consider a receptive field indicated by the oval: a
and b have the same light on the left, dark on the right pattern within that oval (c),
but in a the light side is part of a square against a darker background, and in b the
dark side is part of a square against a lighter background. Neurons that distinguish
between a and b in their response patterns are said to be sensitive to border ownership. What kind of circuit enables this sensitivity is unknown.
the receptive field receives exactly the same light-dark stimuli either
way.3 This is called border ownership sensitivity (Figure 4.6).
Nobody knows exactly how these neurons arrive at this kind of
sensitivity. But it must somehow involve connections coming from
neurons with receptive fields at other locations, connections that mold
the responses to the light within the receptive field based on the pattern
of light over a larger scale.
The vase/profile illusion might reflect the relative contributions of
populations of border-ownership neurons that respond better when the
object is dark versus when the object is light. Perhaps when you see the
faces, it is because your dark-border-ownership neurons have the slight
upper hand. If you see the vase as the object, the light-border-owner-
84
M A K I N G SPAC E
FIGURE 4.7The Kanizsa triangle (a): You should perceive a white triangle partially
obscuring three black circles. The edges of the triangle are not explicitly formed by
anything in the image except at the corners/black circles. These are known as illusory
contours. Possible neural circuitry that might contribute to the perception of illusory
contours (b): The dashed outlines indicate the receptive fields of three neurons sensitive to vertical bars, with excitatory responses for dark bars in the center of the receptive field. If the corners of the Kanizsa triangle image align as shown with two of the
three receptive fields, these neurons would exhibit net excitatory responses. Note
that the stimulus is not the perfect, or optimal, stimulus for these receptive fields, but
it is likely good enough to excite the neurons to some degree. If these two neurons
share their excitation with the neuron in the middle via a synaptic connection, then
all three neurons would be active, and their activity might be interpreted as indicating
a line or contour extending continuously from one corner to the other.
85
* Of course, its not really a direct response; the neuron responds because it receives
input from other neurons, activity that originated with photoreceptors. The distinction here is between activity that is triggered by an actual stimulus as opposed to
inferred from the activity of neighboring neurons.
86
M A K I N G SPAC E
-16
-14
400
10
rays
10 22
-12
Blue
10
10 20
10
10
10 16
-8
10
-6
Green
Wavelength (
600
, nm)
Yellow
-4
10 12
10
Infrared
10 14
Visible spectrum
500
-10
X rays
10 18
10
-2
AM
10 6
700
10
10
10
10 0
10 2
Increasing wavelength ()
10
10 4
Increasing frequency ( )
Radio waves
FM
10 8
10
Red
Microwave
10 10
l (m)
(Hz)
of electromagnetic radiation is between about 400 and 700 nanometers, we can see it, and we therefore call it light.
PLATE 1 Electromagnetic radiation is all around us, but only a tiny portion is visible to the human eye. If the wavelength
10
10 24
Ultraviolet
PLATE 2 Pierre-Auguste Renoirs painting Luncheon of the Boating Party is annotated to illustrate how
light travels in straight lines but in multiple directions, leading to the mixing of light from many sources.
Reflected light spreads outward from any given location in this scene, such as the wine bottle. At
any other location, such as the shoulder of the young woman on the left, light converges from many
sources. With such a mlange, how is it possible to know where any given photon came from?
PLATE 3 A photograph of the state capitol building in Madison, Wisconsin. When viewed with a
pair of spectacles with a red lens in front of the
left eye and a blue lens in front of the right eye,
the distance of the statue in front of the state
capitol will pop out. The blue lens filters out
the blue image, leaving only the red one, and
vice versa. Thus, each eye will see the statue at
a slightly different position on the retina, at relative positions that correspond to the statue being
much closer to the viewer than the capitol dome.
PLATE 6 The cyclist is at different positions at different moments in time. To detect this motion, the brain
compares the visual scene at different moments in time, delaying neural activity in response to one
image so that it co-occurs with neural activity in response to a later (real-time) image.
Tyto alba
an ant on stilts for its return journey home after a day of foraging
far afield, and it will overshoot the
nest. Trimming an ants legs so that
it must walk on shortened stumps
causes it to begin looking for the
nest before it has covered the
necessary distance.
Head-direction cell
Grid cell
North
Location
East
South
West
less
more
Activity Level
Activity Level
Place cell
Activity Level
less more
Location
PLATE 10 The neighborhood of the hippocampus contains three types of signals relevant to spatial
navigation. Neurons known as place cells (a) respond when the rat is in a particular location in its environment. This example shows greater activity when the rat is in the bottom left side of a circular enclosure. Head-direction neurons (b) are tuned for the direction a rat is facing, schematically depicted here
as a preference for facing southeast. Grid cells (c) are active as a rat moves throughout its environment
at a regular spatial interval. Different place, head-direction, or grid cells would respectively prefer
different locations, directions, or spatial intervals (finer or coarser granularity).
ory
Motor
s
osen
at
Som
Visual
Visual
Olfactory
Auditory
PLATE 11 Most of the brain has been implicated in sensory or motor processing of
one kind or another. This figure shows views of the lateral (a) and medial (b) sides
of the cortex of rhesus monkeys, flattened out in c, showing the kinds of sensory or
motor activity that have been found in each area. In humans, there are many more
areas, whose response patterns have not all been fully explored.
Motor
ry
enso
atos
Som
Visual
Attention
Visual
Memory
Planning
Olfactory
Auditory
PLATE 12 Studies of memory, attention, and planning have generally found that the
areas of the brain that seem to play a role in these tasks seem to overlap with the
areas of the brain that show sensory and motor-related activity.
Head, neck
Trunk
Shoulder
Arm
Wrist
Eye
Nose
Fingers
Thumb
Lips
Teeth, gums, jaw
Tongue
Pharynx
Intra-abdominal
Leg
Foot
Genitals
87
entire length of the legs. This happens because there are more receptors per unit area on the face and hands and comparatively few on your
shin. More receptors mean more brain tissue is required to process that
information. But these densely packed receptors provide more finegrained information about where the stimulus is located, how big it is,
what shape, and so on.
Heres a test you can do to feel how this works. It works best if you
close your eyes and have a partner do this to you, but you can try it by
yourself. Take a paper clip and unfold it so that the ends are sticking out
together like a U, about a quarter of an inch apart. Poke the fleshy part
of your thumb with the prongs. You should be able to clearly feel both
prongs. Now, poke the back of your thumb. It should be hard to tell that
it is two prongs, not one. If you pry them apart a bit more, maybe threeeighths or a half inch, you should be able to clearly distinguish each of
the two prongs on the back of your thumb. But now try that on your
forearm. Not wide enough, are they? Here, an inch of separation may
be required.
We have the highest spatial resolution in areas of the body that play
the most important roles in tactile exploration or position sensing. Our
hands reach out to touch objects and provide information about their
shape and texture. When eating, our mouths are very sensitive to how
soft or hard our food is, and our tongues position each bite for our teeth
to chew on. Babies and many animal species also use their mouths to
explore things that are not food. And when we talk, our tongues play yet
a different role, helping to shape the sounds of speech phonemes; very
subtle differences in tongue position can mean the difference between
an R sound or an L sound. In contrast to your hands and mouth, your
backs role is primarily to physically support your arms and legs, so it
88
M A K I N G SPAC E
has no need for a fine-grained sense of spatial location. Oh well, its not
like you can easily reach your back to slap at mosquitoes anyway!
This distortion of space also happens in the visual system, at least in
humans and some animals. Photoreceptors are stuffed cheek by jowl
into the center of the retina, an area known as the fovea. This supports a
very high degree of spatial resolution, but only for the part of the visual
scene whose image lands on the fovea. Stare right here at this spot on
the page right now, and without moving your eyes, try to identify the
letters a line or two up or down from where you are reading. It should
be pretty hard to do. You can really only see clearly what the fovea is
looking at. The foveal part of the visual scene commands a larger swath
of territory in visual areas of the brain, similar to the magnified zones
for touch.
We have seen that the brain uses maps to keep spatial information
organized, and that neurons have receptive fields and patterns of
connections with other neurons that parallel key aspects of spatial
perception. But how do we know that these biological features really
cause the mental experience of perceiving? Certainly, not all brain activity produces awareness. For example, you probably have no idea how
fast your heart is beating right now, even though your brain controls
your heart rate just as surely as it does your thoughts and sensations. So
how does the mental experience of awareness, or consciousness, arise?
While the mind-brain connection is far from well understood, evidence
is mounting that some of these maps make an important contribution
to our mental lives. Consider the following examples.
People who lose an arm or a leg through amputation or injury may
still feel their missing limb. The lost body part can feel itchy, sore,
89
or achy, for example. The limb is gone, so such sensations are not real:
they cannot be triggered by an actual case of poison ivy, a blow from
an errant baseball, or an ache from too much skiing. But such feelings
are real to the amputee. And they probably occur because neurons in
one of these brain maps, in an area that used to receive input from the
missing body part, are active by accident, without some actual stimulus
to drive them. These neurons may be exhibiting what we call spontaneous activityelectrical pulses that occur even though the neurons no
longer receive their usual inputs. And the next set of neurons, those that
receive input from these neurons, have no way of distinguishing activity that was triggered by a real input signal from activity that occurred
spontaneously. All signals are interpreted as real. So when neurons that
are no longer getting input from the now-absent limb fire anyway, their
activity creates a sensation that the limb is still there and that something is touching it or making it sore or itchy. This phenomenon is
known as phantom limb.5
Also, gaps in your maps cause you to be unable to perceive stimuli
at the missing location. This can occur due to brain damage, such as
when a stroke interrupts the blood supply to a particular area of the
brain, depriving the neurons there of oxygen and killing them. The
area damaged by a stroke can be small if the blood vessel is small, or
massive if it is a larger pipe that fails. The small strokes are the interesting ones, for our purpose here. If a small vessel in the visual cortex
leaks or is blocked, then the dead neurons correspond to just a particular part of the visual scene, and you are blind at that location, but just
that location.
Gaps in sensory maps also occur naturally. In fact, everyone has two
holes in their layers of photoreceptors, one in each eye (Figure 4.9). In
90
M A K I N G SPAC E
Optic nerve
FIGURE 4.9 The retina has a gap with no photoreceptors, known as your blind spot,
where the optic nerve forms. You are unable to see anything at that position.
mammals, the photoreceptors are at the back of the retina, and the other
neurons, including retinal ganglion cells, lie in front of them. Where
the axons of the retinal ganglion cells collect together and pass through
the layer of photoreceptors to form the optic nerve, there is a gap in the
map. The photoreceptors are pushed aside. This leaves you with a blind
spot where no photoreceptors are available to monitor light. You probably dont notice this very often. When blind spots are small, our brains
fill in the missing details by extrapolating from whatever is around
the hole. Have a look at Figure 4.10.
Cover your right eye, and look right at the cat with your left eye. Try
to keep your gaze steady on the cat. Now move the page slowly closer
to your face. Is there a distance at which the mouse disappears? This
91
FIGURE 4.10 Cover your right eye and look at the cat. If you hold your gaze steadily
enough and if you hold this illustration at the correct distancelikely about a foot
awaythe mouse should disappear. What else do you see when this happens? The
correct distance depends on the size of this illustration. Your blind spot is about 15
degrees away from your fovea. If the cats face is about 3 inches from the mouse
about right for the print version of this bookhold the drawing a little less than a foot
away. If you are viewing this on a smaller screen, and the cat is closer to the mouse,
hold it correspondingly closer to your eyes.
should happen when the page is about twelve inches away from your
eye. What do you see when the mouse disappears? Nothingness? Or
does it look like the cage is intact, but now empty? The bars of the cage
should fill in the blind spot. The bars extend slightly beyond the edges
of the blind spot, whereas the image of the mouse is entirely contained
within the missing zone. So the brain assumes that the bars would
continue straight through. Since there are no mouse edges to be seen,
there is no way to fill in the actual mouse. This filling-in process is akin
to that involved with the Kanizsa triangle as well. Edges interrupted
by a gap but continuing smoothly on the other side are assumed to
continue straight across. Your brain makes stuff up, but it does so from
the data it has available, in both cases by extrapolating from nearby
visual cues.
92
M A K I N G SPAC E
93
motion involves comparisons of sensory informationlight at different locations. But for motion, the comparisons occur not just in space,
but also in timehow the light at different locations changes (Plate 6).
The precise structure of brain maps is important for detecting motion.
When such maps are damaged or tricked into erroneous activity, curious illusions result.
Evolutionarily speaking, things that move are important to notice
because they may be alive. A living, moving thing can eat yoube it
a bear or a mosquitoor it might taste good if you can catch and eat
it yourself. Your odds of survival increase if you can detect movement.
Put negatively, if you could not tell when something in your environment is moving, youd be toast pretty quickly. Natural selection has
therefore built a specialized system in the brain designed to identify
movement.
Fundamentally, determining if something is moving involves
comparing what we see at one moment in time with what we see at
another moment in time. If they are different, something in the scene
might be moving. But how does the brain compare information across
time? In the case of motion vision, the brain is thought to shift samples
of the visual scene in time. By processing some samples rapidly and
others with a delay, the brain can compare newer and older information
to determine if anything is moving.
To accomplish this, the brain is thought to use specific electrical and
chemical qualities of neurons to manage traffic (in this case, action
potentials), much as a system of speed limits and traffic lights might do.
One way involves variations in conduction delaysthe time it takes for
signals from one stage of processing to reach the next. Action potentials
94
M A K I N G SPAC E
95
more distant past along the slow route. Via the fast input channel, they
would see what the world looks like now.* But via the slow channel, they would see what the world looked like a few milliseconds ago.
By comparing these two snapshots of the visual scene, the brain has a
handy way of detecting any moment-by-moment changes, changes that
could signify that something has moved.
Just as we saw earlier for edge detection, combinations of excitatory
and inhibitory synapses can serve to highlight changes in the visual
scene in time. The input from one pathway would use an excitatory
synapse and the input from the other would use an inhibitory synapse
neurotransmitter receptors that make the recipient neuron less likely to
fire its own action potentials. Suppose it is the slower pathway that uses
inhibitory synapses. For places in the visual scene where the now and
2-ms-ago images are the same, the now-excitatory and the 2-ms-agoinhibitory signals would be in balance and the membrane potential
would stay close to its baseline. But where the image has changed, the
now and 2-ms-ago signals would be out of balance with each other,
with either greater-than-baseline or less-than-baseline activity.
But changed doesnt necessarily mean moved. The dual, fast-and-slow
neural circuit Ive just described would respond well both to things that
move and to things that dont move but merely change. For example,
the flashing yellow light at an intersection is a stationary but changing
* In fact, as we discussed in Chapter 2, the photoreceptors in the retina are quite slow
to convert light to neural activity, so now in the visual system is delayed right from
the start by tens of milliseconds, and delays accumulate at each subsequent stage even
for the shortest and fastest routes. Mentally subtract off these fixed delays for the
purposes of this discussion.
96
M A K I N G SPAC E
signal, whereas the cars are actually changing in position. To distinguish between simple change and actual movement, and to tell what
direction and speed something is moving, the brain is thought to employ
one additional trick: it compares not only snapshots from the same location and different times, but also snapshots from different locations and
different times. When snapshots from two different locations and times
are identical to each other, that can mean something has moved from
the first place to the second. In the illustration above, imagine our two
snapshots both contain the same bicycle. If the same bicycle can be seen
at two different locations at different times, the brain quite reasonably
concludes that the bicycle is moving.
The brain needs many such circuits, to monitor all locations in the
visual scene and to consider a range of possible directions and speeds.
The connections would likely involve a range of spatial and temporal
offsets. Interestingly, there are limits to the ranges of space and time
that the brain evaluates for the presence of motion, and there are corresponding limits to the motion that we perceive. For example, when you
look at a clock, the second hand will clearly appear to move, but the
minute and hour hands do not. In truth, all three hands are moving, but
the minute and hour hands move too slowly to be identified by the brain
as moving. Stimuli can also move too fast to be perceived as moving.
When a ceiling fan is on its lowest setting, the rotation of the individual
blades is perceptible, but at a fast setting, the blades blur togetherthey
are too fast for our motion system to pick out. (Just like the too-fast-tosee flicker rates of movies, televisions, and computer monitors.)
Although it is important to know when something is moving, motion
also presents a problem for vision: the features of moving objects are
harder to see clearly. If the image of an object keeps moving across your
97
visual map, any individual neuron will get only a brief glimpse. The
densely packed photoreceptors in the fovea might have only a limited
opportunity to provide a high-resolution view of the object.
To help solve this problem, we use a special kind of eye movement called smooth pursuit. During smooth pursuit, your eyes rotate
(smoothly) at the same direction and speed as a moving target. By
moving the retina the same way the target is moving, these tracking
eye movements stabilize the position of the image of the target on the
retina. Thus neurons at the corresponding locations in the brains visual
maps will have a chance to evaluate this image for longer periods, until
you decide to stop tracking it.
An interesting aspect of smooth pursuit eye movements is that you
cant move your eyes this smoothly unless you see something moving.
If you try, your eyes will naturally move in fast, jerky movements from
one place to another. These are the saccades we talked about in Chapter
3 (and which we will revisit in Chapter 6). You can demonstrate this to
yourself by working with a partner and watching how his eyes move. Sit
facing each other, and move your finger back and forth in front of your
partner. Have him follow the motion of your finger with his eyes. You
should be able to see his eyes gliding along, tracking your finger. Then
put your hand down and ask your partner to move his eyes following
the same trajectory from memory, without an object to follow. Now
you should see his eyes jumping abruptly from place to place.
Smooth pursuit also only works when the target is moving within
a certain range of speeds, neither too fast nor too slow. Eye movement
speeds are described in terms of degrees per second, corresponding
to the angular rotation of the eye about its axis. The slowest possible
smooth pursuit speeds seem to be about 1 to 2 degrees per second. A
98
M A K I N G SPAC E
99
100
M A K I N G SPAC E
cannot discern whether stimuli are moving. Most people when asked
to follow the motion of their own fingers can do so, and do so better
when they can see their fingers than when their eyes are closed or they
are sitting in the dark. But for LM, seeing them doesnt help; she can
track her fingers no better in the light than in the dark. Objects that
move seem to exist in one place and then in another, without being
seen as moving between them. Pouring a glass of water without causing it to overflow is difficult. Crossing the street is dangerous because
she cannot tell whether cars are approaching and how fast. She has
compensated for this by using sound, as a fully blind person would.
Studies in animals have confirmed that damage in area MT specifically
impairs vision of motion. In animals, lesions can be made deliberately
by injecting a small amount of a toxic drug or by using an electrode to
burn a small amount of tissue. Such deliberate lesions are useful because
they can be restricted just to a small area of the brain region, whereas
humans with MT damage have typically had strokes or brain injuries
that affect a much larger area. (The other advantage of animal studies,
from a scientific perspective, is that you can investigate how well the
animal can detect motion before, as well as after, the lesion, so you can
do a well-controlled experiment.) Small lesions in MT in animals cause
motion blindness only for the location in the visual scene that matches
the receptive fields of the lesioned areaconfirming that it is the loss of
the cells killed by the lesion that causes the perceptual deficit.
Damage is one way of assessing how neurons contribute to perceptionremove their contribution and see how perception is affected.
What about the opposite: what if neurons are forced to be active when
nothing is actually happening in the real world? This too is possible to
101
102
M A K I N G SPAC E
blood vessels. Some brain structures occupy reasonably consistent positions with respect to recognizable landmarks such as the characteristic
folding pattern of cortical tissue. But such landmarks are only visible if
the surgeon opens up a large area of the skull. Today, MRI and CT scans
allow surgeons to visualize the whole brain noninvasively. But in the
1930s to 1940s, that was not possible.
Penfield developed a method of mapping the brain that involved electrically stimulating neurons in different regions. For this to work, the
patient had to be awake during the surgery, so he or she could report
what he or she experienced in response to the stimulation. Fortunately,
brain surgery is not like other kinds of surgery. Although the brain has
about 85 billion neurons,8 these neurons report whats going on elsewhere in the body, not whats happening in their own backyardthere
are no pain sensors in the brain itself. So, local anesthesia can be used to
prevent pain associated with the scalp incision and opening of the skull,
with the patient otherwise awake. An electrode can be poked right into
the brain without the patient feeling it. When small amounts of electrical current are delivered to this electrode, neurons near the tip become
active. The more current, the broader the swath of neurons activated
by the electrode. Depending on the brain area, electrical currents as
1
small as a few microamps (about 1000
of the current drawn from a watch
battery) can activate enough neurons to cause the patients to experience something indicative of the role of the tissue being stimulated.
Penfield stimulated all over the brain in his patients, and he asked
them to describe what they noticed. Patients reported a variety of
sensationssights, sounds, touchdepending on where the electrode
was placed. When Penfield stimulated the somatosensory cortex, the
sensation would be felt at a particular location on the body surface.
103
104
M A K I N G SPAC E
105
SHERLOCK EARS
Sound is mechanical, involving the physical back-and-forth oscillation
of molecules (Figure 5.1). You can think of sound waves as wind, but
wind that jitters back and forth. The air molecules dont go very far,
unlike in wind where individual air molecules actually travel long
distances. They just bump their neighbors, and those neighbors bump
their neighbors, and so on. Sound waves also travel much faster than
windabout 770 miles per hour.
Some aspects of hearing are tied directly to these mechanical attributes. When a drummer strikes the head of the drum, the surface of the
drum moves rapidly up and down, displacing air molecules as it goes.
The more densely air molecules become compressed during this vibration, the louder the sound. The higher the rate at which they jitter back
and forth, the higher the frequency of the sound wave, and the higher
the pitch we perceive. When a soprano sings, the air molecules vibrate
more rapidly than when a tenor does. Our perceptions of loudness and
pitch derive from different physical aspects of the pattern of motion of
the air molecules.
FIGURE 5.1Sound consists of propagat-
compression
rarefaction
Air pressure
wavelength
amplitude
Distance
108
M A K I N G SPAC E
Oval window
(behind stapes)
Incus
Malleus
Ear
drum
Stapes
Round
window
Cochlea
and cause the eardrum to vibrate. The vibrations are transmitted to the cochlea by a
series of tiny bones: the malleus, incus, and
stapes. These bones concentrate the force
of vibrations onto a smaller membrane that
is part of the cochlea, the oval window,
causing it to vibrate in synchrony with the
eardrum. The bones provide a mechanical
advantage that focuses the sound energy
onto a smaller surface area, overcoming the
inertia of the fluid-filled cochlea. (You can
also see some of the structures involved in
balance in this diagram, a topic well return
to in Chapter 8).
When sound waves reach the ear, they cause the eardrum to vibrate
in synchrony with the air movements. The low-air-pressure segments
pull the eardrum outward, and the high-air-pressure segments push the
eardrum back in. This is the first biological response the body makes
to sound, but as of yet, it is not a neural response. The eardrum itself is
just a skin-like membrane and has no neurons.
To get to neurons, the vibrations of the eardrum are transmitted by
tiny bones to a spiral-shaped structure deep within the ear called the
cochlea (Figure 5.2). There, sound vibrations cause movement of another
membrane, the basilar membrane, so named because it is attached on
one end to the base of the cochlea (Figure 5.3). At the other end, it
floats loosely like a ribbon in the fluid. When the vibrations reach the
cochlea, the basilar membrane starts to undulate in time.
SHERLOCK EARS
109
Cochlea, unfurled
Tectorial membrane
Basilar membrane
Tectorial membrane
Basilar membrane
Hair cells
110
M A K I N G SPAC E
bend in one direction, these ions flood through the pores, and as the
hairs deflect in the opposite direction, the pores close and the motion
of the ions stops, creating electrical signals that track the sound. Thus,
movements of air molecules have been converted into electrical signals
in neurons! The hair cells, in turn, transmit their electrical signals to
individual neurons that make up the auditory nerve, sending information from the ear into the brain.
The physical nature of sound makes its location complicated to detect.
As described in Chapter 2, the pupil and lens ensure that light can only
reach a given location on the retina from a given direction in the world.
But such image formation cant happen for hearing because sound
waves are too bendable. Although sound can and does travel along
straight lines, it also curves around the objects and barriers in its path.
This isnt anything special about sound, really. Its just that how much a
wave bends around an object depends on the relative sizes of the object
and the wavelength. Light has a much shorter wavelength (less than
1
1000 of a millimeter) than the objects we see. Air pressure waves that we
hear as sound have frequencies between 20 Hz and 20,000 Hz, which
correspond to wavelengths between seventeen millimeters and seventeen metersmuch larger, and thus more bendable.
As sound waves are funneled down the ear canal to the eardrum, all
points of origin are pooled together (Figure 5.4). So, information about
sound location would seem to be lost, with no way for the spatial layout
of cochlear receptors to signal the spatial layout of sounds. But the
brain has evolved a clever solutionusing two ears, not one. Although
neither ear by itself can tell us where a sound is located, by working
collectively the two ears can provide the clues necessary to deduce the
SHERLOCK EARS
111
FIGURE 5.4 Unlike the eye, the ear cannot create an image of where sounds come
from. Sound waves coming from all directions travel together down the ear canal,
mixing all points of origin.
112
M A K I N G SPAC E
Sound
source
Greater distance
to far ear
the other ear at another point in time. The solution may also involve a
similar mental time shifting.
Heres how it is thought to work. Each ear sends neural signals evoked
by the sound into the brain via the auditory nerve on that side. Some
neurons in the auditory pathway of the brain are binaural, receiving
incoming signals from both ears (Figure 5.6). The binaural neurons are
thought to be sensitive to the relative timing of the sound wave reaching the two ears. Some neurons respond better when the sound arrives
at the left ear first, and other neurons respond better to the opposite
pattern. As we saw for visual motion, this kind of sensitivity can arise if
the signals originating in each ear travel along paths that take different
amounts of time. The head start that one ear got because the sound
arrived at that ear first could be eliminated by an appropriate handicap,
so that by the time the signals arrive at the binaural neuron, they are
coincident in time. Some binaural neurons would have longer paths to
SHERLOCK EARS
113
FIGURE 5.6Binaural neurons in the auditory pathway are sensitive to whether the
sound arrives at the left ear first or the right ear. This is thought to involve a mechanism similar to that involved in visual motion processing: the signal from the near ear
must be delayed relative to the far ear so that the signals from the far ear, which start
later, have a chance to catch up. The neuron in a would receive coincident inputs from
both ears if the sound is located to the left. The neuron in b would be more sensitive
to sounds located to the right.
receive input from the left and would therefore respond better to leftward sounds (because the right ear would have had time to catch up).
Others would show the opposite pattern and prefer rightward sounds.
That our brains can do this is pretty incredible. The time differences involved are infinitesimal, particularly for sound locations that
are not due left or due right, but some direction closer to straight
ahead. The smallest time differences we can detect are about 10 to 30
1
of a second)during which
microseconds (10 microseconds is 100,000
a sound can travel about one-eighth to three-eighths of an inch. This
corresponds to a separation between the sound sources of about one
to three inches if you are about five feet away. This is why you can tell
114
M A K I N G SPAC E
1,000 Hz
that your cell phone is ringing, even if it is sitting on a table right next
to someone elses.
Differences in loudness across the two ears can also be used to infer
location (Figure 5.7). The near ear receives the sound wave on an unobstructed path, which is louder. To reach the other side of the head, the
sound wave has to bend or diffract, producing a softer sound. People
with hearing loss in one ear often turn the good ear toward the sound
to eliminate the barrier formed by their own heads.
How much a sounds loudness is attenuated by the head-as-obstacle depends on where it is coming from. From straight ahead, there is
no difference at the two ears, but a sound located due right might be
fifteen to twenty-five decibels louder in the right ear than the left. The
SHERLOCK EARS
115
116
M A K I N G SPAC E
Fourier Analysis
a
Timbre
d
Penny whistle
Banjo
Air pressure
Air pressure
Flute
Time
Time
FIGURE 5.9 Most sounds can be thought of as consisting of more than one frequency
SHERLOCK EARS
117
118
M A K I N G SPAC E
Amplitude of wave
in ear canal
Amplitude of wave
at source
Frequency
Frequency
Location A
Location B
Location A
Location B
SHERLOCK EARS
119
120
M A K I N G SPAC E
Even if you hear through only one ear, spectral cues can give you
a rough sense of sound location. People who are deaf or hard of hearing in one ear (either permanently or temporarily due to congestion
or an ear infection) lack the usual interaural timing or loudness difference cues. Operating on their own, these spectral cues are less precise
than when used in concert with timing and loudness cues, but they are
certainly better than nothing.
You can test the importance of spectral cues for yourself. For example, to see how well you can localize sounds with spectral cues alone,
try putting a simple earplug in one ear. Then, close your eyes and
have someone snap his or her fingers. When you point to where you
think you hear the sound, you should be less accurate than when you
have unimpeded hearing. Plugging one ear distorts level differences
and spectral cues, although the brain may still be able to detect timing
differences (if you can still hear through the plugged ear).
Changing the pinna to alter your spectral cues will also affect
your ability to localize. To make a much bigger external pinna, take
a dinner plate and hold it up behind your ear. This big smooth surface
will catch more sound, producing a slightly different pattern of
filtering than your ear alone. You can also eliminate the folds of your
pinna by filling them with clay (just be careful not to get the clay in
the ear canal). Modifying your pinna size or structure in these ways
should change perceived sound location, particularly in the vertical and
front-back dimensions. Timing and loudness cues will likely persist in
providing accurate information about the horizontal dimension.
Timing, level, and spectral cues let humans localize sounds fairly
well for our purposes. But certain predatory species (owls, for example)
have evolved more sophisticated systems for sound localization to aid in
SHERLOCK EARS
121
nighttime hunting. Hunting aloft and over an extensive zone of darkness (both horizontal and vertical), an owl faces challenges an earthbound human does not. Humans often assume that important sounds
are at our own level or within a few feet of the ground. If someone calls
out to you from a second-floor window, you may take a few seconds
before you think to look up. But a flying owl must swiftly and accurately
consider not only left/right, but also forward/behind and up/down. The
cone of confusion for an owl stands in the way of a tasty lunch.
How they circumvent this has been studied extensively among a
particular species, the barn owl (Plate 7). Barn owls have specialized
features that facilitate the ability to localize sounds in both the horizontal and vertical dimensions, minimizing the cone of confusion
described earlier. Barn owls faces are crooked: their ears actually aim
in different directions. One ear points slightly upward, and the other
is directed slightly downward. This gives the barn owl a sound level
difference cue for the vertical component of sound location. Overall,
barn owls can localize sounds to within two degrees in both the horizontal and vertical dimensionsremarkable, and particularly so given
that their small head sizes mean that the timing and level difference
cues involved are smaller than those experienced by humans.
Humans are no barn owls, but we are still normally able to easily tell
from sound alone which direction a car is approaching from or which
student in the class has asked a question. Why then do we find it so
hard to localize a failing smoke detector? I recently queried a Listserv of
auditory neuroscientists and received a range of opinions.
Some attributed the problem to the vertical dimension and the cone
of confusion. However, this does not match my experience, largely
122
M A K I N G SPAC E
because it is not the vertical dimension that most puzzles me. Prior
knowledge that smoke detectors are mounted in the ceiling solves my
cone of confusion.
Some attributed the problem to an issue known as aliasing. For
certain types of sounds, it is difficult to determine when the sounds
heard at each ear line up with each other in time. Consider a sound
with a single frequency of about 2,000 Hz (about four octaves above
middle C in music). The wavelength of the sound will be about seven
inches, approximately the diameter of an adult humans head. If it is
located due left, it will arrive at your left ear first and at your right ear
about a half millisecond later. But as the sound goes on, relative timing
becomes ambiguous. Many successive peaks and troughs of the sound
wave arrive at each ear. Given a sound wavelength equal to head diameter, a peak will arrive at your left ear at the very moment that an identical peak from an earlier moment in time arrives at your right ear.
Since the two peaks are identical and are experienced simultaneously,
your brain may (logically but erroneously) conclude the sound is actually straight ahead. The possibility of matching up the sound waves
in multiple equally good ways creates ambiguity regarding a sounds
original location.
For sounds that are much lower in frequency than 2,000 Hz, this
problem disappears. If the wavelength of a sound is larger than the
diameter of the head, only one interpretation of the relative delay time
between corresponding parts of the waveform is possible. Consider a
500-Hz sound, with a wavelength of about four times the width of the
head. When it is located straight leftward, the peak arrives in the left
ear first, and by the time it arrives in the right ear, the left ear is only
about a quarter of the way through the cycle of the waveform. By the
SHERLOCK EARS
123
potential correspondences
Air pressure
Time
time the next peak arrives in the left ear, the peak in the right ear is
long gone. For such low-frequency sounds, waveform alignment possibilities other than the correct one would require our heads to be much
larger than they actually are. And even for higher frequency sounds,
if more than one frequency is present, the waveform becomes sufficiently complicated that a unique solution is generally available (see
Figure 5.12)provided the additional frequencies are not harmonics,
or integer multiples of the lowest frequency, in which case the period is
determined by that lowest frequency.
So does this aliasing problem actually apply to smoke detectors? In
my opinion, the answer is probably no. Aliasing likely applies when
the detector sounds an alarm steadily in response to actual smoke, but
not when it merely chirps briefly to indicate a dying battery. I recently
124
M A K I N G SPAC E
Power
Sound pressure
10
15
Time, ms
20
25
30
3,000
6,000
9,000
12,000
Frequency, Hz
FIGURE 5.13 The sound wave (a) and frequency spectrum (b) of a smoke detectors dead-battery chirp.
The sounds frequency is above 3,000 Hz, which you can see from the peaks on the spectrum graph. (Like
the display on a stereo equalizer, such graphs show how much energy of each frequency makes up the
sound.) This puts the sound in the aliasing danger zone. But the sound also has a sudden onset, which
should make it possible to locate based on timing. The multiple frequencies present here (the multiple
peaks on the spectrum graph) are harmonics, so they dont help very much with aliasing problems.
recorded a smoke detectors battery chirp in our house and found that
it had a fundamental or main frequency of about 3,100 Hz, with several
harmonics at higher multiples of that frequency (Figure 5.13). Thats
high enough to suffer from aliasingbut only when you have missed
the beginning of the sound. When sounds have a clear starting and
stopping point, as they do for dead battery chirps, those edges can be
used to overcome the aliasing that can occur during the middle of the
sound. The steady sound of an actual alarm would be another matter.
Two strong possibilities remain. The first involves how we make
deductions from the spectral content of sound. Interpreting spectral cues requires a basis for comparison. In principle, this comparison
can occur in a few different ways. We might compare the frequency
content of the left and right ears sound samples, or we might evaluate the pattern of attenuation within a single ear but across different
frequencies: which frequencies are quiet and which are loud? Or, we
SHERLOCK EARS
125
126
M A K I N G SPAC E
FIGURE 5.15 Preventing echoes requires lining the surface with sound-absorbing material. The picture
on the left shows foam wedges mounted on the walls of a special room called an anechoic chamber.
Sounds in a room like this are similar to those in a large, open space outdoors, with soft ground and
few surfaces to reflect sounds and create echoes. Its a much more convenient, not to mention safer,
place to study echo-free hearing than the method used in the photo on the right, from a 1936 study.
SHERLOCK EARS
127
FIGURE 5.16 Locating a smoke detector is hard when the sound is coming from a
different room. If the alarm sounds in the room youre in, the precedence effect will
eliminate echoes. Timing and level differences will confirm the location of the straightpath copy of the sound (thick line). But if the alarm sounds in a different room, the
sound arrives at your ear indirectly, having traveled through a doorway and reflected
off other surfaces. The direction from which the sound comes may have little relationship to its origin.
128
M A K I N G SPAC E
SHERLOCK EARS
129
sounds, such a comparison isnt possible, and it can be difficult to distinguish a near, soft sound from a far, loud one. Our scant experience with
the loudness of smoke detector beeps makes it hard to tell whether they
are coming from this room or the next.
Echoes provide the only cue to distance that does not require prior
knowledge (Figure 5.17). Because they travel farther than the straightto-the-ears copy of a sound, they are quieter and they take longer to
arrive. How much farther they travel determines how much slower and
quieter they are. For a sound that is close to you with respect to the
surrounding scene, the straight-path version is much louder than the
echoes reflected off the more distant walls. Coupled with the arrival
delay, that ratio (first-sound/echo loudness) helps indicate how far away
the sound is located. Note that this method alone cant provide information about the absolute distance of a sound: your brain cannot convert
this to a precise number of meters away. But it can indicate the relative
distance of a sound compared to your surroundingsnear the center
of the room or closer to the far wall, for example. When the auditory
direct-to-echo ratio is combined with a visually based estimate of the
distance of those walls, the brain may gain a sense of the likely distance
of the sound within that setting.
Echoes also provide clues to the size, distance, and hardness of walls,
ceilings, and even furnishings. Most people can hear the difference
between what acousticians call live, highly reverberant spaces and dead
ones with few echoes. Sighted individuals may not be consciously aware
of hearing these cues, but blind people make extensive use of them.
They even generate sounds for the purpose of hearing echoes, a process
known as echolocating. Tapping a cane makes a short, sharp sound,
ideal for detecting distinct echoes reflecting off adjacent surfaces. Even
130
M A K I N G SPAC E
FIGURE 5.17 Sounds arrive at your ears on a straight-line path that is much shorter
than the path the echoes take. The difference between the straight-line path and the
echo paths depend on how close the sound is compared to the reflecting surfaces.
For nearer sounds, the loudness differential of the straight-line path compared to its
echoes will be greater than for farther sounds.
SHERLOCK EARS
131
FIGURE 5.18Bats
echolocate, or use
sonar, to find their prey. After emitting
brief chirps, they can detect not only the
location but also details of an objects
size and shape from the resulting echo.
This illustration, similar to one by bat
specialist Donald Griffin, shows bat- and
insect-emitted chirps and echoes approximately to scale.
and thus the delay to hear the echo informs them about the distance of
the object. They are so good at detecting the echoes that they can fly
through a room strung with piano wires without crashing, and they
can catch a meal worm tossed into the air in midflight.1
The ability to deduce sound location from the clues of sound timing,
loudness, frequency, and echoes has to be learned and practiced. This skill
develops over the first few months of life and is continually fine-tuned. As
our heads grow, the separation between our ears increases, so the amount
of time it takes a straight leftward sound to reach the right ear after
arriving in the left ear increases as well (Figure 5.19). Our brains must
compensate for these changes during youth, and as we age, most people
experience some degree of hearing loss. If this loss is asymmetric, loudness differences across the two ears will be altered and will require mental
adjustment. Typically, hearing loss occurs so gradually that we are not
aware of recalibrating to retain an accurate perception of sound location.
132
M A K I N G SPAC E
60
50
40
30
0
5 10 15 20 25 30 35
Age (months)
Adult
her mothers. As she grows, she will have to continually readjust her interpretation of the timing differences of sounds at each ear and what they mean
for the location of the source of the sound. Where
sounds look like they are coming from, and how
the sound waves reaching the ear change as she
moves, will help her calibrate.
With a little practice, we can even learn how to localize sound with
a new set of ears. John van Opstal and his team at the University of
Nijmegen placed molds in the outer ears of several human participants.2
The molds changed the shape of the folds of the outer ears, so that the
participants experienced a pattern of spectral filtering different from
what they were accustomed to. The participants wore these molds day
in, day out for up to six weeks, periodically coming to the lab to have
their ability to localize sounds tested (Figure 5.20). The researchers found
that after a week or two, the participants had started to relearn how to
tell where sounds were coming from. By the end of the experiment, they
were nearly as good with their new ears as with their old ones.
But how did they figure out the right answer? If you arent sure
how to interpret the sound timing, loudness, or spectral information
from your ears, what helps you tell when you have done so correctly?
Two types of feedback are thought to be important: your own movements and vision.
SHERLOCK EARS
133
When you move your head (and ears), the timing differences, loudness differences, and spectral cues change. Monitoring how they change
with movement can reveal their relationship to locations in space.
Consider the cone of confusion: suppose the sound is somewhere to
your right, but youre not sure whether it is in front of you or behind
you. If it is in front and you turn your head to the left, the sound will
be now be farther to the right (relative to your ears) (Figure 5.21). If
it is behind you, it will be closer to your midline. Whether the interaural timing and loudness differences increase or decrease when you
turn your head toward or away from the sound can help you determine
whether the sound is in front or behind you.
Similarly, cocking your head a little to the side can help resolve the
ambiguity in the vertical dimension. When you tilt your head, you tilt
that axis and now differences in true vertical location should produce
(modest) differences in timing and loudness across the two ears (Figure
134
M A K I N G SPAC E
5.22). Of course, to do all this requires that your brain know how much
you have turned your head! So, localizing sound also involves your
sense of body position, as we saw in Chapter 3.
The second way of calibrating sound localization cues involves
vision. When you see something that could be making the sound you
hear, you are likely to think the sound is coming from that visual object,
an illusion known as visual capture of sound location. Ventriloquists
have capitalized on this tendency for millennia. The word ventriloquism
comes from the Latin for to speak from ones stomach. The practice of ventriloquism dates back to ancient Greece and was said to be
used at the Oracle of Delphi. Somewhat more recently, ventriloquism
was popularized during the vaudeville era, notably by Edgar Bergen.
Comedians such as Shari Lewis brought ventriloquism to television in
the 1950s, and more recent practitioners, such as Jeff Dunham and Nina
Conti, are easily found on YouTube.
SHERLOCK EARS
135
136
M A K I N G SPAC E
from the location of the person on the screen. When you watch a ventriloquist on a video screen, you are experiencing the illusion twiceyou
hear the ventriloquists voice coming alternately from her own mouth
and then from the puppet. Neither is actually true!
Vision is able to fool us in part because we normally have good
reason to trust it: many sounds are accompanied by something visible. When you see more than one thing, your brain has to determine
which one is making the sound. Ventriloquists subtly alter the weight
of the evidence in favor of the puppet instead of their own mouths,
overcoming our previous knowledge that puppets dont actually talk.
The synchrony of the movements of the puppet with the sounds of the
speech supports this illusion.
Returning to our chirping smoke detector, vision actually impedes
the process of finding the correct alarm. When you are in the wrong
room, ventriloquism by the image of the smoke detector may make
you think you have found the right one when you have not. Only if you
have the kind where an LED flashes to indicate the dead battery does
vision help in this circumstance.
The effects of vision can be both strong and persistent and seem
to contribute to our learning of sound location. Compelling evidence
for this principle has been obtained by Stanford neuroscientist Eric
Knudsen and his team. Knudsen and his colleagues tested the effects of
chronically shifting the visual scene on barn owls and their ability to
localize sounds. To do this, they made prism glasses small enough to fit
the birds. Like George Strattons prism goggles, these glasses changed
the visual scene, but not as severely. Rather than reverse it, the visual
scene was merely shifted a small amount to the side. The birds were not
quite as flexible as Stratton in adapting to their new visual scene, but as
SHERLOCK EARS
137
138
M A K I N G SPAC E
chapter, well take a detour to discuss a kind of auditory map for nonspatial information, and how this map is critical for restoring hearing with
an auditory prosthetic. Then in Chapter 6, Ill turn to an alternative to
maps, a different form of brain coding used for some types of auditory
and movement-related spatial signals.
As we saw previously, visual and somatosensory maps originate in
the topographical layout of photo- and touch receptors. The spatial
layout of the receptors in the cochlea cannot provide a similar image of
sound location, but it does provide information about the soundinformation concerning sound frequency.
Sensitivity to sound frequency relates to resonance in the cochlea.
Resonance is the tendency of a system to oscillate more strongly at
some frequencies than at others. For example, each string on a guitar
resonates at a different particular frequency. Playground swings are
another example. A swing on a long rope will oscillate slowly and a
swing on a shorter one will cycle more rapidly. You must pump your
legs (or be pushed) at a rate that matches the resonance of the swing in
order to maximize its motion.
In mammals, the basilar membrane (the flexible tissue holding the
hair cells) has a resonance gradient along its length. Each location
along the basilar membrane oscillates more strongly for some sound
frequencies than for others, and which frequency produces the strongest oscillations varies along the length of the basilar membrane. The
base is like the shortest swings on the swing setit oscillates rapidly
and most strongly for the highest frequencies. The basilar membrane
is very stiff herelike the tautest string on a violin or guitar. Near the
apex, or tip, of the cochlea, the basilar membrane is more flexible, like
the strings on a bass guitar, and oscillates most vigorously for lower
SHERLOCK EARS
139
frequency sounds. This variation in stiffness, and concomitant variation in sensitivity to sound frequency, produces a neural map of the
frequency content of sound. When the hair cells at a particular location
on the basilar membrane fire (generate action potentials) like mad, it
means the sound contains energy at the frequency at which that location on the basilar membrane resonates most strongly.
Cochlear implants, a type of auditory prosthetic, capitalize on this
map of sound frequency to restore hearing in some patients with hearing loss. In these patients, the hair cells have died, eliminating the
neurons responsible for translating mechanical motion into neural
activity. Without their input, auditory nerve fibers go silent and the
brain has no way to detect sound. Cochlear implants bypass the missing hair cells using an array of electrodes inserted into the cochlea.
The electrodes electrically activate the neurons of the auditory nerve,
replacing the lost function of the hair cells.
The stimulation delivered through these electrodes is patterned to
mimic the way the auditory nerve would normally respond to sound.
A miniature computer analyzes incoming sounds to determine which
frequencies are present (using Fourier analysis). It then activates the
electrodes selectively based on those frequencies. Stimulation is delivered at the base of the cochlea for high-frequency sounds and nearer the
apex for lower frequency sounds. Getting this pattern right is essential
to patients ability to understand speech. The difference between a d
and a t, or an a and an e boils down to different amounts of different
frequencies at different times. The electrode activation pattern mimics
this (though imperfectly; patients must relearn how to interpret the
stimulation-induced activity patterns).
The map of sound frequency is carried forth from the auditory nerve
140
M A K I N G SPAC E
SHERLOCK EARS
141
Weve already discussed one of the particular ways that the brain
encodes and manipulates information, the neural map. In this chapter, well explore what maps are good for, and well see that they have
some limitations. Well discover that the brain has an alternative form
of encoding, which Ill call a meter. In a map, the locus of neural activity
represents information, but in a meter, the amount of activity represents
information.
To illustrate the conceptual differences between maps and meters,
lets develop an analogy with modern electronic devices such as
computers, televisions, cell phonesany electronic-signaling machine.
The representations used by computers and their many electronic cousins come in two flavors, digital and analog, paralleling the brains maps
and meters.
First, digital coding. Suppose the piece of information encoded
in such a representation is a mathematical quantitya number. In
a digital code, a number is physically represented in a computer by
a series of transistors, organized to form logic gates. Each logic gate
can be in one of two states and can be grouped together to represent
many more values by decomposing the number into powers of two
(see Figure 6.1).
FIGURE 6.1 The zeros and ones of a digital code. This example corresponds to 23 + 24
+ 25, or 56.
144
M A K I N G SPAC E
Here, the leftmost digit or bit signals the presence or absence of 128
(27), the next position indicates the presence or absence of 64 (26), and so
forth. Totaling these components and expressing in our most familiar
base-ten representation of numerical quantities gives:
(0 128) + (0 64) + (1 32) + (1 16) + (1 8) + (0 4) + (0 2) + (0 1) or 32 + 16 + 8 or 56.
Neuron activity
on
off
FIGURE 6.2 We can think of the brains maps as being like a digital code. The pattern
145
In both digital codes and maps, the location of each element has
meaning. A zero or a one in a particular location in the logic gate group
signals the absence or presence of a particular power of two. If a given
neuron is firing in the brains visual maps, this signals the presence of
a visual stimulus at a particular location in the visual scene. If it is not
firing, that indicates that no visual stimulus is present at that location.
Of course, this is an oversimplification, as neurons actually have
graded, not binary, responses. That is, neurons are not simply on or
off but can generate widely varying numbers of action potentials.
This property resembles analog, rather than digital, signaling. In an
analog code, the magnitude of some signal scales continuously with the
information to be encoded. In an analog electronic device, the signal is
usually voltage, but rather than being limited to two values (e.g., 0 and
5 volts), the voltage level might scale smoothly and continuously. For
example, the quantity seven might be represented by a signal of 7 volts,
the quantity three might correspond to 3 volts, two would be 2 volts,
and so on. In this type of code, the value and the signal encoding that
value covary smoothly, regularly, and perhaps even proportionately.
Picture a meter with a needle on its display: how far the needle deflects
indicates how much of the measured thing there is (Figure 6.3).
Neural activity displays an ample dynamic range for analog coding,
varying continuously between zero to hundreds of action potentials
per second. Thus, any given population of neurons can simultaneFIGURE 6.3 When the brain uses a meter instead of a
map to represent something, the encoding of information depends more critically on the firing patterns of
individual neurons (less versus more activity) than on
which neurons are responding.
146
M A K I N G SPAC E
Less
neural
activity
More
neural
activity
ously participate in two types of code: which neurons are on and how
strongly they are on.
The two types of codes generally signal something different from
each other. In vision, the amount of firing in a neuron is related to how
bright (or dark) the light is. More photons affect a larger number of
photopigment molecules, producing a larger impact on the electrical
potential of the membrane. This, in turn, causes a larger change in the
number of neurotransmitter molecules that are released and a greater
impact on the level of activity exhibited by neurons farther along the
chain of command. Staring at the sun versus the moontwo visual
objects with very different levels of brightnessproduces a difference
in how strongly neurons respond. Similarly, in touch, the amount of
pressure applied to a touch receptor will cause a proportionate effect
on the receptors response. Squeeze someones hand, and your touch
neurons will respond vigorously. A gentle caress? A milder response in
your brains touch pathways.
In these two cases, neural meters encode nonspatial information.
But in other circumstances, neural firing rates do reflect information
about space. For example, our brains use analog coding to control the
movements of our bodies in space. Reaching out your hand to catch
a Frisbee, lifting your foot to step over a fallen tree branch, dancing a
tangoall depend on the brains use of analog signals.
Movements are analog because the duration and speed of a movement covaries with the strength of a muscle contraction. Muscle contractions are controlled by your brain via neurons that form synapses with
muscle fibers (Figure 6.4). Neurotransmitter molecules released by
so-called motor neurons are detected by receptors on the surface of the
muscle fiber, triggering contraction of the muscle fiber. The amount of
147
148
M A K I N G SPAC E
of the active muscle groups and the ratio of their contractions control
what body part moves and in what direction, and the vigor and duration of the contraction control the speed and extent of the movement.
When a (metered) movement is guided by a (mapped) sensory event,
such as when you point to something you see, the brain must translate
between these two internal languages. Such a map-to-meter conversion
is the brains equivalent of digital-to-analog conversion.
The best-understood case of how the brain accomplishes this
involves saccadesthe eye movements we first talked about in Chapter
3. Recall that the purpose of a saccade is to aim the fovea at some stimulus that youd like to see more clearly. Saccades can bring the fovea
to bear on the sources of sounds (looking for
someone whos called your name) or tactile
stimuli (a mosquito biting your arm or the
vibrating posts in my earlier experiment). But
most commonly, we direct our gaze to stimuli we can see. While looking at this page, you
might notice a bird outside your window with
your peripheral vision. You might then move
your eyes quickly toward the window to see
whether it is a sparrow or a wren.
Which photoreceptor neurons responded to
FIGURE 6.5 The muscles for moving the
the glimpse of the bird tells your brain where you
eye. Six distinct muscles are attached
need to look. The movement is controlled by
to the outside of the eye, organized
in three pairs. The combined action
six extraocular muscles, so named because they
of these muscles rotates the eye in the
attach to the outside of the eye (Figure 6.5). Two
horizontal and vertical dimensions, as
of these muscles rotate the eye horizontally while
well as in a slight twisting motion.
149
the other four pull in various combinations of up, down, and twisting
motions. Looking at the clouds above the bird or the tree branch below it
would all involve contracting the upward muscles but to different degrees
(Figure 6.6). Contract hard and the eyes will shift to the clouds, more
gently and the eyes will aim at the tree branch. Hit the oculomotor gas
pedal a moderate amount, and the eyes will land on the intended target,
the bird.
The digital-to-analog conversion involved in visual guidance of eye
movements is thought to rely on a brain structure known as the superior
colliculus. Colliculus is a Latin word meaning mound. There are two
sets of mounds on the top of the brain stem (the part of the brain that is
connected like a stem to the spinal cord). The lower pair, the inferior colliculi, are involved in hearing (more on them below). The upper pair, the
superior colliculi, are involved in controlling where the eyes are looking.
The superior colliculus serves as a bridge between the maps employed
by the visual system and the meters used in the motor system. Superior
150
M A K I N G SPAC E
colliculus neurons receive incoming signals from the visual system and
use those signals to control movements. They do so using a curious kind
of map: a movement mapsimilar to our hypothetical car with different gas pedals for different trips! Neurons at different locations in the
superior colliculus respond best when an eye movement of a particular
direction and length (or vector) is to be made. Adjacent neurons prefer
similar vectors, and across the whole structure (including the left and
right colliculi), all possible movement vectors appear to be mapped.
Electrical stimulation in the superior colliculus confirms this, causing the eyes to move (not unlike what we saw earlier for area MT)*
(Figure 6.7). The direction and length of the movement depend on
where in the superior colliculus the stimulation is delivered, and closely
match the receptive fields of neurons at each site.1 Stimulation in the
front-left quadrant will cause the monkey to make a small rightward
saccade. Stimulation farther toward the back will produce a similar but
larger movement, while other places will yield movements of varying
directions and sizes. The movements triggered by stimulating the superior colliculus match closely with the receptive fields of neurons at each
stimulation site.
We dont know for sure how the superior colliculus partners with
downstream areas (between itself and the muscles themselves) to
* Although stimulation in both the superior colliculus and MT can evoke eye movements, there are important differences between these brain regions. MT is a more
visual structure and the most common effect of stimulation is likely to modify the
sense of visual motion, which can then be expressed in a variety of different ways
depending on the behavioral task given the monkey. In the superior colliculus, stimulation appears to more directly evoke a specific saccadic eye movement, regardless of
the behavioral task that the animal might be performing.
151
rostral
25
FIGURE 6.7 The movement map of the supe-
lateral
medial
caudal
1 mm
152
M A K I N G SPAC E
Output
Activity
A B C D E
Location
153
154
M A K I N G SPAC E
There is one last step to actually move the eyes. Once you have
converted the superior colliculuss map into a meter somewhere downstream, how does that amount of activity present in the meter actually
cause the eyes to move to the correct spot? We dont know for sure, but
we think it works something like this: A set of neurons downstream
from the superior colliculus exhibit a firing rate that represents how far
youd like to move the eyes. These neurons project to a neural circuit
that implements a feedback loop. The how far the eyes should move
signal gets sent to the muscles, and a feedback signal corresponding to
how far have the eyes moved gets sent back. When the how far have
the eyes moved signal is equal to the how far the eyes should move
signal, the latter signal is turned off, stopping the eye movement.
When it comes to representing sound locations, the brain has multiple options. Vision, touch, and movements are forced to use maps or
meters because of how they are constructed. The brains visual system
uses a map because the optics of the eyes create an image on the retina,
and all the brain has to do to maintain this order is keep the strands
of the visual pathway from getting tangled. Same for touch and the
body surface. For movements, the mechanics of the muscles require
an amount-of-firing code. In short, the choice of digital versus analog is
dictated by the available hardware.
But for hearing, all bets are off. As we saw in Chapter 5, the brain
doesnt receive a handy-dandy guide to sound location directly from
the ear, the way the visual and somatosensory systems do from their
sensory receptors. Instead, it has to build one. So there are some alternatives. Does it build a map? Or does it build an amount-based, analog,
proportionate meter to represent sound location?
155
To answer this, lets revisit the brains mechanism for detecting binaural timing differences (probably the best studied of the brains multiple
cues for sound location). As we saw in Chapter 5, detecting a difference
in sound arrival time across the two ears requires neurons that receive
input from both ears. Slightly delaying the input from one ear relative to
that of the other can make the two ears signals, originating at the two
eardrums at different times, arrive at a comparison neuron at about
the same time. The comparison neuron responds better when the two
inputs line up; that is when the sounds additional physical travel time
to the far ear is precisely compensated by a matching excess of neural
travel time from the near ear (recall Figure 5.6).
So, does that make the code digital or analog? A map or a meter?
Surprisingly, the answer seems to be that it depends. In 1948, Lloyd
Jeffress proposed a theory suggesting that there were many delay
lines of varying lengths, corresponding to varying time shifts.4
Different comparison neurons, called coincidence detectors, would
respond best for different relative delays. This would create selectivity for limited spatial ranges. A coincidence detector neuron would
respond most strongly if the relative timing of the left- and right-ear
signals was just right for that neuron, based on its own unique set of
neural input pathways originating in the left and right ears. A neuron
with equivalent input timing from both ears would respond best to
sounds located straight ahead (or behindremember the cone of
confusion?). A neuron wired up to receive input from the left ear with
a slight delay on comparison to the right ear would respond best when
the sound got to the left ear with a matching head start. Thus, each
neuron would have a receptive field in space based on the specific staggered start created by the input path to that neuron. As a result, differ-
156
M A K I N G SPAC E
150
100
50
0
50
90
60
30
30
60
90
157
to most sounds, but how strongly they respond varies with location.
Neural responses are more vigorous the larger the interaural time
difference favoring the preferred ear. This suggests our old friend,
the meter: the amount of neural firing varies with how far the sound
is located to one side or the other. An analog code for the horizontal
component of sound location!
Loudness differences probably also produce an amount-based code
in neurons. Like the case for sensing time differences, loudness differences require some form of comparison between signals reaching the
two ears. In this case, the comparison may involve a combination of
excitation and inhibition, as we saw in Chapter 4 for detecting object
boundaries. The comparison neurons for detecting loudness differences might receive excitatory input from one ear and inhibitory input
from the other. For example, a left-preferring neuron might receive an
excitatory input from the left ear and an inhibitory input from the right
ear. The farther the sound is to the left, the louder it is in the left ear in
comparison to the right ear. The left-preferring neuron would receive
the strongest excitation and the weakest inhibition for leftward sounds,
and the weakest excitation and strongest inhibition for rightward
sounds. Similarly, right-preferring neurons might exhibit the opposite
pattern. In both groups of neurons, sounds on the preferred side evoke
larger responses than sounds on the opposite side. The amount of firing
in these neurons thus serves as an analog code for loudness differences
and can be used to indicate the horizontal component of sound location.
Weve recently discovered that maps and meters can actually be
mixed together. The superior colliculus employs a map when signaling
the location of a visual target of an eye movement, but in monkeys, for
example, it can use a meter when guiding the same eye movement to
158
M A K I N G SPAC E
the location of a sound! 6 In this circumstance, neurons retain the characteristics of the inputs they receive, that is, mapping of visual information but metering of sound location. The connections between the
superior colliculus and the muscles themselves must fit either style of
signal, getting the eyes where you want them to go regardless of how
the request to move was initiated.
Ill leave this topic with a final mystery. How the brain encodes the
frequency-based spectral cues to sound location is not at all clear. The
attenuation in the spectrum of a given sound varies with the vertical
position of sounds, but in a complex way. A given frequency will be
most muffled when the sound comes from a particular height, less so
for other heights, but different frequencies follow different patterns. For
example, when your friend calls to you from the second-floor window,
some frequencies contained in her speech will be dampened compared
to others; when she speaks to you from the base of a staircase, different frequencies will be most strongly affected. It is not known how
the proper interpretation of these spectral cues is created by the input
patterns to individual neurons. Memory of the spectrum of a familiar
sound or a comparison of the spectra across the two ears must somehow be involved.
Now that weve considered several different kinds of brain codes for
spatial information, were ready to turn to a related question: How are
spatial locations defined? As well see in Chapter 7, spatial locations can
be defined in a variety of different ways, and the brain must switch
among these different methods when synthesizing signals from different sources.
159
162
M A K I N G SPAC E
you dont see the blur. Even this very gap in vision eludes our consciousness: we see no blackout; we have no sudden sense of darkness.
These eye movements are useful because of the fovea, where the
photoreceptors are so densely packed. If the entire retina were similarly replete, there would be little to gain from shifting gaze around
objects could be viewed under high resolution anywhere. But with only
a small megapixel zone, the brain must constantly reorient it.
Chances are you have never really noticed this. And thats because
the brain stitches together the high-resolution viewpoints from successive eye movements, fusing them so that you think you can see the
whole scene clearly even though you really can only see a small piece
this way at a time (Figure 7.3). Assembling this composite requires
knowing where each clear area lies, or what precise direction the eyes
were looking when you saw it. Knowledge of eye position helps keep
the visual scene not only looking sharp but also stable, preventing the
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
163
FIGURE 7.3 We see clearly only the part of the scene we are looking right at, that
is, with the fovea. The rest of the scene looks substantially less clear. These partially
blurred photos illustrate what this might look like for several different eye positions
(ac). But you would not perceive it this way. Instead, your brain combines the clear
portions from different eye positions to create a mental picture that appears high
resolution throughout, as in d. Aligning these multiple, clear, fixated regions correctly
to make a clear whole requires accurate knowledge of how your eyes movedin what
direction and how far.
164
M A K I N G SPAC E
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
165
needs feedback to find out what happened in response to its requests. But
for eye movements, the request itself might be enough.
Helmholtz tested this theory by poking his own eye to see whether
manually moving it would affect his perception of the visual world.
The hypothesis was that if he could shift his eye position by hand,
without the brain having commanded the movement in the usual
way, then there would be a mismatch between the eyes actual position
and the position of the eyes recorded by the brain based on its history of
motor commands to the eye muscles. By pressing gently on his eyeball,
Helmholtz could slightly alter his direction of gaze. Confirming his
theory, he observed that the world did appear to move in a compensatory fashion. Even though the image motion was much less than
what occurs during a regular eye movement, the brain interpreted that
motion as movement of the visual scene rather than the eyeball.
You can try this yourself: close one eye and press gently on the outer
corner of the open one. The visual scene motion may be small and
subtle, but it should be there. Your stretch receptors will know about
the change in eye position youve caused by poking yourself, because
they are sensing position by measuring stretch, which occurs regardless
of whether your eye was moved by your eye muscles or because you
poked it with your finger. The stretch receptors are the eyewitnesses.
But the eye movement control centers wont know about the finger-poke
movement, since they are only in charge of controlling the eye muscles.
The poking-induced perceived shift of the visual world suggests that the
brain believes the eyes are steady because it has not requested otherwise
and that it ignores the contradicting testimony from the stretch receptors. The brain is behaving like a CEO who assumes his or her orders are
carried out and doesnt verify the actions of underlings.
166
M A K I N G SPAC E
More compelling support for the theory, though, required some way
to modify the amount of force created by a given motor command,
unbeknownst to the brain. In the early 1970s, a scientific team at the
University of Pennsylvania had an idea about how to do that. Their idea
involved the deadly poison curare.
Used by the indigenous people of South America for hunting, curare
causes paralysis by interfering with the communication between motor
neurons and muscles. Animals wounded by arrows or blowgun darts
dipped in curare suffocate when the paralysis progresses to their respiratory muscles. Needless to say, curare is not something you would
want to expose yourself to voluntarily. But the team at Penn, led by
Alan Rosenquist and graduate student John Stevens, did exactly that.2
Stevens and Rosenquist reasoned that when curare disrupted the ability
of eye muscles to respond to the brains commands, they might see an
effect comparable to that observed by Helmholtz: attempting but failing to move the eyes might cause the world to appear to move.
They began with low doses of curarelow enough that they would
still be able to breathe and even walk around. (But just in case, they
conducted the experiment in a fully equipped hospital room with
anesthesiologists on hand, ready to intubate and provide respiratory
assistance.) Although eye movements were still possible at these low
doses, the command issued by the brain no longer matched how far the
eyes actually went. And sure enough, the altered relationship between
effort and outcome of eye movements caused the visual scene to appear
to shift with each eye movement! In a brave (or foolhardy) extension,
Stevens even tested himself under full paralysis, using a faster-acting
related drug called succinylcholine. Since full paralysis would prevent
him from speaking, Stevens wore a tourniquet on one arm, limiting
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
167
the flow of the drug to that hand and preserving a means of communicating through gestures. In these sessions, full paralysis of the eyes
was achieved, and just as before, the visual scene appeared to change in
position whenever an eye movement was attempted.
Later experiments in monkeys using electrical stimulation confirm
that the brain uses copies of its own motor commands to determine
the locations of visual stimuli. In these experiments, a visual stimulus
is flashed up briefly but disappears before the monkey has a chance to
look at that locationwhich normally takes about a fifth of a second.
During that brief delay, stimulation can be applied to a brain area that
produces an eye movement. If the monkey knows about the eye
movement produced by the stimulation, then it will make an eye movement that takes into account that eye movement and look accurately
at the actual location of the flashed stimulus. On the other hand, if the
artificial motor command originates somewhere after the brain derives
its record of eye movements, then the monkey may seek the flashed
stimulus at a location that is shifted by the same direction and amount
of the induced eye movement.
When stimulation is applied to the superior colliculus in this paradigm, monkeys appear to be aware that their eyes have moved and
seek the remembered target accurately. However, when stimulation is
applied to certain locations in the brain circuit between the superior
colliculus and the muscles themselves, monkeys sometimes appear
unaware of their shifted eye position and mislocate the remembered
visual stimulus.3
The spatial information available to the senses of touch and hearing are
quite different from this eye-centered visual worldview. As we saw in
168
M A K I N G SPAC E
Chapter 3, receptors on the body surface inform the brain regarding the
locations of tactile stimuli on the body surface. With moveable bodies
and eyes, the possible relationships between skin locations and retina
locations are legion. If a mosquito bites your racket hand during a game
of tennis, you must look one way to see it during a forehand swing and
another to see the same bite during a backhand. And when we evaluate
both the weight of an object and its color, our brains must determine
which visual and tactile stimuli come from the same location in space
and thus from the same physical object. In short, the brain has to determine the correct skin-to-retina mapping.
The bridge between the body-surface map of the skin and the retinal
map of the visual scene requires information about the angles of all
the joints between your eyes and that particular location on the body
surface. The retinal location of your fingertip, for example, depends
on the relative positions of the eyes, neck, shoulder, elbow, wrist, and
knuckles.
When I was a graduate student, my advisor (David Sparks) and I
conducted a study to test one span of this bridge, exploring how information about eye position is incorporated into our map of tactile information. We knew that neurons in the superior colliculus respond not
only to visual stimuli but also to tactile stimuli. But we wanted to
know if these neurons kept track of the tactile locations with respect
to the eyes, that is, in a visual reference frame. To test this, we trained
monkeys to make eye movements to stimuli they could feel with their
hands, and we had the monkeys make these saccades from a variety
of starting positions. We then tested whether the responses of superior colliculus neurons to a particular tactile stimulus depended on
that initial eye position. We found that the tactile responses of superior
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
169
120
Spikes/sec
Spikes/sec
120
90
60
30
0
400 200
200
400 600
Time (ms)
90
60
30
0
400 200
200
400 600
Time (ms)
FIGURE 7.4 Vision and touch use different reference frames, at least at the moment of
detection in the retina or on the skin. But we frequently use these senses together,
such as when we determine whether a peach is ripe based on a combination of its
color and the softness of its skin. To ensure that the visual and tactile signals from the
same source (e.g., the same peach) can be matched up, the brain must somehow
convert these signals into a common reference frame.
The monkey superior colliculus has a representation of tactile stimuli that seems to
have been converted into a visual reference frame. Here, neural responses to touch
depend on where the eyes are at the time each touch occurs. In the example shown
here, a tactile stimulus delivered to the palm of the left hand evoked a vigorous
response when the eyes were directed below the hand (left), but not when the eyes
were above it (right).
colliculus neurons tracked eye position so that they could stay in register with visual responses (Figure 7.4).4 If a given tactile neuron had a
visual receptive field, then it would respond most strongly if the initial
eye position placed that visual receptive field over the location of the
tactile stimulus.
When the visual and auditory systems try to talk to each other,
they deal with a similar language barrier (Figure 7.5). The cues upon
which the auditory system derives its sense of locationvariations in
170
M A K I N G SPAC E
Line
of sight
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
171
* The other reason is that phones dont carry a full complement of auditory information: when sound is reproduced electronically, not all frequencies are included. Highend stereo systems do a good job of including a wide range of sound frequencies, but
phones usually provide a much more limited set.
172
M A K I N G SPAC E
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
173
head-centered
sound location
an inhibitory synapse
eye position
an excitatory synapse
Output:
eye-centered
sound location
To compute this from the brains available signals may be remarkably simple (Figure 7.6).8 Recall that (in monkeys at least) the horizontal component of sound location is encoded as a meter, with neural
activity scaling proportionally to how close the sound is to one side
of the axis connecting the ears. Eye position, too, resembles a meter,
with activity increasing as the eyes deviate to one side or the other of
the orbits. With both sound- and eye-position signals in meter format,
the brain might calculate sound location with respect to the eyes
using an inhibitory synapse to subtract the eye position signal from
the auditory one.
Weve now covered the three sensory systems that detect the locations
of stimuli external to usvision, hearing, and touch. Weve also considered how we know about the shape of our own bodiesthe positions of
our limbs and joints. Weve seen that the brain has multiple constraints
and options when it comes to representing spatial information and that
174
M A K I N G SPAC E
YO U R S U N G L A S S E S A R E I N T H E M I L K Y WAY
175
GOING PLACES
178
M A K I N G SPAC E
from Your Window contest on Andrew Sullivans blog, the Dish. Do you know where
this photograph was taken? If you dont,
its because what you can see from a given
vantage point is usually not enough to create
a sense of where you are. Knowing where
you are involves knowing how you got there,
which you do by keeping track of your movements. See the footnote at the end of Chapter 8 for the answer to this one.
GOING PL ACES
179
Posterior
canal
Superior
canal
Utricle
Horizontal
canal
Saccule
Cochlea
the three semicircular canals, the utricle, and the saccule. These structures
are located in the inner ear, adjacent
to the cochlea.
Our sense of the direction we are moving comes largely from our
vestibular systemour sense of balance. Though it was left off Aristotles
list of the five major senses, a sense of balance is nevertheless integral to
the way we live. It helps us stay upright and helps us know that we are
movingand where. Even if you sit in a chair (with wheels) and keep
your eyes closed while someone pushes you around, your vestibular
system can signal movement, direction, and speed. And it can signal
when an elevator you are on starts to move, and usually whether it is
going up or down.
Balance uses some of the same mechanisms as hearing and is actually
thought to have evolved first. The sense organ for balance is located in
the inner ear, near the cochlea, and the nerve that carries auditory information also carries information about balance (Figure 8.2). The balance
180
M A K I N G SPAC E
Cupula
Fluid drag
Semicircular canal
Head turn
hair cells, which are embedded in a gelatinous cone-shaped structure known as the
cupula. When the head turns, the fluid and
the cupula lag, deflecting the hairs. The
movement of the hairs is thought to open
or close ion channels in the membrane of
the hair cells, producing an electrical signal
related to the heads rotation.
organs have several parts: three fluid-filled rings called the semicircular
canals, and two other fluid-filled cavities called the utricle and the saccule.
Each ear has a complete set of these component structures, and each
component is responsible for detecting a different kind of movement.
The ring-shaped semicircular canals are responsible for detecting
rotations of the head. The three canals are oriented approximately at
right angles to each other. Two pairs are vertical, one angled diagonally
forward and the other diagonally backward. The third pair lies in the
horizontal plane. The three different orientations allow for detection of
motion in any direction. For example, when the head turns, the casing
of the canals move rigidly with the head (Figure 8.3). But the fluid within
them lags behind. Its like what happens if you rotate a glass holding
your iced tea. The ice cube stays pretty much where it is and doesnt
GOING PL ACES
181
turn with the container. The relative motion of the walls of the canals
with respect to their fluid contents is detected by some of our old friends
from the auditory system: hair cells. The bases of the hair cells are fixed
to the canal casings and move when the head moves. But as the fluid
lags behind, the hairs deflect, creating an electrical signal related to head
motion. How much the hair cells in a given canal deflect depends on the
direction and speed of rotation. The combination of signals in each ear
and in each set of canals allows detection of the exact direction of rotation of the headnodding, shaking, or tilting quizzically.
In contrast, the utricle and the saccule are more concerned with your
movements along a straight line rather than rotation. Like proverbial
boots, these structures are made for walking, rather than for your spins
on the dance floor. They can also sense backward, sideways, or vertical
motion (such as in an elevator), provided this motion is linear rather
than rotational.
Like the semicircular canals, the utricle and saccule are filled with
a fluid, but here this fluid contains tiny rocks called otoconia, Greek for
ear stones (Figure 8.4). The otoconia are about three times as dense as
the fluid, too heavy to be moved much by the motion of the fluid itself.
Instead, they move when the head changes its orientation with respect
to gravity, or when the head accelerates in a straight line. Picture what
would happen if you put a rock in a toy wagon and pulled the wagon
hardthe rock would slide to the back. Similarly if you lifted the
wagon, tilting its angle, it would slide to the low side. The motion of the
ear rocks is similar, and when they move, they deflect the cilia of hair
cells attached to the interior of the utricle and saccule. These hair cells
are organized in a kind of cowlick, so that as a group they can respond
to pressure from the otoconia in any direction.
182
M A K I N G SPAC E
FIGURE 8.4 Scanning electron microscope images of the otoconia, small rocks in the
When your sense of balance is working well, you dont notice it. But
when it makes mistakes, your sense of orientation in space suffers. For
example, spinning around too many times makes us dizzy. Thats from
the fluid in the semicircular canals getting caught up by the prolonged
motion of your head and starting to move along with the canal. When
you stop, the fluid continues sloshing around, creating a signal of relative
motion, which your brain interprets as you turning in the opposite direction. Actually turning that way will stop the sloshing more quicklybut
restrain yourself or youll re-create the problem in mirror image!
Too much alcohol can also make you dizzy. Alcohol affects your sense
of balance because it diffuses into the various substances in your balance
organs, changing their relative density and how much the hair cells
are deflected for a given movement. The resulting mismatch between
GOING PL ACES
183
actual movement and what your vestibular hair cells convey causes you
to feel unsteady.
This unsteadiness actually occurs in several phases. Alcohol first
diffuses into the gelatinous cupula in which the hairs of the hair cells
are embedded. Alcohol makes the cupula less dense or lighter than the
surrounding fluid, known as the endolymph. When someone lies down
after having had a few too many drinks, the cupula will float upward a
little bit, producing a spinning sensation. As time passes, alcohol starts
to diffuse into the endolymph itself. Temporarily, the density of the
endolymph and the cupula will return to their normal ratios and the
spinning should abate. But dont hop in the car to drive home yet
this is the eye of the storm. As the cupula starts to clear the alcohol,
it becomes more dense than the endolymph, sags to the bottom, and
creates a spinning sensation againnow in the opposite direction!1
Even when working well, balance is not perfect, tending to underestimate slow or steady self-motion for example. Fortunately, it doesnt
have to work alone; it collaborates with partners. One such partner is
vision, which helps by detecting optic flow, or consistent motion of the
visual scene as a whole (Figure 8.5). Such large-scale movements of the
visual scene are rarely caused by physical movements of objects in the
world and are therefore interpreted as self-motion. Wide-screen movies
at IMAX theaters or amusement parks take advantage of this effect to
make you feel like you are moving. These movies provide an image big
enough to cover most of the retina. So when the whole projected scene
appears to move, observers feel the motion in themselves, triggering a
balancing reflex and causing you to sway this way and that.
We also monitor the movements weve made. Suppose I were to blind-
184
M A K I N G SPAC E
fold you and take you along a route with a few twists and turns. How
accurately you could then return to your starting point would depend on
whether you had been allowed to walk the outbound route yourself or
whether I had pushed you along in a wheelchair.2 If we only used our senses
of balance and vision to monitor our progress through the environment,
there should be no difference between walking versus being wheeled.
Our greater accuracy when we move actively indicates the importance of
monitoring our own physical actions to keep track of where we arejust
as we saw in Chapter 7 for eye movements and vision.
Clever proof of the idea that keeping track of ones own movements
can be used to aid navigation comes from ants in the Tunisian desert
(Plate 9). These ants travel far afield in search of food to bring back to
the nest. Ants are good at finding their way back home again, and they
seem to count their steps to help do so. This was shown in an ingenious
experiment involving putting ants on stilts. Scientists captured ants just
after they had found a food source and were about to set off for home.
Before they were released again, some ants had their legs glued to pigs
GOING PL ACES
185
Food
Return
Outbound
FIGURE 8.6 An ant might leave its nest and search for food.
10 m
Nest
bristles to extend them. Others had their legs clipped to be shorter than
before. The ants were then released and followed to see whether they
would still be able to find their way home. The ants on stilts overshot,
while the stumpy-legged ones undershotjust what one would expect
if the ants were somehow counting their steps and were unaware of
their changed stride length.3
As our brains combine all these different sources of information, they
lay the groundwork for a sense of place that transcends the specific
details. Whether I drive or walk to the coffee shop, whether I approach
from the north or the south, I have a sense of the coffee shop as being
at a consistent position. This generalized sense of location enables us
186
M A K I N G SPAC E
to plan new routes between places, shortcuts we may not have taken
before. Again, imagine the alternative: suppose after completing a set
of errands to the grocery store, the pharmacy, the hardware store, and
the gas station, you had to return to each of these locations in reverse
sequence to get home again. Instead, typically you will return home
following the most direct available route once you have completed your
tasks. This ability to accurately update your sense of position, allowing
the planning of new routes and shortcuts, is called dead reckoning.4
Organisms that forage at some distance from home territory (or
animals storing food at various locations), must find their way repeatedly to the same spot, and many show this same dead-reckoning ability.
Think of the expressions as the crow flies or make a bee line. Even
our friends the desert ants seem to possess this skill: their outbound
foraging forays zig and zag all about. But when they return to their
nests, they take a short, straight path (Figure 8.6).5
As weve seen in this chapter and the preceding one, our sense of space
often relies on remembered informationthe view from previous fixations, the number and direction of steps weve traveled. In Chapter 9,
well look at this relationship more closely. As well see, many of the
brain mechanisms for space and memory overlap with each other
forming a two-way street in which memory aids our sense of space, and
our sense of space helps us remember.*
* The photo in Figure 8.1 was taken from the Provincetown Public Library,
Provincetown, Massachusetts.
GOING PL ACES
187
190
M A K I N G SPAC E
retina. The memory is revealed by what happens next, when the visual
stimulus turns off. The neurons keep respondingas if the visual stimulus were still there. They continue firing action potentials right up
until the animal moves its eyes. So the activity of these neurons seems
to help the brain store information about the visual stimuluss location
until the eye movement is made. This sustained activity appears to be a
neural mechanism for a very short-term but essential form of memory.
Different types and time scales of information involve different
neural mechanisms. Remembering the location of a brief visual stimulus is quite unlike learning to play a musical instrument (a physical
skill), recalling your first day of school (a personal experience), or
knowing the capital of Peru (a fact). A variety of brain storage methods
are involved in these assorted types of memory.
For example, learning physical skills, such as to how to play the
piano, is known as procedural learning. Neuroscientists investigating
the neural changes associated with procedural learning have generally
found that the changes occur within the neural pathways responsible
for performing the behavior and not in a separate area. Scientists often
try to study general principles such as this by picking a simple system
that can be extensively investigated in the laboratory setting. A popular
model system for this kind of skill learning is the vestibulo-ocular reflex.
The vestibulo-ocular reflex is a fancy name for our ability to keep
our eyes steadily looking at some particular point despite movements of
our head. Your eye muscles compensate for head motion, shifting your
eyes an equal and opposite amount. This keeps the image of the visual
scene quite stable on the retina as we bounce about. The vestibulo part
of the name reflects the role of your sense of balance, or more precisely,
the sense of head orientation and movement provided by the vestibular
S PAC E A N D M E M O R Y
191
organs in your inner ear. The ocular part indicates that the eyes are
involved. And the reflex part reflects that moving the eyes in this way is
fast and automatic. You cant not do this. (Have someone watch or film
your eyes while you turn your head. No matter how hard you try not
to, your eyes will counteract the motion of your head.)
The brain has to learn how far you should move your eyes in the opposite direction to the head movement. It must also dynamically adjust this
reflex to take into account viewing distance because the eyes and head
rotate about different axes. Furthermore, the brain recalibrates whenever
the previously learned gain (ratio of eye movement to head movement)
fails to keep the visual scene stablesomething that happens whenever
you change eyeglasses! Glasses alter the size of the visual scene, enlarging
it for the farsighted and shrinking it for the nearsightedessentially
adding or subtracting a bit of light-bending ability to help the eye focus
properly (Figure 9.1). This rescaling is an essential part of what glasses do:
they add or subtract a bit of light bending when the eyes own lightbending abilities are insufficient to focus an image precisely on the retina.
FIGURE 9.1Eyeglasses to correct near-
192
M A K I N G SPAC E
S PAC E A N D M E M O R Y
193
how many fingers you are holding up in their left visual field, they will
probably be able to tell you. But if you simultaneously hold up fingers in
both their left and right visual fields, they might only be able to tell you
about the fingers on the right side. The stimuli on the right side outcompete those on the left. This type of deficit is called neglect. When only one
side is damaged, the deficit is referred to as hemineglect. The syndrome is
most commonly observed for lesions of the right parietal cortex.
The deficit manifests in other ways as well. For example, when
patients copy or draw pictures of familiar objects, like flowers, the
spatial layout is distorted so that the important features (petals or stems)
are only copied if they are in areas of space that are handled by neurons
that are undamaged (Figure 9.2).3 Sometimes all the details may be
present, but they are squashed into the good side. This suggests that
the patients brains have access to those details, but they somehow cant
properly relate them to spatial locations.
194
M A K I N G SPAC E
Cathedral
A similar odd deficit involves remembered items and their spatial reference frame. In 1978, two Italian researchers, Edoardo Bisiach and Claudio
Luzzatti, tested how patients with damage to one hemisphere of the parietal cortex recalled familiar settings.4 They instructed their subjects to
imagine themselves standing at the foot of the famous Piazza del Duomo,
in Milan, and to name as many landmarks around the square as they
could. The patients had trouble naming landmarks from the side of the
square that corresponded to their impaired region of space (Figure 9.3).
But heres the fascinating part. Bisiach and Luzzatti then asked their
subjects to switch mental perspectives and imagine themselves standing
at the other end of the square, facing the opposite direction. The change in
S PAC E A N D M E M O R Y
195
vantage point, even though only imagined, allowed their patients to recall
landmarks from the opposite side of the square! So the memories were
there all along, and the problem was one of accesslike documents safely
stored in a locked file cabinet, but for which the key has gone missing.
The parietal cortex seems to play a role in unlocking this proverbial file
cabinet, and it seems to do so by calling to mind a spatial representation in
a reference frame based on ones imagined perspective on the scene.
Perhaps the most profound (and well-studied) connection between
space and memory involves the hippocampus, a seahorse-shaped cortical area folded beneath the rest of the cortex (Figure 9.4). That injury
to the hippocampus could cause profound memory problems was first
noticed in the 1950s due to the unfortunate case of a patient named
Henry Molaison. Molaison had suffered from epilepsy after injuring his
head during a bicycle accident at age seven. As a young adult in 1953,
he underwent brain surgery to remove the seizure focus (the place in
the brain where the seizures began). This entailed excising the medial
temporal lobe on both sides of the brain, including about two-thirds of
the hippocampus, as well as some surrounding brain structures.
FIGURE 9.4The hippocampus, a seahorse-shaped
196
M A K I N G SPAC E
Hippocampus
S PAC E A N D M E M O R Y
197
fashion. Also, the activity of neurons in the normal, intact hippocampus could be studied using recording electrodes, even while the rats
were performing these tasks.
In 1971, two scientists in London made an intriguing discovery. John
OKeefe and Jonathan Dostrovsky found that neurons in the hippocampus of rats were sensitive to the rats position in the laboratory apparatus.7 They placed recording electrodes in the hippocampi of rats who
were free to move around a small enclosure. When the rat moved from
one place to another, different neurons became active. These hippocampal neurons were behaving the way visual neurons do, responding
when something was within the receptive field, except the thing in
the receptive field was the rat itself rather than an external visual stimulus. They called this phenomenon a place field.
Later work by OKeefe, Lynn Nadel, and many others confirmed
that the place fields seemed to reflect the rats knowledge of where it
was, not particular stimuli the rat was experiencing from that position.8
It didnt matter what the location smelled like, or what visual stimuli
the rat could see from a particular spot, unless those sensory cues also
helped the rat know where he was.
In one key test, Robert Muller and John Kubie in New York showed
that while prominent visual landmarks influence the place fields,
removal of such landmarks doesnt abolish them (Plate 10a).9 Muller
and Kubie placed rats in a round tub, uniform and circular, surrounded
by dark curtains with no features on them except for a single, large,
white rectangular sheet of paper attached to the curtains. Nothing else
in the scene except this white cue card landmark provided any sense
of orientation. They then mapped hippocampal neurons preferred
locations while the rat wandered around the tub. Next, they took the
rat out of the tub, either moved or removed the cue card, brought the
198
M A K I N G SPAC E
rat back, and remeasured the place fields. When the landmark was
shifted to a new spot, the place fields shifted accordingly. And yet, this
could not have been a simple visual response, because the neurons still
exhibited place fields when the cue card was removed altogetherbut
the location of the place field was often altered, as if the rats sense of
direction was off. This all makes sense if what is happening is that the
hippocampal place fields reflect where the rat feels himself to be.
Subsequent studies expanded the portfolio of location-related signals
in the hippocampal complex (the hippocampus itself and nearby parahippocampal gyrus, which contains cortical regions interconnected
with the hippocampus). One such response pattern involves head direction sensitivity (Plate 10b).10 Neurons sensitive to head direction respond
when the rat is facing or moving in a particular direction, regardless of
starting position. These signals can be thought of as a kind of internal
compass. Another intriguing response pattern involves so-called grid
cells, or neurons that are responsive in a regular grid-like repeating
pattern across space, like the holes of a game of Chinese checkers (Plate
10c).11 Such a stepping-stone response pattern might play a role in keeping track of how far the rat has moved.
The discovery that the hippocampal region has two major attributesprofound sensitivity to ones own spatial location and a vital
role in forming memoriesis intriguing. Why would sensitivity to
space and a role in memory be combined in one structure? Perhaps the
answer is that space and memory are inextricably linked. As I talked
about earlier in this chapter, much of what we remember has a spatial
component to itremembering where we parked our car in the parking
lot or how to drive home in the evening. Furthermore, memories that
are not themselves spatial may still be indexed by the spatial location that
the original event occurred in. When you go to a college reunion, being
S PAC E A N D M E M O R Y
199
M A K I N G SPAC E
S PAC E A N D M E M O R Y
201
10
204
M A K I N G SPAC E
areas that show some kind of sensory or motor activity dominate the
illustration, and the territory left over after all the sensory and motor
areas are colored in is rather paltry.
With so much of the brain apparently involved in sensation and
movement, little brain space remains to accomplish the myriad of other
cognitive functions we possess. If we were to make a map of areas of
the brain that have been shown to play a role in other cognitive functions, such as memory, attention, planning and deciding, wed find
that it would superimpose on top of the sensory/motor map described
above, something like Plate 12.
Furthermore, areas that have only attention or memory signals,
say, and not sensory or motor signals have not been found. To be sure,
there are brain regions for which the most pronounced consequence
of a lesion might be a deficit in attention (e.g., the parietal cortex) or
memory (e.g., the hippocampus), but neurons in these regions seem to
carry sensory and motor-related signals in addition to their attentional
and memory-related signals.
The implication of this overlap between cognition and sensory
and motor processing is that perhaps the operations of cognition are
implemented at least in part via sensory and motor structures. That is,
perhaps thinking also involves activating some subset of sensory and
motor pathways of the brain. For example, when you mentally picture
sitting on the couch in your living room, that thought might be implemented by partially activating the visual, tactile, auditory, olfactory,
and motor responses that would have occurred if you were actually
there. The theory that thought might involve simulating the activity
patterns in our sensory and motor areas of the brain is called grounded
or embodied cognition.2
205
206
M A K I N G SPAC E
207
They found that words that related to specific kinds of actions elicit
responses in populations of neurons thought to be involved in controlling
those actions. For example, words such as kick or walk altered the activity of neurons close to the foot region of the motor cortex, and words
such as lick or talk affected activity close to the mouth region. And in
another study, words related to color were found to affect activity in the
visual pathway.7 Such observations support the idea that when you think
about what a word means, your brain may engage in a partial re-creation
of sensory and motor activity related to that meaning.
Double duty between sensory-motor processing and thought makes
sense for highly concrete concepts, for which there is a clear relationship
between the thought and what you might physically experience when
engaged in the real thing. But, this theory has been ingeniously extended
to more abstract forms of thought as well. Cognitive scientist George
Lakoff and philosopher Mark Johnson have suggested that we utilize
sensory and motor-building blocks for abstract thinking via metaphor.8
We often think of metaphor as something used in literature to make
a story vivid and interesting. But tying the abstract to the concrete also
aids understanding. In fact, I have done this on purpose throughout
this book.9 Ive used analogies between eyewitnesses and sensory receptors (Chapter 7), between Sherlock Holmes and the brains deductions
regarding the locations of sounds (Chapter 5), between file cabinets and
memory (Chapter 9), and many others. These deliberate analogies may
have helped you create a mental picture of the unfamiliar and intangibleat least, I hope they have.
Lakoff and Johnson pointed out that metaphorical language that
connects the abstract to the concrete is ubiquitous and permeates
our everyday way of speaking, such as a hot topic or a big name.
208
M A K I N G SPAC E
Furthermore, such metaphorical connections tend to follow consistent patterns. For example, emotion is often tied to figurative language
involving temperature (a heated debate, cold anger, a lukewarm
reception). And, like the studies by Hauk and colleagues concerning
concrete body-related words, Krish Sathians team at Emory University
has shown activity changes in the somatosensory cortex when subjects
read sentences involving metaphorical uses of texture (She had a rough
day) versus sentences with the same literal meaning but no connection
to tactile experience (She had a bad day).10
Most intriguingly, metaphorical linkages in everyday language
often involve space. For example, time is often referred to using spatial
language (in the days ahead, now that the exam is behind us).11
Social ties are also often expressed spatially (a close friend, a distant
father), as is social rank (lower class, top executive). In music,
there is no physical connection between sound frequency and space,
and yet we refer to pitches as being high or low.
The nature of our spatial representations may in turn shape how
we reason in the abstract nonspatial domains that are linked to space.
Cognitive scientist Lera Boroditsky has argued that differences in how
cultures define space affect how its members mentally represent time.
For example, English speakers commonly use a reference frame relative to themselves to describe relatively near locationssuch as left,
right, in front, behind. In contrast, in an Australian aboriginal community, the Pormpuraaw, the cardinal directions (north, south, east,
west) are the preferred method for defining space. Boroditsky then
tested how American English speakers and Pormpuraawans physically
arranged cards depicting events occurring in a particular temporal
order.12 The Americans tended to organize the cards from left to right
209
210
M A K I N G SPAC E
211
212
M A K I N G SPAC E
213
with respect to the eyes. That means the brain has to be capable of
linking every single location in one domain to every single location
in another. Furthermore, in any specific context it has to choose just
one of these linkages (based on the position of the eyes with respect to
the head). Perhaps neural wetware akin to that used for implementing
these coordinate transformations of sensory stimuli underlies the rapid
and flexible linkages between concepts.
Evidence supporting the idea that the neural circuitry involved in
reference frames is also used for other purposes comes from the mathematical domain and the parietal cortex. Recall that the parietal cortex
is thought to play a role in translating information across reference
frames (for example, patients with parietal lesions selectively recalled
landmarks depending on their vantage point, as discussed in Chapter
9). But the parietal cortex also appears to be important for mathematical thinkinga highly combinatoric process. In humans, lesions of
the parietal cortex produce deficits in mathematical ability (a condition
known as acalculia). Underlying both may be parietal neurons sensitive
to both attributes. Parietal neurons with receptive fields anchored to
both the head and eyes have been identified, as have neurons sensitive
to the number of items being presented.14 Thus, information about the
spatial locations of stimuli and the number of stimuli are both contained
in the firing patterns of neurons in the same brain region. Perhaps it is
no coincidence given that mathematics is a highly combinatoric form of
cognition in which numbers may be combined in infinite ways.
There is an important caveat to all this. The degree of overlap between
sensory and motor signaling and cognitive function in the brain implies
that not all of what appears sensory or motor is necessarily involved in
214
M A K I N G SPAC E
perceiving or moving. If these signals truly did both (at the same time),
then it would seem to be impossible to both think and monitor your
sensory scene at the same time. Perhaps this is why we like to do intense
thinking in quiet places like libraries, where the peaceful sensory scene
provides minimal disruption. But we are certainly capable of thinking
despite simultaneously sensing and moving.
Which sensory-responsive areas, say, are involved in perceiving and
which might be involved in using sensory-like signals for thinking is
completely unknown at this point. Determining which neurons are
actually for sensing and which are for thinking requires experiments that manipulate neural activity. For example, in Chapter 4, we
discussed experiments involving electrical stimulation in the visual
motion area, MT. Those experiments tested the effects of artificially
activating MT neurons on perceptual and motor tasks. In humans,
this general area shows activity not only when participants view real
moving stimuli, but also when they read (stationary!) sentences that
concern motion.15 Might stimulation in MT also then affect thinking
about visual motion, or mentally using visual motion in a metaphorical way? In this particular example, we dont know, but other studies
have found that temporarily interrupting activity in sensory and motor
structures using a technique called transcranial magnetic stimulation can
impair performance on cognitive tasks such as the mental rotation task
described above.16
The dividing line between perception and thought is not likely to be
clean. When we dream, we experience a variety of sensory perceptions
not triggered by real physical events but which can seem very real at the
time. In mental disorders such as schizophrenia, patients suffer from
hallucinations, erroneous perceptions that intrude on their thought
215
216
M A K I N G SPAC E
217
NOTES
1. For some other possibilities, see E.A. Lumpkin and M. J. Caterina, Mechanisms of
sensory transduction in the skin, Nature 445 (2007): 858865.
1. C.W. Oyster, E.S. Takahashi, and D.C. Hurst, Density, soma size, and regional
distribution of rabbit retinal ganglion cells, Journal of Neuroscience 1 (1981):
13311346.
2. D.H. Hubel and T.N. Wiesel, Receptive fields of single neurones in the cats striate
cortex, Journal of Physiology 148 (1959): 574591.
3. H. Zhou, H. S. Friedman, and R. Von Der Heydt, Coding of border ownership in
monkey visual cortex, Journal of Neuroscience 20 (2000): 65946611.
4. W. Penfield and T. Rasmussen, The Cerebral Cortex of Man: A Clinical Study of
Localization of Function (New York: MacMillan, 1950).
5. V.S. Ramachandran and S.Blakeslee, Phantoms in the Brain: Probing the Mysteries of the
Human Mind (New York: William Morrow and Company, 1998).
6. B.L. Sabatini and W.G. Regehr, Timing of synaptic transmission, Annual Review of
Physiology 61 (1999): 521542.
7. J.Zihl, D. von Cramon, and N. Mai, Selective disturbance of movement vision after
bilateral brain damage, Brain 106 (2) (1983): 313340.
8. F.A.Azevedo, L.R. Carvalho, L.T. Grinberg, J.M. Farfel, R.E. Ferretti, R.E. Leite,
W.Jacob Filho, R.Lent, and S.Herculano-Houzel, Equal numbers of neuronal and
nonneuronal cells make the human brain an isometrically scaled-up primate brain,
Journal of Comparative Neurology 513 (2009): 532541.
9. J.M. Groh, R.T. Born, and W.T. Newsome, How is a sensory map read out? Effects
of microstimulation in visual area MT on saccades and smooth pursuit eye movements, Journal of Neuroscience 17 (1997): 43124330.
10. C.D. Salzman, K.H. Britten, and W.T. Newsome, Cortical microstimulation influences perceptual judgements of motion direction, Nature 346 (6280) (1990): 174177;
C. M. Murasugi, C.D.Salzman, and W. T. Newsome, Microstimulation in visual
area MT: Effects of varying pulse amplitude and frequency, Journal of Neuroscience
13(4) (1993): 17191729.
Chapter 5: Sherlock Ears
1. Piano wires: D. R. Griffin and R. Galambos, The sensory basis of obstacle avoidance
by flying bats, Journal of Experimental Zoology 86 (1941): 481506. Meal worms: D.R.
Griffin, J.H. Friend, and F.A. Webster, Target discrimination by the echolocation of
bats, Journal of Experimental Zoology 158 (1965): 155168.
220
2. P.M. Hofman, J.G. Van Riswick, and A.J. Van Opstal, Relearning sound localization with new ears, Nature Neuroscience 1 (1998): 417421.
3. E.Knudsen and P.F.Knudsen, Visuomotor adaptation to displacing prisms by adult
and baby barn owls, Journal of Neuroscience 9 (1989): 32973305; J.F.Bergan, P.Ro,
D.Ro, and E.I. Knudsen, Hunting increases adaptive auditory map plasticity in
adult barn owls, Journal of Neuroscience 25 (2005): 98169820.
4. D. S. Pages and J. M. Groh, Looking at the ventriloquist: Visual outcome of eye
movements calibrates sound localization, PLOS One 8 (2013): e72562.
5. G.H. Recanzone, Rapidly induced auditory plasticity: the ventriloquism aftereffect, Proceedings of the National Academy of Sciences USA 95 (1998): 869875.
Chapter 6: Moving with Maps and Meters
221
U.Werner-Reiss, K.A. Kelly, A.S. Trause, A.M. Underhill, and J.M. Groh, Eye
position affects activity in primary auditory cortex of primates, Current Biology 13
(2003): 554562; D. A. Bulkin, J. M. Groh, Distribution of eye position information in
the monkey inferior colliculus, Journal of Neurophysiology 107 (2012): 785795.
7. J. Lee and J.M. Groh, Auditory signals evolve from hybrid- to eye-centered coordinates in the primate superior colliculus, Journal of Neurophysiology 108 (2012):
227242.
8. K.K. Porter, R.R. Metzger, and J.M. Groh, Representation of eye position in
primate inferior colliculus, Journal of Neurophysiology 95 (2006): 18261842; J.M. Groh
and D.L. Sparks, Two models for transforming auditory signals from head-centered
to eye-centered coordinates, Biological Cybernetics 67 (1992): 291302.
Chapter 8: Going Places
222
N OT E S TO PAG E S 198 2 11
223
13. R.Q.Quiroga, L.Reddy, G.Kreiman, C.Koch, and I.Fried, Invariant visual representation by single neurons in the human brain, Nature 435 (2005): 11021107.
14. J.D. Roitman, E.M. Brannon, and M.L. Platt, Monotonic coding of numerosity
in macaque lateral intraparietal area, PLOS Biology 5 (2007): e208; A.Nieder, D.J.
Freedman, and E.K. Miller, Representation of the quantity of visual items in the
primate prefrontal cortex, Science 297 (2002): 17081711; O.A.Mullette-Gillman,
Y.E. Cohen, and J.M. Groh, Eye-centered, head-centered, and complex coding of
visual and auditory targets in the intraparietal sulcus, Journal of Neurophysiology 94
(2005): 23312352.
15. E.Chen, P.Widick, and A.Chatterjee, Functional-anatomical organization of predicate metaphor processing, Brain and Language 107 (2008): 194202.
16. G.Ganis, J.P. Keenan, S.M.Kosslyn, and A.Pascual-Leone, Transcranial magnetic
stimulation of primary motor cortex affects mental rotation, Cerebral Cortex 10
(2000): 175180.
17. A. K. Fetterman, M. D. Robinson, R. D. Gordon, and A. J. Elliot, Anger as seeing red:
Perceptual sources of evidence, Social Psychological and Personality Science 2 (2011):
311316.
18. J.A. Bargh and I.Shalev, The substitutability of physical and social warmth in daily
life, Emotion 12 (2012): 154162.
224
CREDITS
226
CREDITS
CREDITS
227
228
CREDITS
CREDITS
229
ACKNOWLEDGMENTS
National Science Foundation, and private foundations funded the original research from my laboratory described herein. And Duke University
has provided financial and organizational support for teaching and
research activities synergistic with this book project.
As for the words on the page, Im most grateful to Suzanne Bolt, freelance editor extraordinaire. Suzanne rolled up her sleeves and read this
book very closely and thoroughly. Her insights into the structure of the
arguments and sequence of ideas were trenchant, and her wordsmithing
skills are superb. I learned an immense amount from her suggestions. If
you still find some sentences with too many prepositions, thats on me.
She tried her best.
Several colleagues served as my guides concerning theories of cognition. Cognitive linguist (and collaborator) Edna Andrews prompted me
to consider the relationship between spatial perception and language
and has been my conversation partner for a number of discussions that
have shaped my thinking. Auditory neuroscientist David Poeppel has
been unfailingly generous with his time. David provided me with pointers to a number of important findings and arguments, and he critiqued
several chapters of the book. He likely will not agree with everything
I have written, but I hope he finds it thought provoking. Cognitive
scientist Rafael Nunez was, unbeknown to him, one of the first people
who alerted me to the connection between space and thought, and Ive
enjoyed subsequently discussing these ideas with him.
The development of this book involved a number of stellar undergraduate, graduate, and postdoctoral students and technicians with
whom Ive had the pleasure of working, either in the classroom or
the laboratory. These students were instrumental in creating the
microcosm for discussing these concepts. Many of these students
232
ACKNOWLEDGMENTS
ACKNOWLEDGMENTS
233
234
ACKNOWLEDGMENTS
Two anonymous reviewers took time out of their busy lives and read
the entire manuscript all the way through. Their insightful comments
were particularly valuable and constructive.
Last but by no means least, my family (nuclear and extended)
deserves my fervent appreciation. Whether their support has come
in the form of commentary, interest, critique, or simply quiet time to
think and write, they have contributed to this book in ways that are of
immeasurable importance. I cannot thank them enough.
ACKNOWLEDGMENTS
235
INDEX
acalculia, 214
action potentials, 5659, 57, 60, 61, 7172, 74,
9496, 140, 146; electrically stimulating,
102; and memory, 191; in MT, 99; in
superior colliculus, 154. See also spikes,
electrical in neurons
active sites, of proteins, 17
adjusting to new glasses, 192193
akinetopsia, 100
alcohol, effects on vestibular system, 183184
Alhazen, 8, 1112
aliasing, of sound, 123125
analog coding, in the brain, 146147. See also
digital coding, in the brain; maps, as form
of brain code; meters, as form of brain
code
anechoic, 127
ants, and navigation, 185187, Plate 9
238
INDEX
INDEX
239
240
INDEX
INDEX
241
242
INDEX
INDEX
243
refraction, 3032, 31
relative size, as clue to distance, 4446
Renoir, Pierre-Auguste, 23, Plate 2
representation, 80, 143; brain maps as,
145146; brain meters as, 146147; in
electronics, 144146, 145146. See also
analog coding, in the brain; digital coding,
in the brain; maps, as form of brain code;
meters, as form of brain code
resonance, 139
resting membrane potential, 18, 19, 57, 58
retina, 3, 13, 22, 25, 29, 31, 33, 7374, 7779,
91, 111, 145, 150, 155, 162, 165, 169, 191;
and blind spot, 91, 91; and connection to
thalamus, 7779; and fovea, 89, 163; and
interpreting the inverted retinal image,
3435; movement of by smooth pursuit,
9899; and optic flow, 184; and reference
frames, 170171; size of retinal image
and distance, 4445, 192; stereovision
and 36, 37, 39; and visual maps, 69, 73.
See also center surround organization;
photoreceptors; retinal ganglion cells
retinal ganglion cells, 7778, 91, 93
retinotopic organization. See brain map;
retina; visual cortex
rhodopsin. See photopigment molecules
Rosenquist, Alan, 167
round window, of cochlea, 109
Rubin vase/profile illusion, 83
saccade, 66, 98, 149, 153155; and curved
trajectories, 66; to tactile stimuli, 169170.
See also eye movements, control of by
superior colliculus; eye movements,
feedback control of; eye movements, meter
coding of; eye movements, saccades
saccule, 180, 181182, 183
Sathian, Krish, 209
sclera, 31
244
INDEX
INDEX
245
246
INDEX
Waterfall (Escher), 47
Werner-Reiss, Uri, 172173