Psych 1XX3 Form Perception II Lecture Notes
Psych 1XX3 Form Perception II Lecture Notes
Feature Detectors:
The brain uses a division of labour; with each region along the visual pathway
processing relatively specific information and then passing it on.
Form recognition follows a similar strategy.
Lettvin et al:
For instance, in 1959, Lettvin and colleagues discovered a neuron in the optic
nerve of a frog that responded only to moving black dots, and they called these
cells "bug detectors".
Hubel and Wiesel:
Hubel and Wiesel spent years extending this work in their studies of cells in the
visual cortex of cats and monkeys, eventually earning the Nobel Prize in 1981.
Beginning in 1962, Hubel and Wiesel began their exploration of the visual cortex
by trying to learn what type of stimuli the individual conical cells responded to.
They did this by putting microelectrodes in the cortex of a cat to record the
electrical activity of individual neurons as the cat was shown different types of
visual stimuli, such as flashes of lights.
The problem was that they weren't getting much response from the neurons, until
one day when they presented the cat with a slide that had a crack in it. When the
line that was projected from that crack moved across the cat's visual field, the
neuron started to fire like crazy!
This was a light bulb moment for Hubel and Wiesel, who realized that neurons
must respond to stimuli that are more complex than diffuse flashes of light.
They began using lines of different orientations and thickness that moved in
different directions, and they found that each neuron is very specific about what
will make it fire the most. These cells fire maximally to stimuli of a certain shape,
size, position, and movement, and this defines the receptive field for that cell.
Simple Cells:
Defn of Simple Cell: Responds maximally to a bar of a certain length and
orientation in a particular region of the retina.
For example, this simple cell responds the most to a horizontal bar but if that
same bar is moved outside that particular region and/or changes orientation, then
the cell will be inhibited, and actually fire less than baseline. (See image below.)
Complex Cell:
Defn of Complex cell: Responds maximally to a bar of a certain length and
orientation, regardless of where the bar is located within the receptive field, but
unlike the simple cell, a complex cell does not care about where in its receptive
field the bar is located, and it will even continue to fire if the bar is moving within
the receptive field.
Some complex cells do care about the direction of this movement, such as the cell
in this figure that fires the most when the stimulus is oriented at a certain angle
and moving in a particular direction.
See image on next page.
Hypercomplex Cell:
Defn of Hypercomplex cell: Responds maximally to a bar of a particular
orientation that ends at specific points within the receptive field.
For example, this hypercomplex cell fires the most to a horizontal bar of light that
appears anywhere in the "on" region of the receptive field, but gives only a weak
response if the bar touches the "off" region.
So these cells have an inhibitory region at the end of the bar making them
sensitive to the length of the bar.
These three types of cells should give you some idea about how specifically tuned
our visual cortical cells can be.
Topographical Organization:
The layout of the visual scene is preserved in the visual cortex.
Neighbouring objects in your visual field are processed by neighbouring areas of
your brain, but this mapping from visual field to brain is not exact, because the
largest amount of cortex is devoted to processing information from the central
part of the visual field, which projects onto the fovea.
Nevertheless, each region of the cortex receives some input from a small piece of
the visual field, and within each region, there are cells that analyze specific
features of the scene.
For a particular part of the visual field, there are neurons that fire maximally if
there is something in the scene that has a line of a certain orientation, length, and
movement; other neurons respond maximally if there is something in that tiny
portion of the visual scene that is a specific colour; other neurons respond most
when there is a line that moves in a certain direction.
Cluster of cells in the region of the cortex right beside this region are doing the
same analysis for the neighbouring part of the visual scene.
An important benefit of this parallel processing strategy is speed. (See img below)
Ventral Stream:
Combining Information in the Extrastriate Cortex:
The processing of visual input in the primary visual cortex involves specific cells
responding to relatively specific features from a small portion of the visual field.
But for the visual scene to make any sense, this information has to be combined to
form a meaningful whole.
Subregions in Extrastriate:
This combination begins in the extrastriate cortex, also known as visual
association cortex, which surrounds the primary visual cortex.
The extrastriate cortex has multiple subregions that each receives a different type
of information from the primary visual cortex about the visual scene.
For example, one subregion of the extrastriate cortex will receive information
about the colours in the scene, another about any movement in the scene, and
another about different line orientations in the scene.
It is in the extrastriate cortex where the information begins to be segregated into
two streams according to the type of information that is processed.
One stream is the dorsal stream, also known as the "where" stream, which
processes where objects are located in the visual scene and how they are moving
within that scene.
The dorsal stream takes information from the primary visual cortex to the parietal
cortex, which processes spatial information.
The other stream is the ventral stream, also known as the "what" stream because
it processes information about what the object is, including form and colour.
The ventral stream takes information from the primary visual cortex and sends it
to the temporal cortex, where all the bits of feature information come together.
Infants between 4 and 5 months of age treated the identical bear that was viewed
at two different distances as familiar; but they stared much longer at the large hear
that was viewed from a greater distance.
This suggests that infants at this age had some sense of size constancy and
understood that an object that is farther away should produce a smaller retinal
image.
Naturally, size constancy is far from perfect at this age, and continues to develop
with improvements in perceiving depth, well into the school-age years.
Do we have an Innate Preference for Faces?
Some researchers have argued that even newborns prefer to look at faces over any
other pattern.
This innate preference has evolved to ensure that as infants, we orient toward
other people and not other objects in our environment. This could serve to build a
necessary social bond with our caregivers.
In these face preference studies, infants as young as 4 days old are shown
different patterns, colours, shapes, or even a scrambled face, and the infants prefer
looking at faces. Interestingly by 2 months, infants prefer to look at attractive
faces over unattractive faces, and what an infant considers to be an attractive face
actually coincides with what adults consider to be attractive.
Furthermore, 2-month old infants will look longer at their own mother's face than
faces of other people, and by 5 months, they can begin to detect different
emotional expressions, such as happiness or sadness.
All of these studies suggest that we are born with a readiness to perceive and
prefer face stimuli compared to other stimuli.
However, other studies have found that infants show no preference for faces over
other complex stimuli, and if there is a face preference, it emerges gradually from
all the early experience we have with faces.
The argument is that infants do not have a preference for faces per say, but really
have a preference for complex stimuli that have a lot of contrast between light and
dark, such as the eyes and mouth, as well as moving parts.
In fact, studies that used non-face stimuli that were matched to the face stimuli for
complexity showed that infants did not prefer the face stimuli.
Studies that tracked where an infant was looking when shown a picture of a face
revealed that infants under 2 months of age focused mainly on the outer contours
of the face, such as the hairline or chin.
It was not until the infant was over 2 months old that she looked at regions within
the face, like the eyes and mouth.
Preferential Looking Method: Right from birth, babies would rather look at
something that is patterned rather than something that is plain gray.
If we show babies a card with stripes on one side, and they choose to look to that
side on the card, we know that they can see the stripes.
Over trials, we make the stripes thinner and thinner. The thinnest stripes babies
prefer over grey provide a measure of their vision, just like the smallest letter that
you can read on an eye chart.
Using techniques such as this, we can figure out what the world looks like to a
baby. We now know that babies can see right from birth and that vision improves
rapidly over the first few months of life.
By 6 months of age, acuity is 5 times better than it was at birth. Both visual
experience and maturation of the eye and brain contribute to the rapid
development.
Prosopagnosia:
Another type of visual agnosia is prosopagnosia, which occurs when a person
can recognize regular objects but cannot recognize faces.
A person with prosopagnosia will know that they are looking at a face, and they
will be able to see eyes, a nose, and a mouth, but they won't be able to put those
individual features together and perceive whose face it is, even if they're looking
in a mirror.
These people have to rely on other cues to recognize other people, like their voice,
smell, or the way that they walk.
Curiously people with prosopagnosia can also have difficulty recognizing other
specific stimuli, but they can recognize categories of objects.
For example, they would be able to recognize a dog and a car but they would not
be able to pick out their own dog or their own car.