An Open Introduction to Linguistics 2022
An Open Introduction to Linguistics 2022
AN OPEN
INTRODUCTION TO
LINGUISTICS
USED BY PERMISSION
EDITED AND EXTENDED BY
JENNIFER SPENADER & HESSEL HAAGSMA
Copyright © 2022 Marcus Kracht & Jennifer Spenader
tufte-latex.googlecode.com
Licensed under the Apache License, Version 2.0 (the “License”); you may not use this file except in com-
pliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/
LICENSE-2.0. Unless required by applicable law or agreed to in writing, software distributed under the
License is distributed on an “as is” basis, without warranties or conditions of any kind,
either express or implied. See the License for the specific language governing permissions and limitations
under the License.
1 Introduction 5
2 Phonetics 11
20 Pragmatics 229
Bibliography 275
Index 275
1
Introduction
An introduction
ple tasks like setting an alarm, asking what the weather is like, ask-
ing Alexa to play the more recent news broadcast or to play a song.
What’s interesting about Alexa, google assistant or Siri, is that all
the tasks and programs that these smart speakers can access are also
available on our computer’s or phones, and we frequently and easily
access them. But it’s much more convenient to simply use your voice
to ask in natural language for the same information or task. It’s not
only easier, because you don’t have to find your phone, but it also feels
more natural, because speech is the more natural, and more common
way in which we ask humans to do things for us.
For an Alexa device to speak and understand our commands re-
quired years of computational linguistic research, which built and still
builds on our understanding of human language and how language is
used as a means of communication. While many natural language pro-
cessing tasks (NLP) like speech recognition and natural language un-
Alexa and google home, smart speak-
derstanding (NLU) and natural language generation (NLG) are now
ers.
frequently dealt with using machine learning methods that rely on
large data sets, often these data sets require training and developmen-
tal data that first required linguistic processing or annotation. This
processing relies on an understanding of fundamental linguistic con-
cepts, based on decades of research. One of the goals of this book and
the course that accompanies it, is to give you a solid, if brief, intro-
duction to these linguistic concepts and the methods used in linguistic
analysis. Often linguistic analysis deals with recognizing patterns, and
uncovering hidden representations. Many linguistic analysis tasks are
really like puzzles. Along the way you will be surprised to discover you
already unconsciously know very complex things about language and
language structure. In the following sections I will give a brief sum-
mary of the major topics covered in this book, noting which topics are
most relevant and important to artificial intelligence. Keep in mind
that artificial intelligence research has two sides: on the one hand we
are trying to understand human intelligence, in order to be able to
model or recreate it. To this end, we study humans, and human devel-
opment and thought processes, and we use computational tools to do
this. On the other hand, we want to create intelligent systems that can
interact with humans in a natural, fluent way. In both these aspects
there is a need for understanding the structure and communicative
functions of language.
Phonetics
Phonology
Semantics
Pragmatics
Language Acquisition
How do we learn to speak our native language? Is this different from
how we learn other skills or knowledge, or how we would learn to
speak a new, second language, as an adult? The answer to both these
questions seems to be yes. Language is different from other cognitive
skills, and much research has shown that we possess preprogrammed
language structures as part of our genetic endowment, allowing us to
quickly learn our native language when we are exposed to it as children.
Language acquisition research helps us better understand human
intelligence and human learning. For this reason, understanding the
basics of how language is acquired is important. Further, we can often
better understand the final, adult state of human language knowledge
if we examine how that state was reached. Like other psycholinguis-
tic research, theories of child language acquisition have benefited from
more advanced statistical modelling and computational models of lan-
guage learning. We’ll touch briefly on that topic.
Machine Translation
Machine translation is the task of translating from as source language
to a target language automatically with a computer program. The
history of machine translation research is actually an interesting topic
in itself, because initial efforts to use computers to analyze text were
all focused on the task of translation. It still remains one of the major
research areas in computational linguistics today. It’s also particularly
interesting because it involves all levels of human language. For speech-
to-speech translation, both speech recognition and speech generation
is necessary. Depending on the structure of the language and the
translation technique used, pre-processing steps such as segmenting
the input into morphemes, disambiguating homographs, dealing with
issues such as dropped subjects, etc. are all tasks that require linguistic
analysis.
Topic identification
Automatic topic identification requires analyzing texts to determine
what the dominant topic is automatically. This is useful for all kinds
of applications, including twitter misinformation detection. We’ll talk
a bit about this in the last lecture.
10 an open introduction to linguistics
Most speech sounds are pulmonic egressives: that means the sound
is created by using a stream of air coming from the lungs.
We inhale air, and then we release the air from the lungs in short
puffs. We are unconsciously able to control this exhaling using the
muscles of our rib cage, maintaining even air pressure. The air from
the lungs travels first through the trachea into the larynx which houses
the vocal cords. The larynx is also called the voice box. In men,
the position of the larynx is visible in the neck. It’s therefore colloqui-
ally named the Adam’s apple. The vocal cords are two small flaps
made out of tissue that can be completely open, completely closed, or
The vocal cords when open. Air
partially closed. The gap between the vocal cords also has a name: flows freely throught the glottis.
the glottis. When partially closed the vocal cords create vibrations
which become sound.
The larynx is then considered the first point of phonation, the be-
ginning of speech sounds. Air then travels through the trachea to the
pharynx. The pharynx is a tube-like chamber that connects to the
mouth and nasal passages from the trachea, if you think of the flow
of air from the lungs. But if you follow the pharynx from the mouth
cavity down towards the lungs, it actually splits into two passages: the
larynx and trachea (which goes to the lungs) and the esophagus (which
goes to the digestive tract). In order to prevent food from entering the
esophagus by mistake, there is a flap of tissue called the epiglottis
that covers the esophagus. The pharynx also has resonating proper- The vocal cords are closed, only a
small stream of air flows through the
ties, and it’s shape can be manipulated slightly by the position of the
opening.
back of the tongue, allowing it to also participate in phonation.
The velum determines whether or not air enters the nasal cavity.
If closed, air will only flow through the mouth. Once a flow of air
is present, it can be manipulated to produce different sounds. Most
of this manipulation happens in the oral cavity (the inside of your
mouth), where the tongue can be in different positions. The oral cavity
can also be opened to different degrees and the shape of the lips can
be varied. (Not so) fun fact:
Above we described sound production using an air stream expelled It is actually an unfortunate design
of the human body that the phar-
from the lungs. Sounds can also be produced with air entering the ynx leads to both the lungs and the
mouth going to the lungs but it’s less natural. These are called pul- stomach, because this means eating
and breathing are done via the same
monic ingressives. Pulmonic ingressives are generally only used par- passage way. Because the epiglottis,
alinguistically, that is they are used to convey things like attention, which is supposed to protect the air-
emotion or agreement in a conversation. For example, both in Swedish way from food entering, sometimes
fails in this function, food can get
and Dutch, pulmonic ingressives have been used to signal that one is stuck in the trachea, which leads to
listening (or agreeing) in a conversation, similar to the English use of airflow being cut off i.e. choking.
“uh huh”.
Almost everyone is familiar with the distinction between consonants
and vowels: the articulatory difference has to do with the degree of ob-
phonetics 13
Consonants
problem was recognized early on, and lead to the development of IPA,
the International Phonetic Alphabet. IPA was created in 1888 by the
International Phonetics Association. It’s based on the Latin alphabet,
but has been designed to make it possible to record in writing a good
approximation to all speech sounds in languages of the world. At the
end of this chapter you can find the full IPA chart from (2005) (the
most recent version), but first we’ll zoom in on the different sections
separately. Let’s first examine the part that relates to pulmonary
consonants: How to produce click consonants:
(Diagrams: JKS)
not as profound, but IPA still makes it possible to code them with
diacritic markings. See in the table below how more dental or apical
sounds can be coded.
A given meaningful sound in a language is called a phoneme. We’ll
talk in much more detail about phonemes in the next chapter but
briefly: phonemes are the sounds that are meaningful in language.
Alveolar voiceless stops are meaningful in English. It could be that in
rapid speech, some English alveolar voiceless stops will be produced
as if they were laminal denti-alveolar stops. But English hearers will
classify both sounds as the same meaningful sound or phoneme. And
in fact, the context in which a speech sound is produced will influence
how close the production comes to its ‘ideal target’. For example,
consider the word teetotaler (someone who does not drink alcohol).
This would include three voiceless alveolar plosives ([t]). But if you
pay attention as you say it, you will notice that whereas the first two
stops are pronounced at the front of the alveolar ridge, the third is
pronounced more palatal, more to the back. The reason is the context.
The first vowel, teetotaler is a front vowel ([i]), but a diphthong that
moves to a back vowel ([@U] and then to schwa). These diacritics also
come in handy for giving a more detailed (or narrow) transcription
of sounds in context.
Place and Height The form of the vocal tract affects the sound
that air flowing from the glottis through the mouth makes. We modify
this form mostly by modifying the placement of the tongue, limiting or
permitting airflow, narrower openings in different positions. The places
where airflow can be partial obstructed can be at the front, back or
even a central position in the mouth. Further, we can have the main
body of the tongue either closer to the roof of the mouth, creating a
American English Vowels
more closed cavity, or we can have it more open, by having it flatter
American English has fewer vowels
in the mouth. Sometimes, the same feature is described from the than British RP English.
perspective of the tongue, and some researchers instead talk of high
vowels, synonymous with closed vowels, and low vowels, which are
synonymous with open vowels. Tongue height is a horizontal feature,
and we can think of how closed or open the inside of the mouth is as
a vertical feature. The IPA charts for vowels intuitively captures this.
You can imagine a sagittal section with the mouth opening to the left
superimposed on the chart, to understand how the chart relates to the
actual moth (see margin picture).
So for English, the sound in beet and feet is the most closed front Japanese vowels Japan has a vowel
vowel (or high front vowel), while the sound in the words bat and cat system with only five vowels. Accord-
ing to WALS, in the 564 languages
is the most open front vowel (or low front vowel). There are several surveyed, 287 languages have a vowel
vowels in between (see the margin note) so you can actually say a inventory between 5-6 vowels.
sequence of words with only front vowels, and notice how with each
vowel the mouth opens a bit more, e.g. Say beet-bit-bait-bet-bat. You
can also notice the difference between front and back vowels by saying
the English words in the following sequence: Say bet-but-bought.
Vowel inventories
Each language has its own vowel inventory, and these can differ enor-
mously. Among European Languages, Danish is the front-runner in Daniel Jones Inventor of the cardinal
vowels.
monophthongs with 32 vowels, followed by Norwegian with 19. French,
German and Swedish have 17 and Dutch, Finnish and Icelandic have
16. There are many languages that have a small vowel inventory. In
particular, languages of the indigenous peoples of north American, and
the indigenous languages of Australia have small vowel inventories of
less than 5 vowels. In fact, the Aymara, Cherokee and Haida have only
three vowels.
Even among vowels that IPA would classify as the same vowel, the
actual vowel in a given language may differ in exactly how it is pro-
duced. We can talk about a language having a high front vowel, but
that high front vowel might be more or less high in one language com-
pared to another. These differences are signalled by the way in which
the vowel inventory for different languages is positioned in an IPA style
vowel diagram. Consider the charts for American English, Japanese Cardinal vowels 17 of Daniel Jones’
cardinal vowels, without the mid vow-
els.
phonetics 21
the symbols show the variation. This shows clearly how production of
vowels aims at a target, but that that target is not always met. We
see that for the top three vowels the individual speakers do not overlap
with themselves in their productions, but that there is some overlap in
the lower two vowels. Still, even if vowel productions are off, hearers
can use the context to determine what vowel was intended.
Extra Material
It is useful to have a look at the active sound chart at http://web.
uvic.ca/ling/resources/ipa/charts/IPAlab/IPAlab.htm. You can
go there and click at symbols to hear what the corresponding sound is.
Another very useful source is the Wikipedia entry for the IPA (http://
en.wikipedia.org/wiki/International_Phonetic_Alphabet) and the
entries for individual sounds, e.g. http://en.wikipedia.org/wiki/
Voiced_Velar_Fricative.
Study Questions
1. In your own words, describe the difference between phonetics and
phonology. Do you think there is a clear divide between the two?
5. Write the IPA symbol for the following speech sounds: You can find the complete
IPA chart here: http://www.
a. voiceless velar fricative internationalphoneticassociation.
org/sites/default/files/IPA2005_
b. glottal stop 3000px.png .
a. phone
b. theatre
c. ceiling
Persian (Farsi) Vowel Chart.
d. very
e. complete.
Meaningful Distinctions
The last chapter introduced phonetics. Phonetics is the study of the
speech sounds that can be produced and perceived by humans. This
chapter introduces phonology, the study of meaningful sound systems.
to different objects in the world, so clearly the use of the [b] rather than
the [m] is meaningful. There are many pairs of words that contrast
[b] with other sounds, e.g. ball-call, book-look, bed-head, etc. These
contrasts are all evidence that [b] is meaningful in English. Pairs of
words (with distinct meanings) which differ only in the pronunciation
of one sound have a special name in linguistics. We call these minimal
pairs. We can use minimal pairs to identify the meaningful sounds
in a given language. These language specific meaningful sounds are
termed phonemes of that language.
Systematically examining minimal pairs to determine which speech
sounds are meaningful in a given language helps us determine the
28 an open introduction to linguistics
.
Categorical perception allows us to perceive phonemes as distinct
categories, despite the continuous nature of sound. English has about
40 distinct phonemes, depending on the dialect. Each phoneme con-
trasts with many of the other 39 phonemes, but not necessarily all.
Consider the following data from Burushaski, a language spoken in
Srinagar, India. In this language there are phonemic voicing contrasts
in plosives word initially, but there are not clear minimal pairs for all
cases. For example, the word [tam] means swimming, and the word
[dan] means stone. These do not qualify as minimal pairs, but because
phonology i: features and phonemes 31
Word Broad Narrow Word Broad Narrow Table 3.1: Different allophones of /p/
in English
pat /pat/ [ph a:t] peace /pis/ [ph i:s]
prince /prIns/ [prIns] play /pleI/ [pleI]
spin /spIn/ [spIn] pin /pIn/ [ph In]
happy /hæpi/ [hæpi] tap /tæp/ [tæp]
stamp /stæmp/ [stæmp] pump /p@mp/ [ph @mp]
The above table shows the words in standard English spelling, fol-
lowed by two transcriptions in IPA, broad transcription and narrow
transcription. Narrow transcription gives all the details of the pro-
nunciation. Broad transcription is a phonemic transcription: in other
words, the broad transcription just shows the phonemes in English.
An English speaker reading the broad transcription will pronounce
the words like the narrow transcription, but they will (without train-
ing) not be aware of the features of their pronunciation that are shown
in the narrow transcription. Most English speakers are not aware that
they aspirate word initial [p]’s. That’s because the contrast between [p]
and [ph ] is not phonemic. There are no minimal pairs in English that
contrast them. However, they are clearly related in articulation. Both
are bilabial voiceless plosives, and, as the example shows, the context
predicts where each sound occurs. None of the contexts overlap, so this
means that the two sounds are in complementary distribution.
Sounds that are in complementary distribution, are allophones.
Allophones belong to a given phoneme. The sounds [p] and [ph ] are
both allophones of the phoneme /p/. Note that we talk about both
sounds being allophones of /p/ (a phoneme, so phonemic transcrip-
tion) and not allophones of /ph /. That’s because if we look at the
32 an open introduction to linguistics
or word finally?
(2) Do the sounds occur before or after vowels? Which vowels? Do
they have features in common? (front, back, low, high etc.).
(3) Do the sounds differ in distribution of nasals or other consonants
next to them? If you make a list of the contexts in which each of
the three sounds occurs, you should usually be able to identify a clear
pattern.
In Maasai, [p] occurs word initially, word finally, and always before
or after a vowel. In contrast, [b] occurs only after a nasal. Finally,
in the example given, [B] only occurs between two vowels. All three
sounds are thus in complementary distribution. This means also that
they are all allophones of the same phoneme. Speakers of Maasai
perceive all three sounds to be the same sound. In the case of Maasai,
it’s a bit harder to determine how to represent the allophonic set: what
symbol should we use for the phoneme? There isn’t really evidence to
prefer one over the other in terms of the number of contexts they
occur in, so we can perhaps choose [p] to represent the set because
word initial sounds seem more primary than word internal sounds, but This joke rests on the fact that in
this is fairly arbitrary. Then we can say that /p/ is a phoneme in General American, the central vowel
Maasai that is made up of three allophones: [p],[b] and [B]. [2] is often reduced to a schwa, e.g.
[@]. However, not always. So in fast
Now we’ve covered phonemic contrasts and allophonic variation. speech, photograph will be pronounced
There is only one more important type of variation possible when as [foUt@græf], but in single syllable
words it cannot be reduced, so we do
comparing two sounds within a language: free variation. It’s possible say cut [c2t] and luck [l2k].
that two sounds can be used interchangeably. Speakers are aware of the
differences, but the distinction is not meaningful (there are no minimal
pairs. This often is found between two dialects of the same language,
but also within an individual speaker. The aunt pronunciation in
English is an example of this. While each of the two pronunciations is
dominant in different dialects in English, some speakers may freely use
both. Note that technically, sounds in free variation are also allophones
of the same phoneme. They are just not in complementary distribution
with each other.
There’s a decision tree you can use to identify whether two sounds
you are analyzing in a language are different phonemes, allophones of
the same phoneme, or if they are in free variation. See the chart below:
phonology i: features and phonemes 35
The range of values for place is obviously different from that of man-
ner, since ‘dental’ is a value of the former and not of the latter. A set
of features is called an attribute value structure (AVS). You have
already seen AVSs in the first chapter. The notation is as follows. The
attributes and values are arranged vertically, the rows just having the
attribute paired with its value, separated by a colon:
place : dental
manner : fricative (3.3)
voice : +
Notice that the following are also legitimate AVSs:
place : dental
place : dental place : dental
place : uvular (3.4)
manner : fricative
voice : +
voice : +
Note that the fact that you can still
The first is identical to (3.3) in the sense that it specifies the same distinguish between [p] and [b] while
whispering shows you that the distinc-
object (the features are read conjunctively). The second however does tion is not exclusively a voicing con-
not specify any sound, since the values given to the same feature are in- trast! There are additional differences,
one of which is that the lip tension is
compatible. (Features must have one and only one value.) We say that higher in [p].
the second AVS is inconsistent. Notice that AVSs are not sounds,
they are just representations thereof, and they may specify the sounds
only partly. I add here that some combinations may be formally con-
sistent and yet cannot be instantiated. Here is an example:
" #
consonantal : −
(3.5)
voice :−
This is because vowels in English are voiced. There are a few languages,
for example Mokilese, which have voiceless vowels. To understand
how this is possible think about whispering. Whispering is speaking
without the vocal chords vibrating. In effect, whispering is the same
as systematically devoicing every sound.
Not everything goes for AVSs, however. The following AVS, for
example, is illegitimate because it assigns a value to place that is
outside of its value range.
" #
place : fricative
(3.6)
voice : +
phonology i: features and phonemes 37
There are a number of arguments for why features exist. First and
foremost the features encode a certain linguistic reality; the features
that we have spoken about so far have phonetic content. They speak
about articulatory properties. It so happens that many rules can be
motivated from the fact that the vocal tract has certain properties.
For example, in German all syllable-final plosives are voiceless (see the
discussion on Page 53). This is the case even when there is reason
to believe that the consonant in question has been obtained from a
voiced consonant. Thus, one proposes a rule of devoicing for German.
However, it would be unexpected if this rule would turn [g] into [t].
We would rather expect the rule to turn [g] into [k], [b] into [p] and
[d] into [t]. The questions that arise are as follows:
À Why is it that we expect matters to be this way?
Thus, the resulting sound is indeed [p]. You may experiment with
other AVS to see that the rule really operates as expected. Notice
that the rule contains [consonantal : +] on its left but does not
change it. However, you cannot simply eliminate it. The resulting rule
would be different:
h i h i
voice : + → voice : − / # (3.10)
This rule would apply to vowels and produce voiceless vowels. Since
German does not have such vowels, the rule would clash with the con-
straints of German phonology. More importantly, it would devoice
every word final vowel and thus—wrongly—predict that German has
no word final voiced vowels (counterexample: /Oma/ [oma] ‘grand-
mother’).
Suppose that you have to say this without features. It is not enough
to say that the voiced consonants are transformed into the voiceless
ones; we need to know which voiceless consonant will replace which
voiced consonant. The tie between [p] and [b], between [k] and [g] and
so on needs to be established. Because features have an independent
motivation the correspondence is specified uniformly for all sounds. As
will be noted throughout this course, some rules are not really specific
to one language but a whole group of them (final devoicing is a case
in point). This seems to be contradictory, because the rules are stated
using phonemes, and phonemes are language dependent, as we have
seen. However, the fact that language has a contrast between voiced
and voiceless is independent of the exact specification of what counts,
say, as a voiced bilabial stop as opposed to a voiceless bilabial stop.
Important is that the contrast exists and is one of voicing.
For example, Hungarian, Turkish and Finnish all have a rule called
vowel harmony. Simplifying somewhat, all rules agree that words can-
not both contain a back vowel and a front vowel. On the other hand,
the front close-mid rounded vowel of Finnish (written /ö/) is pro-
nounced with more lip rounding than the Hungarian one, also writ-
ten /ö/. Nevertheless, both languages systematically oppose /ö/ with
/o/, which differs in the position of the tongue body (close-mid back
rounded vowel). The situation is complicated through the fact that
Hungarian long and short vowels do not only contrast in length but
also in a feature that is called tension. Finnish /ö/ is tense even
when short, while in Hungarian it is lax (which means less rounded
and less close). However, even if short and long vowels behave in this
phonology i: features and phonemes 39
way, and even if back and front vowels are different across these lan-
guages, there is good reason to believe that the contrast is between
‘front’ and ‘back’, no matter what else is involved. Thus, among the
many parameters that define the actual sounds, languages decide to
systematically encode only a limited set even though one still needs
to fill in details as for the exact nature of the sounds. Precisely this
is the task of realization rules. These are the rules that make the
transition from phonemes to sounds. They will be discussed in the
next chapter.
Natural Classes
Suppose we fix the set of attributes and values for a language. On the
basis of this classification we can define the following:
This is still not as clear as I would like this to be. First, we need
to say something about the classification system used above. Let P
be our set of phonemes. Recall that this set is in a way abstract.
It is not possible to compare phonemes across languages, except by
looking at their possible realisations (which are then sounds). We
then define (using our (pre-)theoretical insights) some features and
potential values for them. Next we specify which sounds have which
value for which attribute. That is to say, for each attribute A and
value v there is a set of phonemes written [A : v ] (which is therefore a
subset of P ). Its members are the phonemes that are said to have the
value v for the attribute A. In addition, we require that the following
holds:
this is to add a specific value to the attribute, call it ?, and then de-
clare that all vowels have this value. This ‘value’ is not a value in the
intended sense. But to openly declare that vowels have the ‘non-value’
helps us be clear about our assumptions. To show the first, an intersection of no
subsets of P is defined to be identical
Definition 2 (Natural Class) Let S be a classification system for to P , so that is why P is natural. To
show the second, let H be the intersec-
P . A subset U of P is natural in S if and only if it is an intersection
tion of all sets [A : v ] that contain p.
of sets of the form [A : v ] for some attribute and some legitimate value. I claim that H = {p}. For let p0 6= p.
Then there are A, v, and v 0 such that
I shall draw a few conclusions from this. v 6= v 0 p ∈ [A : v ] and p0 ∈ [A : v 0 ].
However, p0 6∈ [A : v 0 ], since the sets
1. The set P is natural. are disjoint. So, p0 6∈ H. Finally,
for the third, let there be at least two
2. For every p ∈ P , {p} is natural. phonemes, p and p0 . Then there are
A, v and v 0 such that p ∈ [A : v ],
3. If P has at least two members, ∅ is natural. p0 ∈ [A : v 0 ] and v 6= v 0 . Then
[A : v ] ∩ [A : v 0 ] = ∅ is natural.
The reason is that the vowel [e] is obligatorily followed by [j] and there-
fore mentioning of [j] is needless. However, unless you know English
well you need to be told this fact. The sequence [aı] is different in
that [a] is not necessarily followed by [ı], hence writing the sequence is
unavoidable. I shall indicate now how Table 3.5 establishes a classifi-
cation system and how it is written down in attribute value notation.
To make matters simpler, we concentrate on the consonants. There
are then three attributes: place, manner, and voice. We assume
that the features have the following values:
The sounds with a given place features are listed in the columns, and
can be read off the table. However, I shall give them here for conve-
nience:
[ manner : stop] = {/p/, /b/, /t/, /d/, /tS/, /dZ/, /k/, /g/}
[ manner : fricative] = {/f/, /v/, /T/, /ð/, /s/, /z/, /S/, /Z/}
[ manner : nasal] = {/m/, /n/, N /}
[ manner : l approx] = {/l/}
[ manner : c approx] = {/w/, /j/, /ô/}
(3.13)
[ voice : +] ={/b/, /d/, /g/, /dZ/, /v/, /ð/, /z/, /Z/, /m/, /n/,
/N /, /w/, /l/, /ô/, /j/}
[ voice : −] ={/p/, /t/, /k/, /tS/, /f/, /T/, /s/, /S/}
(3.14)
So, one may check, for example, that each sound is uniquely character-
ized by the values of the attributes; /p/ has value bilabial for place,
42 an open introduction to linguistics
Binarism
The preceding section must have caused you to think that for a given
set of phonemes there must be several possible classification systems.
Indeed, not only are there several conceivable classification systems,
phonologists are divided on the issue of which one to actually use.
There is a never-ending debate on the idea of binarism of features.
Binarism is the idea that features have just two values: + and −.
In this case, an alternative notation is used; instead of [att : +] one
writes [+att] (for example, [+voiced]) and instead of [att : -] one
writes [−att] (for example [−voiced]).
Although any feature system can be reconstructed using binary val-
ued features, the two systems are not equivalent, since they define
different natural classes.
Consider, by way of example, the sounds [p], [t] and [k]. They
are distinct only in the place of articulation (bilabial vs. alveolar vs.
velar). The only natural classes are: the empty one, the singletons
or the one containing all three. If we assume a division into binary
features, either [p] and [t] or [p] and [k] or [t] and [k] must form a
natural class in addition. This is so since binary features can only cut
phonology i: features and phonemes 43
a set into two parts. If your set has three members, you can single
out a given member by two cuts and only sometimes by one. So you
need two binary features to distinguish the three from each other.
But which ones do we take? In the present case we have a choice of
[+labial], [+dental] or [+velar]. The first cuts {[p], [t], [k]} into {[p]}
and {[t], [k]}; the second cuts it into {[t]} and {[p], [k]} and the third
into {[k]}, and {[t], [p]}. Any combination of two of these features
allows for the singleton sets as natural classes. If you have only two
features then there is a two element subset that is not a natural class
The choice between the various feature bases is not easy and hotly
disputed. It depends on the way the rules of the language can be sim-
plified which classification is used. But if that is so, the idea becomes
problematic as a foundational tool. It is perhaps better not to enforce
binarism.
Study Questions
1. It is very useful to get to grips with making minimal pairs. Take
some of the sounds of English (see the table of English phonemes in
the chapter) and make some minimal pairs (that are not already in
the book) to show that these sounds belong to distinct phonemes?
a. [f@nEIks]
b. [SUô]
c. [ôaItIN]
d. [Igzæm]
e. [noU]
f. [@bæUt]
a. skunk
44 an open introduction to linguistics
b. heath
c. nudging
d. uncommon
e. English
8. In the following sets of four sounds, three sounds belong to the same
natural class, while the remaining phoneme, the ’odd one out’, does
not. Give the phoneme that is the odd one out and the attribute-
value pairs which describe the difference. There might be more than
one solution.
a. [d T x Z]
b. [ð Z S s]
c. [Y W u ø]
d. [I 3 2 æ]
p ph b bh m
t th d dh n
ú úh ã ãh ï (4.1)
c ch , é éh ñ
k kh g gh N
To describe these sounds we use the following features and values: By the way, if you read the sounds as
they appear here, this is exactly the
cons(onantal) :+ way they are ordered in Sanskrit. The
Sanskrit alphabet is much more logi-
manner :stop, fric(ative) cally arranged than the Latin alpha-
bet!
place :bilab(ial), dent(al), retro(flex), velar, palat(al)
(4.2)
asp(irated) :+, −
nas(al) :+, −
voice :+, −
p ph b
place:bilab place:bilab place:bilab
asp :−
asp :+
asp :−
nas :− nas :− nas :−
voice :− voice :− voice :+
(4.3)
bh m
place:bilab place:bilab
asp :+
asp :−
nas :+
nas :−
voice :+ voice :+
48 an open introduction to linguistics
All other classes are intersections of the ones above. For example, the
class of phonemes that are both retroflex and voiced can be formed by
looking up the class of retroflex phonemes, the class of voiced phonemes
and then taking the intersection:
m ph ã
place:bilab place:bilab place:retro
asp :−
asp :+
asp :−
(4.6)
nas :+
nas :− nas :−
voice :+ voice :− voice :+
The first is nasal, but the others are not. So the description can-
not involve nasality. The second is voiceless, the others are voiced.
Therefore, the description cannot involve voicing. The same goes for
aspiration and place. It means that the smallest natural class that
contains this set is—the entire set of sounds. Yes, the entire set of sounds is a nat-
This example was in some sense easy: there was no feature that the ural class. Why? Well, no condition
is also a condition. Technically, it cor-
phonemes shared. To see another example, look at the set {[p], [ph ], responds to the empty AVS, which is
[b]}. Agreeing features are blue, disagreeing features are red (I have denoted by [ ]. Nothing is in there,
so any phoneme fits that description.
marked the agreeing features additionally with +):
p ph b
+ place:bilab place:bilab place:bilab
asp :−
asp :+
asp :−
+ nas :− nas :− nas :−
voice :− voice :− voice :+
(4.7)
X →Y / C D
(4.9)
Input → Output Context
a→b (4.10)
Let’s look at an example. The rules of spelling require that one uses
capital letters after a period (that’s simplifying matters a bit since the
period must end a sentence). Since the period is followed by a blank—
written ␣, in fact, maybe there are several blanks, but let’s ignore that
too—the context is ‘.␣ ’. D is omitted since we place no condition
on what is on the right of the input. So we can formulate this for the
letter a as
a → A/.␣ (4.14)
k → g/ #b (4.17)
k# → g/ b (4.18)
The omission of the voicing specification means that the rule applies
to any feature value. Notice that on the right hand side we find the
pair [ voice : +]. This means that whatever voice feature the original
sound had, it is replaced by [ voice : +].
Next, if the consonant of the following word is a nasal, the preceding
consonant becomes a nasal. The choice of the nasal is completely
determined by the place of articulation of the original stop, which is
not changed. So, predictably, /p/ is changed to /m/, /t/ to /n/, and
so on.
h i voice:+ h i
nas:− → asp :− / # nas:+ (4.21)
nas :+
Thus, we use the rules (4.19), (4.21), (4.23), (4.24) and (4.25). Given
that the latter three abbreviate two rules each this leaves us with a
total of 8 rules as opposed to 60.
Neutralization of Contrast
There are also cases where the phonological rules actually obliterate
a phonological contrast. We’ll discuss an example of this called final
devoicing. In Russian and German, stops become devoiced at the end
of a syllable. It is such rules that cannot be formulated in the same
way as above, namely as rules of specialization. This is so since they
involve two sounds that are not allophones, for example [k] and [g] in
German. We shall illustrate the German phenomenon, which is rather
widespread. The contrast between voiced and voiceless is phonemic:
Now look at the following words: Rad (wheel) and Rat (advice). They
are pronounced alike: ["Ka:t]. This is because at the end of the syllable
(and thus at the end of the word), voiced stops become unvoiced:
" #
h i +stop
+stop → / # (4.27)
−voiced
So how do we know that the sound that underlies Rad is [d] and not
[t]? It is because when we form the genitive the [d] does not disappear:
This is because the genitive adds an [s] (plus an often optional epen-
thetic schwa) and this schwa suddenly makes the [t] and [d] non-final, In phonology, epenthesis means the
so that the rule of devoicing does not apply. addition of one or more sounds to
a word, especially to the interior of
a word. http://en.wikipedia.org/
Phonology: Deep and Surface wiki/Epenthesis
The fact that rules change representations has led linguists to posit
two distinct sublevels. One is the level of deep phonological repre-
sentation and the other is that of surface phonological represen-
tation. The deep representation is more abstract and more regular.
For German, it contains the original value for voicing, regardless of
the environment the consonant is in. The surface representation, on
the other hand, contains the phonological description of the sounds
that actually appear; so it will contain only voiced stops at the end
of a syllable. The two representations are linked by rules that have
the power of changing one representation into the other. In terms of
IPA–symbols, we may picture the change as follows:
["Ka:d]
↓ (Final Devoicing) (4.30)
["Ka:t]
However, what we should rather be thinking of is this:
−vowel +vowel −vowel
+approximant +open +stop
#
+velar +front +labiodental
↓ (4.31)
−vowel +vowel −vowel
+approximant +open +stop
#
+velar +front +labiodental
+voiced +long −voiced
There is another option: we can
deny that the voicing contrast at the
We may in fact view all rules proposed above as rules that go from
end of a syllable is phonological—a
deep to surface phonological mapping. Some of the rules just add fea- voiced stop like [b] is just realized
tures while some others change features. The view that emerges is (i.e. put into sound) in two differ-
ent ways, depending on the environ-
that deep phonological structure contains the minimum specification ment. This means that the burden is
necessary to be able to figure out how the object sounds, while pre- on the phonology–to–phonetics map-
serving the highest degree of abstraction and regularity. For example, ping. However, the evidence seems to
be that the syllable-final [b] is pro-
strings are formed at deep phonological level by concatenation, while nounced just like [p], and therefore
on the surface this might not be so. We have seen that effect with final simply is transformed into [p].
devoicing. While on the deep level the plural ending (or a suffix like
chen) is simply added, the fact that a stop might find itself at the end
of the word (or syllable) may make it change to something else. The
picture is thus the following: the word /Rad/ is stored as a sequence
of three phonemes, and no word boundary exists because we might
decide to add something. However, when we form the singular nomi-
native, suddenly a word boundary gets added, and this is the moment
the rule of final devoicing can take effect.
phonology ii: realization rules and representations 55
m a U/ı s
+vowel
+stop " # +fric
+low +vowel (4.32)
+nasal +alveol
+back +lower high
+bilab −voiced
−rounded
m a U/ı s
+vowel
+stop " # +fric
+low +vowel
+nasal +alveol ← [+ front]
+back +lower high
+bilab −voiced
−rounded
(4.33)
By convention, this is
m a ı s
+vowel
+stop +vowel +fric
+low
(4.34)
+
nasal +lower high +alveol
+back
+bilab +front −voiced
−rounded
The singular then has the representation ← [+ back]. You may have wondered about the
A downside of this approach, however, is that it leaves us with a fact that the so-called root of a noun
like /cat/ was identical to its singular
more complex picture of phonological representation. It does not just form. This is not actually the case; the
contain items that define sounds but also funny symbols that act on root does not have a word boundary,
while the plural does (as nothing can
sounds somewhere else in the representation. attach itself after the plural suffix).
Notes
You may have wondered about the use of different slashes. The slashes
are indicators showing us at which level of abstraction we are work-
ing. The rule is that /·/ is used for sequences of phonemes while [·]
is used for sequences of actual sounds. Thus, one rule operates on
the phonemic level while another operates on the phonetic level. Fur-
thermore, writers make a distinction between the surface phonological
level and the deep phonological level. (There could also be a deep
phonetic and surface phonetic level, though that has to my knowledge
not been proposed.) In my own experience textbooks do not consis-
tently distinguish between the levels, and for the most part I think
that writers themselves do not mentally maintain a map of the levels
and their relationships, which is a popular source of mistakes. Since I
have not gone into much detail about the distinction between the two,
obviously there is no reason to ask you to be consistent in using the
slashes. It is however important to point out what people intend to
convey when they use them.
In principle there are two levels: phonetic and phonemic. At the
phonemic level we write /p/, and at the phonetic level we write [p].
Despite the transparent symbolism there is no connection between the
two instances of p. Phonemics does not know how things sound, it
only sees some 40 or so distinct categories of sounds. Writing /p/ is
therefore just a mnemonic aid; we could have used ♠ instead of /p/
(and get rid of the slashes, since it is now clear that we are not on the
phonetic level). Phonemes may be organised using features, but the
phonology ii: realization rules and representations 57
Another is to translate /p/ into the ‘broad’ vowel [p] and leave the
specification to phonetics. Both views have their advantages, though
I prefer the first over the second. I have made no commitment here,
though to either of the views, so the use of /·/ versus [·] follows no
principle and you should not attach too much significance to it.
I also use the slashes for written language. Here they function in
a similar way: they quote a sequence of characters, abstracting from
their surface realisation. Again, full consistency is hard to maintain
(and probably not all that necessary). A side effect of the slashes is
that strings are easier to identify in running text.
With regards to our sketch in Chapter 1 I wish to maintain (for the
purpose of this course at least) that phonology and phonetics are just
one level; and that phonology simply organises the sounds via features
and contains the abstract representations, while phonetics contains the
actual sounds. This prevents us from having to deal with too many
levels.
Study Questions
1. In your own words, describe what a natural class is. Is an intersec-
tion of two natural classes also a natural class? What is the largest
natural class for a given set of sounds? And the smallest?
5. For each of the following AVSs that define a natural class, give the
set of phonemes from English (see Table ?? on Page ??) that belong
to that natural class.
58 an open introduction to linguistics
a.
consonant:+
:bilabial
place
voice :−
b.
consonant:−
height :open
rounded :−
voiced :+
c.
" #
consonant:+
place :dental
a. Identify the allophones of the phoneme /n/. Make sure you do not confuse the com-
bination of I and n, In, and m!
b. Write a set of rules that describes their occurrence in this data
c. Can you give an explanation for why this variation and distri-
bution is found?
a. In what environments do [l] and [&l ] occur? Hint: If you don’t see the answer im-
mediately, carefully write out what
b. Write a phonological rule that captures this variation. sound occurs before and after each us-
age to try to find a pattern.
Utterances are not mere strings of sounds. They are structured into
units larger than sounds. A central unit is the syllable. Words consist
of one or several syllables. Syllables in English typically start with one
or more consonants, followed by a vowel or a diphthong and then one
or more consonants again. The first set of consonants is the onset,
the group of vowels the nucleus and the second group of consonants
the coda. The combination of nucleus and coda is called rhyme. So,
syllables have the following structure:
[stô [E NT]]
Onset Nucleus Coda
(5.2)
Rhyme
Syllable
Thus, the onset consists of three consonants: [s], [t] and [ô], the nucleus
consists just of [E], and the coda has [N] and [T]. We shall begin with
some fundamental principles. The first concerns the structure of the
syllable.
This principle is not fully without problems and that is the reason that
we shall look below at a somewhat more general principle. The main
problem is the unclear status of some sounds, for example [Ä]. They
are in between a vowel and consonant, and indeed sometimes end up
in onset position and sometimes in nuclear position, for example in
/bird/ [bÄd].
The division into syllables is clearly felt by any speaker, although
there can be hesitation as to how to exactly divide a word into sylla-
bles. Consider the word /atmosphere/. Is the /s/ part of the second
syllable or part of the third? The answer is not straightforward. In
particular the stridents (that is, the sounds [s], [S]) enjoy a special sta-
tus. Some claim that they are extrasyllabic (not part of any syllable
at all), some maintain that they are ambisyllabic (belonging to both
syllables). We shall not go into that here.
A validation for the existence of rhymes can be given by looking at
verses (which also explains the terminology): words that rhyme do not
need to end in the same syllable, they only need to end in the same
rhyme: /fun/ – /run/ – /spun/ – /shun/. Similarly, the idea of the
coda is sustained by the fact that it is the domain of a rule that affects
many languages: For example, in English and Hungarian, within the There is an interesting problem
caused, among other things, by nasals.
coda the obstruents must either all be voiced or unvoiced; in German
Nasals are voiced by default. Now try
and Russian, all obstruents in the coda must be voiceless. The notion to find out what will happens when
of onset was used in Germanic verse in the Middle Ages, which used a you pronounce words with a sequence
nasal+voiceless stop in the coda, such
rhyming technique where the onsets of the rhyming words had to be as /hump/, /stunt/ and /Frank/.
the same.
It is worthwhile to remain with the notion of the domain of a rule.
Many phonological constraints are seen as conditions that concern two Rhyming words with the same on-
adjacent sounds. When these sounds come into contact, they undergo set is also called alliteration. It al-
lows to rhyme two words of the same
change to a smaller or greater extent, for some sound combinations are stem; German had a lot of Umlaut and
easier to pronounce than others, such as Sandhi, which we have dis- ablaut, that is to say, it had a lot of
cussed at length in Chapter 4. Another example is the Latin word /in/ root vowel change, making it impos-
sible to use the same word to rhyme
‘in’, a verbal prefix, which changes in writing (and in pronunciation) with itself (say /run/ – /ran/).
to /im/ when it precedes a labial (e.g. /impedire/). Somewhat more
radical is the change from [ml] to [mpl] to avoid the awkward combi-
nation [ml] (the word /templum/, for example, derives from /temlom/,
with /tem/ being the root, meaning ‘to cut’).
There is an influential theory in phonology, autosegmental phonol-
ogy, which assumes that phonological features are organized on differ-
ent scores (tiers) and can spread to adjacent segments independently
from each other. Think for example of the feature [±voiced]. The
condition on the coda in English is expressed by saying that the fea-
ture [±voiced] spreads along the coda. Clearly, we cannot allow the
feature to spread indiscriminately, otherwise the total utterance would
be affected. Rather, the spreading is blocked by certain constituent
boundaries; these can be the coda, onset, nucleus, rhyme, syllable,
foot, or word. To turn that on its head: the fact that features are
blocked indicates that we are facing a constituent boundary. So, the
existence of voicing harmony indicates that English has a coda.
The nucleus is the element that bears the stress. We have said that
phonology iii: syllable structure & stress 61
Syllable Structure II
Within a syllable the sonoricity follows a strict pattern: it in-
creases, then decreases again and is highest in the nucleus.
This means that the nucleus must contain at least one sound which is
at least as sonorous as any of the others in the syllable. It is called
the sonoricity peak. Thus, in the onset, consonants must be orga-
nized so that the sonority rises, while in the coda it is the reverse.
These conditions say nothing about the nucleus itself. In fact, some
diphthongs are increasing ([ı@] as in the British English pronunciation
of /here/), while others are decreasing ([aı], [oı]). This explains why
the phonotactic conditions for the onset are the opposite of those for
the coda. You can end a syllable in [ôt], but you cannot begin it that
way. You can start a syllable by [tô], but you cannot end it that way
Let me briefly note why diphthongs are not considered problematic
in English: it is maintained that the second part of the diphthong is
actually a glide (not a vowel), and as such is part of the coda. Thus,
/right/ would have the following structure:
r a j t
(5.3)
O N C C
will say [estôenZ]. The effect of this manoeuvre is that [s] is now part
of the coda of an added syllable, and the onset is reduced to just [tô], French is similar, but French speak-
ers are somehow less prone to add
or, in their pronunciation, most likely [tr]. Notice that this means
the vowel. (French has gone through
that Spanish speakers apply the phonetic rules of Spanish to English, the following sequence: from [st] to
because if they applied the English rules they would still end up with [est] to [et]. Compare the word
/étoile/ ‘star’, which derives from
the onset [stô] (see the Maximise Onset principle below). Latin /stella/ ‘star’.
given and changed as more things come in. But this viewpoint compli-
cates matters unnecessarily. As syllabification is largely predictable,
there is good reason to believe most of it is not stored. It is enough
to just insert syllable boundary markers in the lexicon where they are
absolutely necessary and leave the insertion of the other boundary
markers to be determined later. Another reason is that rate of speech
influences pronunciation, which in turn influences syllable structure.
Language Particulars
Languages differ in what types of syllables they allow. Thus, not only
do they use different sounds, they also restrict the possible combina-
tions of these sounds in particular ways. Finnish allows (with few
exceptions) only one consonant at the onset of a syllable. Moreover,
Finnish words preferably end in a vowel, a nasal or ‘s’. Japanese sylla-
bles are preferably CV (consonant plus vowel). This has effects when
these languages adopt new words. Finns, for example, call the East
German car /Trabant/ simply [rabant:i] (with a vowel at the end and
a long [t]!). The onset [tr] is simplified to just the [r]. The same can
be seen in other loanwords: /koulu/ [kOulu] ‘school’ has lost the ‘s’,
/rahti/ [rahti] ‘freight’ has lost the ‘f’, and so on. Notice that it is
always the last consonant that remains.
English too has constraints on the structure of syllables. Some
are more strict than others. We notice right away that English does
allow the onset to contain several consonants, similarly with the coda.
However, some sequences are banned. Onsets may not contain a nasal
except in the first place (exception: [sm] and [sn]), although there are
some loanwords that break the rule, e.g. /mnemonic/ and /tmesis/.
The sequence sonorant-obstruent is also banned ([mp], [ôk], [ôp], [lt]
and so on; this is a direct consequence of the sonoricity hierarchy.
Stridents are found only in the first place; exceptions are [ts] and [ks],
found however only in non-native words. The cluster [ps] is reduced to
[s]. It also makes a difference whether the syllable constitutes a word
or is peripheral to the word. Typically, syllables have to be simpler
inside the word than at the word boundary.
Syllabification is based on expectations concerning syllable struc-
ture that derive from conditions on the well-formedness of syllables.
However, these still leave some room for variation. ["pleito] (/Plato/)
can be syllabified ["plei.to] or ["pleit.o]. In both cases, the syllables we
get are well formed. However, if a choice exists, then preference is
given to creating open syllables and/or syllables with onset.
All these principles are variations of the same theme: if there are
consonants, languages prefer them to be in the onset rather than the
coda, also known as Maximise Onset.
Maximise Onset
Put as many consonants into the onset as possible.
The way this operates is as follows: first we need to know what the
nucleus of a word is. In English this is easy: vowels and only vowels are
nuclear. Thus, we can identify the sequences consisting of coda+onset
64 an open introduction to linguistics
but we do not yet know how to divide them into a coda and a subse-
quent onset. Since the first syllable of a word can only begin with a
(null) onset, we say that a sequence of consonants x is a legitimate
onset of English if there is a word such that the largest stretch of
consonants it begins with is identical to x.
Legitimate Onsets
A sequence of sounds is a legitimate onset if and only if it is the
onset of the first syllable of an English word.
Stress
Syllables are not the largest phonological unit. They are themselves or-
ganised into larger units. A group of two, sometimes three syllables is
called a foot. A foot contains one syllable that is more prominent than
the others in the same foot. Feet are grouped into higher units, where
again one is more prominent than the others, and so on. Prominence is
marked by stress. There are various ways to give stress to a syllable.
Ancient Greek is said to have marked stress by pitch (the stressed syl-
lable was about 3/2 of the frequency of an unstressed syllable). Other
languages (like German) use loudness, while other languages use a
combination of the two (Swedish). Within a given word, there is one
syllable that is the most prominent. In IPA, it is marked by a pre-
ceding ["]. We say that it carries primary stress. Languages differ
with respect to the placement of primary stress. Finnish and Hun-
garian place the stress on the first syllable, French on the last. Latin
put the stress on the second-to-last (penultimate) syllable if it was
long (that is to say, had a long vowel or was closed); otherwise, it was
the syllable that preceded it (the antepenultimate). Thus, we had
pe.re.gri.nus (‘foreign’) with stress on the penultimate (gri) since
the vowel was long, but in.fe.ri.or with stress on the antepenulti-
mate /fe/ since the /i/ in the penultimate ri was short. Sanskrit
is said to have had free stress, that is to say, stress was free to fall
anywhere in the word.
Typically, within a foot the syllables like to follow each other in a
specific pattern. If the foot has two syllables, it consists either of an
unstressed syllable followed by a stressed syllable (iambic metre), or
vice versa (trochaic metre). Sometimes a foot carries three syllables
(a stressed syllable followed by two unstressed ones, a dactylus). So,
if the word has more than three syllables, it will have more than one
foot, and therefore there will be a syllable that is more prominent than
its neighbours while not carrying main stress. You can try this with
the word /antepenultimate/. You will find that the first syllable is
more prominent than the second but less so than the fourth. We say
that it carries secondary stress: [antEp@n"2ltımEt].
The so-called metrical stress theory tries to account for stress as
follows. The syllables are each represented by a cross (×). This is a
Layer 0 stress. Then, in a sequence of layers, syllables get assigned
more crosses. The more crosses, the heavier the syllable. The number
of crosses is believed to correspond to the absolute weight of a sylla-
ble. So, a word that has a syllable of weight 3 (three crosses) is less
prominent than one with a syllable of weight 4. Let’s take the word
/assimilation/, [@sım@"leıSn].
Layer 0 × × × × ×
(5.4)
@ sı m@ leı Sn
We have five syllables. Some syllables get extra crosses. The syllable
[sı] carries primary stress in /assimilate/. Primary stress is always
marked in the lexicon, and this mark tells us that the syllable must get
a cross. Further, heavy syllables get an additional cross. A syllable
66 an open introduction to linguistics
Layer 1 × ×
Layer 0 × × × × × (5.5)
@ sı m@ leı Sn
Layer 2 ×
Layer 1 × ×
(5.6)
Layer 0 × × × × ×
@ sı m@ leı Sn
If larger units are considered, there are more cycles. The word /maintain/
for example has this representation by itself:
Layer 2 ×
Layer 1 × ×
(5.7)
Layer 0 × ×
meın teın
Layer 3 ×
Layer 2 × ×
Layer 1 × × × × (5.8)
Layer 0 × × × × × × ×
meın teın @ sı m@ leı Sn
schwa: [@]). On the other hand, we can also say that a vowel is that
way because it is unstressed. Take, for example, the pair ["ôi@laız] and
[ôi@lı"zeıSn]. When the stress shifts, the realisation of /i/ changes. So,
is it rather the stress that changes and makes the vowel change quality
or does the vowel change and make the stress shift? Often, these
questions have no satisfactory answer. In this particular example it The tendency to simplify consonant
clusters in Finnish has diminished
seems that the stress shift is first, and it induces the vowel change. somewhat in recent times. Consider
It is known that unstressed vowels undergo reduction in time. For recent loanwords such as: /trolli/
example, the reason why French stress is always on the last syllable [trol:i], ‘internet troll’ and /spammata/
[spAm:AtA], ‘to spam’, as compared
is because it inherited the stress pattern from Latin, but the syllables to older loans like /ranska/ [rAnskA],
following the stressed syllable eventually eroded away. Here the stress ‘French’, from Swedish ‘franska’. This
is likely due to an increased knowl-
was given and it drove the development! edge of foreign languages, especially
English. Similarly, Finnish dialects
spoken in close proximity to Swedish-
Study Questions speaking areas of Finland have gener-
ally been more tolerant of word-initial
1. Describe how the following terms relate to each other: coda, foot, consonant clusters than others.
syllable, stress, rhyme, nucleus, onset, open, closed.
Hint: try pronouncing the sounds
from Table 5.1 in hierarchical order
2. What are the sonoricity hierarchy and sonoricity peak used for? To
what property of articulation is the notion of sonoricity related?
3. What are the arguments for the idea that syllable information is
not stored in the mental lexicon? Are there any exceptions?
4. What principle explains the fact that we syllabify the word /pro-
prietary/ as /pro.pri.e.tar.y/, and not, for example as /prop.-
ri.e.tar.y/?
6. For each of the following English words, give the syllable structure
of the word and mark the primary (and, if applicable, secondary)
stress.
a. restricted
b. disappeared
c. stressful
d. encyclopaedia
e. protester
a. [Imp6sIbl]
b. [fr2streItINli]
c. [trEZ@]
68 an open introduction to linguistics
8. We have seen that languages differ in the way they fit loanwords
into their phonology. Finnish, for example tends to adapt loanwords
so that they do not have word-initial consonant clusters and end in
a vowel (e.g. /koulu/), while English generally leaves them intact,
even if they do not fit in with general phonological patterns (e.g.
/tsunami/). What are the advantages and disadvantages of these
approaches? Which do you think is better?
6
Phonology IV: Rules, Constraints
and Optimality Theory
All these choices give different results. We shall illustrate this with an
interesting problem, the basic form of the past tense marker in English.
The question is: what is the source of these differences? Sure, it is pos-
sible to say that the past has three different forms and that depending
on the verb a different form must be chosen. This, however, misses
one important point, namely that the choice of the form is determined
solely by the phonological form of the verb and can be motivated by
phonological constraints of English. Therefore, the variation should
be explained in the phonological stratum, not anywhere else The facts
can be summarized as follows.
À [d] is found if the last sound of the verb is voiced but unequal to
[d].
Á [t] is found if the last sound of the verb is voiceless but unequal to
[t].
∅ → [@] /C C0
(6.2b)
(C and C0 similar)
The symbol ∅ denotes the empty string. The first rule should actually
be written as two rules in our feature system:
" # " #
voice : + voice : −
→ /[−voice] #
manner:stop manner:stop
" # " # (6.3)
voice : + voice : −
→ /[−voice] #
manner:fricative manner:fricative
The second rule effectively says that it is legal to insert a schwa any-
where between similar consonants. Since we have two rules, there is a
choice as to which one shall be applied first. We shall initially schedule
(6.2b) before (6.2a). This means that the rule (6.2b) is applied to the
original form F0 , giving us an output form F1 , and then we apply rule
(6.2a) to get F2 . Each rule applies only once, so the final output is F2 .
This yields the following result:
Notice that, when we say that a rule does not apply, this does not
mean that no output is generated. Instead, it means that the form is
left unchanged by the rule. Textbooks sometimes indicate this by a
hyphen, as if to say that there is no output. But there is an output, it
just happens to be identical to the input. Notice, too, that sometimes
a rule does apply, but does not actually change anything.
72 an open introduction to linguistics
" # " #
+voice −voice
→ /[−voice] # (6.9a)
+obstruent +obstruent
[@] → ∅ /C C0 (6.9b)
0
(C and C dissimilar)
Initially, we apply rule (6.9b) before rule (6.9a):
that onsets may not contain more than one consonant. This constraint
exists in many languages, for example in Japanese and Finnish. But
the two languages apply different repair strategies. While Japanese
likes to insert vowels, Finnish likes to cut the onset to the last vowel.
Both repair strategies are also attested in other languages. However,
we could imagine a language that simplifies an onset cluster to its first
element: with this strategy /trabant/ would become not /rabantti/
but /tabantti/ in Finnish, and Germanic /hrengas/ would become
not /rengas/, but /hengas/ instead, and so on. This latter strategy
is however, not attested anywhere. So, we find that among the infi-
nite possibilities to avoid forbidden cluster combinations, only some
get used at all, while others are completely disfavoured.
To a certain degree, we can think of reasons why some strategies
exist and why others are not used at all. Look at schwa insertion as
opposed to schwa deletion, for example. While the insertion of a schwa
is inevitably going to improve the structure (because all languages
agree that CV is a syllable), the deletion of a schwa has the risk of
producing clusters that are illegitimate. Thus, we would expect that
there is a bias towards the repair by schwa insertion as opposed to
schwa deletion. Note, however, that all these arguments have to be
taken with care. If a rule causes stress changes, for example, there
might be good reason to reduce or eliminate a vowel.
Optimality Theory
There are two main approaches to describe the regularities of language:
Note that generative linguists do not always propose that rules are
real, that is, in the head of a speaker, nor that derivations take place
in real time. They would say that the rules are simply a way to system-
atize the data. If that is so, however, it is not clear why we should not
adopt a purely descriptive account, describing all and only the legal
representations rather than pretending that they have been derived
in a particular way. The discussion is complicated by the fact that
many arguments drawn in favour of either analysis come from data
on child development and language learning. To interpret the data
coming from learning, we need to have a theory of the internal knowl-
edge of language (‘language faculty’). This theory is related to the
approach to language, since this internal knowledge may either consist
of representations and conditions on the representations, or in fact rep-
resentations plus a set of rules. The discussion concerning the problem
of whether we should have rules or not is probably a never-ending one.
phonology iv: rules, constraints and optimality theory 75
b æ t d
(6.12)
• • • •
• • • •
↓ ↓ ↓ ↓ (6.13)
• • • •
b æ t t
b æ t d
• • • •
↓ ↓ ↓ (6.14)
• • •
b æ t
b æ t d
• • • •
↓ ↓ ↓ ↓ (6.15)
• • • • •
b æ t @ d
Recover Obstruency
If a segment of the UR is an obstruent, it must be an obstruent
in the SR.
Recover Voicing
If a segment is voiced in the UR, it must be voiced in the SR.
Syllable Equivalence
The UR must contain the same number of syllables as the SR.
Recover Adjacency
Segments that are adjacent in the UR must be adjacent in the
SR.
À Suppose that π = hI, Oi and π 0 = hI, O0 i are such that for all
constraints that π violates there is a more valuable constraint that
π 0 violates. Then O is called optimal with respect to O0 .
The actual output form for I is the optimal output form for I; if there
are several optimal output forms, all of them are chosen. So, given I,
if we want to know which SR corresponds to it, we must find an O
which is optimal. OT uses the following terminology: O
and O0 are called candidates and
they get ranked. Note that candidacy
is always relative to the UR.
phonology iv: rules, constraints and optimality theory 77
Let’s apply all this to [bætd]. We rank the constraints NTS (Not-
Too-Similar), RObs (Recover Obstruency), RMorph (Recover the Mor-
pheme) and RAdj (Recover Adjacency) as follows:
The first three are ranked equal, but more than the fourth. Other
principles are left out of consideration.
The forms [bætd], [bænd], [bæt] all violate constraints that are higher
ranked than RAdj. [bæt@d] violates the latter only. Hence it is optimal
among the four. (Notice that we have not counted a violation of Re-
cover Obstruency for [bæt], even though one obstruent was dropped.
This will be discussed below.) Note also that the optimal candidate
still violates a constraint.
We now turn to the form /sæpd/. Assume that the ranking is (with
VAgr = Voice Agreement, RVoice = Recover Voicing)
b æ t d
• • • •
↓ ↓ ↓ ↓ (6.20)
• • • •
b æ n t
78 an open introduction to linguistics
b æ t d
• • • •
↓ ↓ ↓ (6.21)
• • • •
b æ n d
Notes
The transition from underlying /d/ to /t/, /@/ or /d/ has mostly been
described as morphophonemics. This is a set of rules that control
the spelling of morphemes in terms of phonemes. In this view, there
is one morpheme, which is denoted by, say, /PAST/M and there are Recall that / · /M means: this is a unit
rules on how to realise this morpheme on the phonological level (recall on the morphological level
Study Questions
1. What is the advantage of using the concept of a single realisation
of a morpheme, modified by repair rules, compared to the use of
multiple realisations?
3. Describe how the following terms and concepts, all part of Op-
timality Theory, relate to each other: underlying representation,
segments, constraints, surface representation, optimality, ranking
of rules.
4. What is the main argument for dealing with the [d], [t], [@d] vari-
ation on the phonological, and not the morphological level? And
what about irregular past tense markers? Can these be explained
by phonological considerations, too?
Word Classes
Morphology is the study of the smallest meaningful units of language.
It studies the structure of words from a semantic viewpoint rather
than from the viewpoint of sound. Morphology is intimately related to
syntax, but their domains are separate. Everything that is larger than
a word falls under the scope of syntax, while morphology is concerned
with the structure within words only.
The first thing to notice is that words come in different varieties,
called word classes (also called parts-of-speech). For example,
there are verbs (/to imagine/) and there are nouns (/a car/), there
are adverbs (/slowly/) and there are adjectives (/red/).
a few English affixes and lists the word classes to which they can be
attached. We see that the list of affixes is heterogeneous, and that
affixes do not always attach to all members of a class with equal ease
(/anti-house/, for example, is yet to be found in English). Still, it is
a powerful test that reveals a lot of the division of words into different
classes.
Affix Attaches to Forming Examples Table 7.1: English Affixes and Word
Classes
anti- nouns nouns anti-matter, anti-aircraft
adjectives adjectives anti-democratic
un- adjectives adjectives un-happy, un-lucky
verbs verbs un-bridle, un-lock
re- verbs verbs re-establish, re-assure
dis- verbs verbs dis-enfranchise, dis-own
adjectives adjectives dis-ingenious, dis-honest
-ment verbs nouns establish-ment, amaze-ment
-ize nouns verbs burglar-ize
adjective verbs steril-ize, Islamic-ize
-ism nouns nouns Lenin-ism, gangster-ism
adjectives nouns real-ism, American-ism
-ful nouns adjectives care-ful, soul-ful
-ly adjectives adverbs careful-ly, nice-ly
-er adjectives adjectives nic-er, angri-er
The SUSANNE Corpus on the other hand has 355 distinct tags.
This all shows that word classes are not written in stone, and there
is much theoretical and practical debate about which ones could
84 an open introduction to linguistics
Compounding
In English, a compound can be recognised by its stress pattern. For
example, the main stress in the combination adjective+noun is on the
noun if they still form two words (black board [blæk "bOôd]) while
in a compound the stress is on the adjective part (/blackboard/
["blækbOôd]). Notice that the compound now is a single word; the ad-
jective has lost its status as an adjective through compounding, which
explains the new stress pattern.
In English, multi-word constructions are favoured over compounds.
It has to be said, though, that this difference is sometimes purely
orthographical; the use (or not) of spaces between the words does
not really tell you whether you are dealing with one or two words.
For example, although one writes /rest room/, (American Eng. for
public toilet) on the phonological level the stress pattern sounds as if
it is a single, compounded word.
There are languages where the compounds can be distinguished
from multi-word constructions by other aspects than stress patterns.
German is such a language: Funny compound from
https://www.orphicpixel.com/illustrated-
Regierung-s-entwurf (7.5) by-keren-rosen/
government proposal
Schwein-e-stall (7.6)
pig sty
Räd-er-werk (7.7)
wheel work = mechanism
Derivation
English has a fair amount of derivational affixes. Table 7.1 shows some
of them. We said that derivation often changes the category of the
word, but this is not always the case. I.e. if an affix changes the word
86 an open introduction to linguistics
class, it is derivational, not inflectional, but if the word class does not
change, derivation and inflection have to be distinguished in another
way. One difference is that derivation is optional (while inflection is
not), and can be iterated (i.e. done multiple times consecutively; inflec-
tion cannot be iterated). Additionally, inflectional affixes are typically
found outside of derivational affixes. To give an example: you can
form /republican/ from /republic/. To this you can add /anti/:
/antirepublican/ and finally form the plural: /antirepublicans/.
But the plural marker /s/ is an inflectional suffix. You could have
added the ‘s’ to /republic/, but then you could not go on with the
derivation; there is no word ∗ /republicsan/. Similarly, the word
/antirepublics/ has only one analysis: first /anti/ is added and
then the plural suffix /s/ is added.
The distinction between a root word and a word formed by deriva-
tion is not always clear, either. For example, is /resist/ formed by
affixation of /re/ to a root /sist/ meaning “to sist again”? Actually,
it was in the original Latin (/re-sistere/), but in English it is not.
This is because we do not have a verb ∗ /sist/. Thus, derivation may
form words that initially are perceived as complex, but later lose their
transparent structure. This may be because they start to sound dif-
ferently because some sort of sound change has occurred, or because
the base form disappears from the modern vocabulary.
Very often language users are completely unaware of the original
origin (or etymology) of the words they use. For example, the word
/nest/ in English (e.g. as in a bird’s nest) was once was a complex
word made from two morphemes combined: ∗ /nizdo/. Here we use
an asterisk to mark that the word form is reconstructed by historical
linguists based on evidence. The word is believed to be derived from
the words ∗ /ni/ or /nether/ meaning (‘down’) and ∗ /sed/ (‘sit’). This
is what historical linguists have determined. But in modern English,
we would simply treat nest as an indivisible root, and (most) speakers
are entirely unaware that it is originally a complex form.
Inflection
0
88 an open introduction to linguistics
0
Let’s work through a few examples to see how this works. Consider
the following two English words with suffixes.
1. govern+ment (as in the Dutch government)
V inf.
see -s
V infl.
deriv. V -ed
re- visit
Note that re- combines with verbs to
Diagramming the morphological structure becomes more interesting make new verbs, so we know that re-
had to have combined with nationalize
when more than two morphemes are involved. Consider the derivation (there is no *renation or *renational.
of ungracefully and renationalize. : If we put it in the past-tense, then the
inflectional suffix -ed would be added
Adv last.
Adj deriv.
un- N deriv.
grace -ful
deriv. V
N deriv. -ize
nation al
What’s interesting is that every native speaker, and also non-native
speakers, on some unconscious level know that this is the derivation of
these words. However, it requires explicit teaching (and some analysis
and practice) to be come aware of the hidden structure.
morphology i: basic facts 91
Study Questions
1. Say you want to determine the word class (also called part-of-
speech, or POS) of a certain word, i.e. whether it is a noun, verb,
adjective or adverb (actually, there are many more (or less!) word
classes, depending on who you ask). What are the various methods
you could use? Which one do you prefer?
2. What is the main distinction between root words and affixes? How
can you check whether something is a root or an affix?
a. noun
b. preposition (e.g. /to/, /over/)
c. adjective
d. verb
5. For each of the following words, identify all morphemes which make
up the word.
a. sadness
b. unkind
c. underworld
d. kingdom
e. simplified
f. highlighters
g. irreplaceable
h. antidisestablishmentarianism
a. rational
b. John’s
c. bigger
d. untie
92 an open introduction to linguistics
e. eaten
f. criticize
g. studied
Now consider the following. The plural morpheme has four different
morphs (in our example, that is) . The first is /z/, the second is root
vowel change, the third is ∅ (the empty morph) and the fourth is
/@n/. Which morph is used in a particular case depends on the noun
to which we apply the morpheme. On the deep morphology level we
only see a combination of a noun root with the plural morpheme; the
noun root-based choice of what morph to use is made at the surface
morphology level.
Circumfixes
A circumfix consists of a part that is added to the beginning of a word
and another part that is added to the end. It is thus a combination of
prefix and suffix. The German perfect tense is marked by a circumfix.
It consists of the prefix /ge/ and a suffix which changes from stem to
stem (usually it is /en/ or /t/). The infinitive is a suffix /en/ as can
morphology ii: morphemes and morphs 95
The two parts (the prefix and the suffix) are historically of different
origin (the prefix /ge/ did not exist originally, and is not found in
English). In present-day German, however, there is no reason to see
the two as being separate.
The superlative in Hungarian also is a circumfix:
Infixes
Infixes insert themselves inside the string. Look at the following data
from Chrau.
The string /an/ is inserted after the first consonant! The string is cut
into two pieces and the nominaliser inserts itself right in between.
Reduplication need not copy the full word. For example, in Latin
some verbs form the perfect in the following way. The first consonant
together with the next vowel is duplicated and inserted:
Pingelapese is a language spoken in
pendit ‘he hangs’ pependit ‘he has hung’ Micronesia, by approximately 3,000
tendit ‘he stretches’ tetendit ‘he has stretched’ people. It belongs to the Micronesian
(8.10) branch of the Austronesian language
currit ‘he runs’ cucurrit ‘he has run’ family. Map from CIA.
spondet ‘he promises’ spopondit ‘he has promised’
Note that the last example shows that /s/ is exempt from reduplica-
tion.
Morpheme or word?
There is no way to predict whether some piece of meaning is expressed
by a morpheme, a separate lexeme or both, and languages differ greatly
morphology ii: morphemes and morphs 97
in this respect: some languages pack all kinds of meanings into the verb
(Inuit, Mohawk), some keep everything separate (Chinese), and most
languages do something in between. Because of this, morphology is
more important in some languages than in others.
However, even within a single language, there are varying ways to
express something. Adjectives in English, for example, have three
forms: positive (normal form), comparative (form of simple com-
parison) and superlative (form of absolute comparison). Now look
at the way they get formed:
The first set of adjectives take suffixes (zero for the positive, /er/ for
the comparative and /est/ for the superlative). The second set of
adjectives take a separate word, which is added in front (/more/ and
/most/). The third set is irregular. The adjective /good/ changes
the root in the comparative and superlative before adding the suffix
(/bett/ in the comparative and /b/ in the superlative), while /bad/
does not allow an analysis into root and suffix in the comparative and
superlative. We may define a suffix for /worse/,
but it will be a one-time-only suffix,
since there is no other adjective like
Allomorphy /bad/. It therefore makes not much
sense to define a separate comparative
In the above discussions of different inflectional and derivational mor- suffix for /worse/. However, /worst/
is a debatable case.
phemes we have ignored the fact that actually sometimes the same
meaning is expressed by different forms, and sometimes the choice of
these forms is phonologically conditioned, just as allophones are. Let’s
look at an example from the Native American language Cree:
the present-tense of the verb. If we look at the word for canoe and
compare it to my canoe data we can already see that a prefix has been
added in the possessive form, [ni-]. We can guess that this means my.
We can then check this preliminary hypothesis by looking at the rest of
the nouns and their corresponding possessive forms. Our guess seems
to be true for examples (c.) and (d.), and (e.) and (f.) but when we
get to [e:mihkwa:n], spoon, the first-person possessive form seems to be
[nit-]. The same seems to be true for my hat. Now, just like allophones,
different allomorphs are often phonologically conditioned, so we should
look at the root form it attaches to in order to see if we can predict
when the form [ni-] appears and when [nit-] is used. It is actually quite
simple: it seems that if attaching to a vowel, [nit-] is preferred, which
makes sense from an articulation perspective: pronouncing two vowels
next to each other is hard: you either have to add a glottal stop (e.g.
[?] or you have risk turning it into a dipthong. With this hypothesis,
you can also make a very good guess as to how you would say my pipe
in Cre: [nitospwa:kan].
So now we have identified two allomorphs for the first-person pos-
sessive prefix in Cree. We can then write a rule predicting their use.
Just like with allophones, we should try to write the specific cases,
and then note that the more common form is used elsewhere. IN this
case, given that there are more consonants than vowels, it would be
logical to write a rule saying that we should use the [nit-] prefix before
a vowel, and the form [ni-] everywhere else.
Now let’s consider the verb forms. Here we see there is only one
lexical word for two (or more) English words, so we can guess that
morphological marking, rather than pronouns, are being used to signal
person. The example only shows singular forms, so we don’t have to
worry about number.
Looking at she/he counts and I count, we can see two things. First,
a similar prefix seems to be used for the first person here as in the
possessive forms: before a vowel we see [nit-] in both I count and I
sit. and I rest. We can also notice that all the first person forms end
with an [-n], whereas the third person forms end with a [-w]. Realizing
this makes it possible for us to analyze all the present tense forms
given. However, the last four forms are future tense forms, and here
we see something different. Compare s/he sits, [apiw] with s/he will
sit: [kaapiw], and what we see is the addition of a morpheme ka- before
the root. If we hypothesize that [ka- means will, then we can analyze
these forms as well, except for the prefix [ni+]. But here we can guess
that the same phonologically based allomorphic variation is probably
at work that we saw with the genitive form: Cree doesn’t like vowels
next to each other so if the root begins with a vowel the first person
morpheme is realized as [nit-, but when attaching to a consonant, [ni-
is preferred. Given this I would also guess that the double vowels in the
example are actually lengthened vowels, because they are two vowels
of the same kind (but to know for sure we would need to look this up
in a grammar of Cree.
morphology ii: morphemes and morphs 99
Study Questions
1. Define the following terms and relate them to each other: mor-
pheme, affix, morph, allomorph, suffix, prefix, circumfix, infix, trans-
fix.
a. -ment
b. -ize
c. -al
d. un-
e. -ance/-ence
f. -ee
g. en-/em-
4. Words that are made up of more than two morphemes possess some
kind of internal structure, i.e. a word is not simply a linear com-
position of its morphemes. For example, the word goalkeepers has
the structure [[[goal][keeper]]s], i.e. goal and keeper are combined
first, after which the plural suffix is added. For the following En-
glish words, give the morphemes of which they consist, the type of
morpheme (root/suffix/prefix and inflectional/derivational, if appli-
cable) and the internal structure of the word (either as a tree or by
using brackets).
a. irreparable
b. replacements
c. premeditated
d. surcharges
8. Allomorphs: Egaugnal
Consider the following verbs in present and past tense in Egaugnal
(‘language’ spelled backwards!), a fake language I created.
morphology ii: morphemes and morphs 101
Occurrences
Sentences are not only sequences of words. There is more structure
than meets the eye. Look for example at:
Visually we can see that the words are ordered. When spoken the
words will be ordered in time. This ordering is linear. It satisfies the
following word postulates:
we find the word /the/ three times (we do not distinguish between
lower and upper case letters here). However, the definitions above
suggest that /the/ precedes /the/ since it talks of words. Thus, we
must change that and talk of occurrences. The best way to picture
occurrences is by underlining them in the string:
Constituents
Consider for example /a fortune/. This is a sequence that has very
different properties than a sequence like /costs a/. We can see ev-
idence of their difference by looking at whether or not the sequence
can be replaced by a single word. We can easily replace /a fortune/
with the single word /much/ without affecting grammaticality:
{{1, 2, 3, 4, 5}, {1, 2}, {1}, {2}, {3, 4, 5}, {3}, {4, 5}, {4}, {5}} (9.15)
Typically, the brackets around single words are omitted, though. This
gives the slightly more legible
2 • • 3
4 • • 5 6 • • 7
a fortune
• The Sand People took [the robot]. –> The robot was taken
(by the Sand People).
Beware!
All these tests have to be used with caution. Not all constituents
pass all tests. In general you should always try several tests to
confirm that what you think is a constituent, actually is.
Categories
~x ~y (9.18)
syntax i: categories, constituents, trees and context free grammars 109
An example of a context is
that /villa costs/ is a constituent ...?) And the same three tests fail with:
Bob ate too many [apples yesterday].
We agree to call arbitrary strings of English words constituents
• Clefting *It was apples yesterday
just in case they occur as constituents in some sentence. The wording that Bob ate too many.
here is important: I do not say that they have to occur everywhere • Answer ellipsis *What did Bob
as constituent. It may happen that a sequence occurs in one sentence eat too many of? Apples yesterday.
as a constituent, but not in another. Here is an example. The string • Passivization *Apples yesterday
were eaten too many (by Bob).
/you called/ is a constituent in (9.21) but not in (9.22):
First we give some criteria for constituent occurrences. ~x has a • I saw the fox that wore a hat.
constituent occurrence in ~z only if there are ~u and ~v such that: The string the fox is a constituent in
the first sentence, but not in the sec-
ond. First note that in the second sen-
À ~z = ~u␣~x␣~v ; tence the string the fox cannot be re-
placed with a pronoun, e.g.
Á ~z ∈ E (that is, ~z is a grammatical sentence);
• *I saw it that wore a hat.
Thus, given that ~x and ~y are members of the same category, if ~u~x~v is
a grammatical sentence of English, so is ~u~y~v and vice versa.
The definition of a category is actually similar to that of a phoneme;
phonemes were defined to be substitution classes that preserve mean-
ing. In syntax we define substitution classed based on the ability to
preserve grammaticality (more or less, since we have the caveat about
‘constituent occurrences’ because we are dealing with the substitution
of strings for strings, not just of an item for another item). We give
an example. The intransitive verbs in the 3rd person singular form a
category:
S
Figure 9.2: A Labelled Syntactic Tree
•
DP• • VP
D• • NP •V •DP
a fortune
S → DP VP (9.30)
S). Consider by way of the example the following grammar. The start
symbol is S, the rules are
S → DP VP (9.31a)
DP → D NP (9.31b)
D → a | this (9.31c)
NP → villa | fortune (9.31d)
VP → V DP (9.31e)
V → costs (9.31f)
Here, the vertical stroke ‘|’ is a disjunction. It means ‘or’. For example,
the notation D → a | this is a shorthand for two rules: D → a and
D → this.
This grammar says that (9.1) is of category S. How does it do that?
It says that /this/ is a determiner (D; Rule (9.31c)), and /villa/ is
a noun phrase (NP; Rule (9.31d)). By Rule (9.31b) we know that the
two together are a DP. Similarly, it tells us that /a fortune/ is a DP,
and that /costs a fortune/ is a VP. Finally, using Rule (9.31a) we
get that the whole is an S.
Given a set of rules R, we define a derivation as follows. An R-
derivation is a sequence of strings such that each line is obtained
from the previous by doing one replacement according to the rules of
R. If the first line of the derivation is the symbol X and the last the
string ~x we say that the string ~x has category X in R. The following
is a derivation according to the set (9.31a) – (9.31f).
S (9.32)
DP VP
D NP VP
D NP V DP
D NP V D NP
this NP V D NP
this villa V D NP
this villa costs D NP
this villa costs a NP
this villa costs a fortune
Notes
Let’s return to our definition of constituent occurrence. Context free
grammars provide strings with a constituent analysis. However, not
every such analysis conforms to the definition of constituents. To make
everything fall into place we require that the grammars of English have
the following universal rule schema. For every category symbol X they
contain a rule
X → X␣and␣X (9.33)
X→w
~ (9.34)
for some word. If that is so, the criteria À – Ã will be met by all
constituents assigned by the grammar.
Study Questions
1. Show with at least two constituency tests that the bracketed words
in the following examples are constituents, if possible. If they are
not constituents, show how they fail at least two tests.
4. Create rules for a Context Free Grammar that can generate the
following set of sentences.
But actually, you do not want your CFG to produce these sentences.
How could you prevent this? Explain your answer by showing ex-
amples of how your CFG-rules would have to be modified.
10
Syntax II: Argument Structure
ran.
The child
lost the battle of Waterloo.
Napoleon
My neighbour’s dog
ate the cake.
is beautiful.
(10.12)
Projection
How does syntax know that a verb like ran doesn’t take an object
argument and uses a CGF rule like VP –> V without an object, but
that a verb like greet must have an object argument? Where does this
information about the structures associated with lexical items come
from? Modern syntax believes that the structures associated with
words are projected from the words themselves, with the specifics of
what type of projections each word allows coded in the mental lexicon.
Thus:
This means that a verb phrase like fight the empire gets its structure
from the projection of the verb fight.
VP
V DP
fight D NP
the N
empire
In the mental lexicon, the word and the structure are stored, with
empty nodes for the other arguments. We know what type the other
arguments are, because we know that fight takes an object argument:
118 an open introduction to linguistics
VP
V DP
fight D NP
In the trees we’ve drawn above, we’ve only show binary branching.
In our context-free grammar in the previous chapter we also only have
rules that would lead to binary branching because we only have rules
where a non-terminal LHS can be rewritten into two symbols on the
RHS. But we could have rules that lead to three branches, and it might
seem that this would make sense given sentence structures having a
subject, a verb, and an object, see below.
Flat S
DP V DP
Binary Branching S
DP VP
Luke V DP
Flat DP
DP
D ADJ N
D NP
this ADJ N
red lightsaber
Verb-centric syntax
Verbs can restrict the types of sub-
But actually the binary branching structure we have above is not quite jects they can occur with. For ex-
right. We know that an intransitive verb does take one argument: a ample, only animate things can eat
or revise things, so *The lamp ate
subject argument. Further, a transitive verb takes two arguments, a and *The daisy revised the speech are
subject and an object. These are arguments of the verb, but in the weird. This is further evidence that
features of the subject argument are
binary branching examples above, we’ve somehow let the sentence itself controlled by the verb.
generate the subject argument. But actually, the verb, which describes
the event that the sentence communicates, also constraints the types
of subjects it can occur with so it makes sense to treat everything as
projected from the verb. In essence, that means we think of sentences
as a kind of verb phrase. And because every phrase should be a project
of a lexical category, it fits with our goal to get a uniform analysis that
we get rid of S as a start symbol (because there is no lexical category
S) and instead make VPs our start symbol.
VP
DP VP
Luke V DP
phrase and then has a level 1 V, a level one verbal projection. (10.15)
is for intransitive verbs, (10.16) is for transitive verbs. So actually we
need to add an attribute to make sure that only transitive verbs get
objects. To do this we’ll add an attribute trs with values + and −. TRS = Transitivity, TRS : + = tran-
sitive while TRS : − = intransitive.
" # cat :V
cat :V
→ proj:0 (10.17)
proj:1
trs :−
" # cat :V " #
cat :V cat :D
→ proj:0 (10.18)
proj:1 proj:2
trs :+
cat :V
→ sit | walk | talk | · · · (10.19)
proj:0
trs :−
cat :V
→ take | see | eat | · · · (10.20)
proj:0
trs :+
This means that at some point we will need to add rules to allow
certain verbs to take sentential complements which will be CPs.
prep:of
This tells us that /accuse/ wants a subject (because all verbs do),
a direct object (because it is transitive) and a PP opened by /of/
(because this is how we defined the meaning of prep:of). To get the
positioning of the phrases right, we propose to add the following rules:
cat :V cat :V cat :P
proj :1 → proj :1 proj :2 (10.26)
prep:of prep:of prep:of
cat :P cat :P
proj :2 → proj :1 (10.27)
prep:of prep:of
cat :P cat :P " #
cat :D
proj :1 → proj :0 (10.28)
proj:2
prep:of prep:of
cat :P
proj :0 → of (10.29)
prep:of
This works similarly for all prepositions. Notice that we have used the
feature [prep: of] also for the preposition itself to make sure that the
right preposition is used.
The rules say that the PP is to be found to right of the direct object,
if there is one. This is generally the case:
X-bar syntax
Let’s step back and consider for a moment what our ideal syntactic
theory would be. We can probably agree that a theory with general
rules will be preferred to a theory with many exceptions. This is in part
because a simpler theory is preferred over a complex theory (Occam’s
razor). But stronger yet, linguists believe that a natural theory will
give a uniform analysis to structures based a few, general principles.
After much research, a general structure has been proposed for syn-
tactic phrases. It’s called X-bar syntax, where “X” stands for a word
class from which structure will project. X-bar syntax claims that all
phrases have the same structure. They have a core that is of a certain
word class in the language. This core, X, is called the head of the
phrase. This head can be modified by a phrasal structure to its left,
called the specifier position. Take a simple sentence for example,
like Luke fought the empire. We said above that the start symbol
for our grammar treats a sentence as a VP. In this VP then, the head
must be the verb itself, fought. The specifier is then the subject con-
stituent. Each phrase can also contain a phrasal complement. In this
example it would be the empire. Question: What is the specifier of the
What is then proposed is that all phrases have the following struc- verb in a sentence like John sleeps.?
Note that, because phrasal projec-
ture: tions (maximal projects) have their
source in the lexical information of
their head, the word class of a phrase
XP
and its head will always be the same.
So it’s not possible to have a rule
where XP –> Z’ YP. A Z word, if a
(YP) X0 head, will always project a ZP.
specifier X (YP)
head complement
syntax ii: argument structure 125
XP → (YP) X0 (10.45a)
0
X →X (YP) (10.45b)
In X-bar theory all phrases will
The projection of X on the right is called the head. Subjects are have the same structure (though
specifiers of verbs, direct objects complements of verbs. The YP not all positions will be used!). The
structure of a DP is as follows:
in Rule (10.45a) is called the specifier of the phrase, the one in
Rule (10.45b) is called the complement. DP
We note the following general fact about English.
ZP D
In English, specifiers are to the left, complements to the right of
the head.
Specifier D YP
All these terms are relational. An occurrence of a phrase is not a
complement per se, but the complement of a particular phrase XP in head Complement
a particular construction. Now English, like all languages, also allows
adjuncts. A phrase or a bar-level category can combine with another The structure of an NP will be as
phrase as an adjunct. So to make our rules complete we need the follows:
ZP N
XP → XP YP (10.46a)
X0 → YP X0 (10.46b)
Specifier N YP
0 0
X →X YP (10.46c)
Recall that adjuncts are always optional. Further, you can have head Complement
multiple adjuncts. The rules above allow this by permitting e.g. an
And the structure of a DP with an
X-bar to expand to another X-bar by using e.g. (10.46b):
NP complement will be as follows:
DP ]
X
ZP D
YP X
Specifier D NP
X YP
head ZP N
If we look at other languages we find that they differ from English
typically only the relative position of the specifier, complement and
Specifier N YP
adjunct, not the hierarchy. Thus some languages make use of the
following rules:
head Complement
0
XP → X YP (10.47a)
XP → YP XP (10.47b)
0
X → YP X (10.47c)
It’s also the case that the order of combining categories can change
from category to category, even within the same language. German
puts the verb at the end of the clause, so the complement is to its left.
The structure of nouns and PPs is however like that of English nouns.
Japanese and Hungarian put prepositions at the end of the PP. (That
126 an open introduction to linguistics
are different from other verbs in several ways. For example, in the
third person singular form they never take the inflection ending -s.
This category can project a phrase. Let’s consider a sentence with an
auxiliary verb.
DP TP
C3PO T V0
can V
speak Bocce
But then we seem to now have rules like the following:
• VP –> DP V
• VP –> DP TP
• TP –> T V
• V –> V DP
This however ruins the uniformity of the system that we had. There
seem to be three major problems. First, there is no T-bar level, and
instead we have two different kinds of VPs, VPs with modal and VPs
without. Second, we’ve now separated our VP from the V, but the
VP is supposed to be the projection of the V. How did that TP get
in between them? Third, the subject argument is now separated from
the verb in sentences with auxiliaries, but it is also supposed to have
been generated as an argument of the verb. All this suggests that this
is not the right analysis.
One solution to regain uniformity would be to simply argue that all
verb phrases have TPs, and actually that what we previous called a
VP, was actually a TP.
Notice that the correct elliptical answer is not speaks, but speak. The
third person singular inflectional affix -s is not part of the constituent!
Instead, it seems to be like can. That’s exactly what we are going
to assume. That inflectional morphemes that mark person or tense
(such as the regular past tense suffix -ed) are of category T, along
with auxiliary verbs. But we have to deal with problems two and
three as well. We don’t want to separate the verb from the subject or
the VP. A solution then could be to simply put the TP on top, and let
TPs take VP complements (instead of the other way around), so then
C3PO can speak Bocce will get the following structure:
TP
T VP
can DP V
C3P0 V DP
speak Bocce
(Spec) T
T VP
can DP V
C3P0 V DP
speak Bocce
So now we solved problems two and three. As for problem one, sure
we can put inflectional markers in T, and for verbs that don’t take
any inflections we’ll just add a zero-inflection, and then we’ve solved
the problem of not having a uniform analysis. However you may have
noticed that we’ve introduced a new problem: the surface order
of the sentence and the order of the words in the tree is no
syntax ii: argument structure 129
Step 1: S
TP
(Spec) T
T VP
can DP V
CP30 V DP
speak Bocce
Step 2: S
TP
DP T
C3P0 T VP
can DP V
tCP 30 V DP
speak Bocce
We will need to restrict movement, so there are some constraints.
We want to remember where things came from, so we’ll leave some-
thing behind, which we’ll call a trace. Further, it’s actually the case
that you can only move things to positions that accept that sort of
category. So we can move the DP to Spec-TP because that’s a po-
sition for a phrase (recall from X-bar structure, a YP can go there).
This is phrasal movement. Finally, there’s a structural constraint that
forces movement to be leftward and upwards: moved elements must
C-command their traces. C-command is something we’ll go into more
detail in the next chapter.
Now the current structure, with movement, will work fine for sen-
130 an open introduction to linguistics
tences like C3PO can speak Bocce. But we still have a word order
problem with sentences like C3PO speaks Bocce in that their analysis
becomes C3PO -s speak Bocce. There are two problems with this sen-
tence. The first is that the words are not in the right order. But the
second is that we said in the chapter on morphology that inflectional
affixes cannot stand alone. So it’s a bit odd that we let -s just hang
there as a node. It actually needs to be attached to the verb. But
hey, maybe we can solve both of these problems with movement as
well! What if we let the verb move to T? V and T are both the same
level of category, Projection level 0, and this movement would satisfy
the c-command requirement. We then let the verb and the inflectional
affix merge to end up with the correct surface structure: C3PO speaks
Bocce. Actually, this type of movement is called Head-to-head move-
ment and in the following diagram you can see it’s identifeid with a
dashed red line. Try drawing the steps involved for yourself.
We still have one final problem. We have S as a start symbol, but
this ruins our nice X-bar uniformity, and it means we have a phrase
project that doesn’t come from a lexical entry. It would be much nicer
if this S was something us. A potential candidate is a CP. Recall that
CPs are sentential complements to verbs like know and discover. It
would make everything even more uniform if all sentences, whether
matrix sentences or sentential complements, belong to the same cate-
gory! Win-win! Let’s therefore assume that all sentences are CPs, and
matrix sentences have an empty C category.
CP
C TP
(that) DP T
C3P0 T VP
speak+s DP V
tC3P 0 V DP
tspeak Bocce
∗
The sailors accused the captain on the ship. (10.56)
∗
The sailors accused the captain on the ship during the storm at midnight.
(10.57)
∗
The sailors accused the captain of treason of cruelty.
(10.58)
Study Questions
1. What are the three major syntactic processes?
2. Write out the CFG rules that capture the information AVS struc-
tures 10.26-10.28.
3. Write attribute value rules that you would need to add to 10.26-
10.29 in order to parse a prepositional phrase headed by the prepo-
sition to, e.g. to the lighthouse. Hint: you will need a verb that can
take to, like run. You will need four rules.
4. Identify the specifier and the complement of the verb in the follow-
ing sentences.
7. In the text we gave tree structure examples of the X-bar based rules
that can form DPs, with NP complements. Now write out the same
rules using AVS structures.
9. Draw a tree structure for the sentence CP3O talks. You will need
to use both VP-to-Spec TP and Head-to-head movement.
11
Syntax III: Local Dependencies and Constraints: Se-
lection, Agreement and Case Marking
Grammatical Features
Nouns come in two varieties: singular and plural. The feature of being
singular or plural is called number. We have already mentioned a few
ways in which the singular and plural of nouns are formed. For syntax,
the different ways of forming the plural are not relevant; only the fact
of whether a noun is singular or plural is relevant. A way to represent
the number of a noun is by adding an attribute num, with values sing
and pl (in other languages there will be more...). This means that the
syntactic representation of ‘mouse’ and ‘mice’ is
cat :N cat :N
mouse : proj:0 , mice : proj:0 (11.1)
num :sing num :pl
We say that like nouns, determiners have a singular and a plural form
and that the determiner agrees with the NP complement in number.
This will be put into the rules as follows. The letter α may be instanti-
ated to any legitimate value, in this case sing or pl. Notice that within
a single rule, each occurrence of α must be replaced by the same value.
However, whether or not the letter α also occurs in another rule is of
no significance. Grammatical Gender vs. Nat-
ural Gender Languages generally
make two distinctions: features that
mark natural gender mark whether
cat :D cat :D or not an animate indvidual is male
proj:2 → proj:1 (11.12)
or female. Grammatical Gender is a
num :α num :α type of noun classification system that
manifests as agreement between syn-
cat :D cat :D cat :N tactically related words in an utter-
rance. Some languages have mascu-
proj:1 → proj:0 proj:2 (11.13)
line and feminine (e.g. French, Urdu),
num :α num :α num :α some languages have common and
neuter, (e.g. Dutch, Swedish) while
cat :D some languages instead distinguish be-
→ this | the | a | · · · (11.14) tween Animate and Inanimate, (e.g.
proj:0
Basque, Hittite). Some languages, like
num :sing Swahili, have 18 noun classes.
cat :D
proj:0 → these | the | ∅ | · · · (11.15)
num :pl
if, roughly speaking, they refer to a specific entity given by the con-
text or if they are uniquely described by the DP itself. Otherwise they
are called indefinite. The determiner of definite DPs is for example
/this/, /that/ and /the/. The determiner of indefinite DPs is for
example /a/ in the singular, /some/ in singular and plural, and ∅ in
the plural.
In other languages, nouns have more features that are syntactically
relevant. The most common ones are case and gender. Latin, for ex-
ample, has three genders, called masculine, feminine and neuter.
In English the three genders have survived in the singular 3rd person
pronouns: /he/, /she/, /it/. The Latin noun /homo/ (‘man’) is mas-
culine, /luna/ (‘moon’) is feminine, and /mare/ (‘sea’) is neuter. The
adjectives have different forms in each case (notice that the adjective
likes to follow the noun, but it does not have to in Latin):
This is just like English /inflict/, which wants a direct object and a
PP[on].
For us, a grammatical feature is anything that defines a syntactic
category. You may think of it as the syntactic analogue of a phonemic
feature. However, beware that elsewhere the usage is a little different.
Grammatical category and projection level are typically discarded, so
that grammatical feature refers more to the kinds of things we have
introduced in this section above: number, gender, definiteness, and
case. A fifth one is person. (This list is not exhaustive.)
The position of subject can be filled with DPs like /the mouse/,
/a car/, but also by so-called pronouns. Pronouns are distinct from
nouns in that they express little more than that they stand in for a DP.
However, they have many different forms, depending on the grammat-
ical feature. In addition they show a distinction in person. Across
languages, there is a pretty universal system of three persons: 1st, 2nd,
and 3rd. First person means: includes speaker. Second person means:
includes hearer, and third person means: includes neither speaker nor
hearer. So, /I/ is first person (it includes me and no one else); it is
also singular, because it refers to just one thing. The plural /we/ is
used when one refers to several people, including speaker. By contrast,
/you/ is used for individuals or groups including the hearer (there is no Note that /we/ might also include the
distinction between singular and plural). The third person pronouns hearer, in addition to the speaker and
other people. The distinction between
distinguish also gender in the singular (/he/, /she/, /it/), but not /we/ with and without the hearer is
in the plural (/they/). Moreover, as we explained above, pronouns not made explicit in English, but must
be derived from the context. Other
distinguish nominative from accusative. Thus, there are more mor- languages do make this distinction ex-
phological distinctions in the pronominal system in English than there plicit, using different pronouns for the
first person plural with and without
are in the ordinary DPs.
the hearer.
More on Case
Cases have many different functions. One function is to indicate the
nature of the argument. A verb has a subject, and the case of the
subject is referred to as nominative. The direct object has a case
that is referred to as accusative. Some people believe that English
has cases (because pronouns still reflect a distinction between subject
and object: /she/ : /her/, /he/ : /him/, and so on). On the other
hand, this is confined to the pronouns, and nouns show no distinction in
case whatsoever. This is why most people say that English has no case.
Strictly speaking, it means only that there is no distinction in case.
(One might say: there is one case and only one. This is useful. For
example, there is a famous principle of syntactic theory which states
that nouns need case. If there is no case, this principle fails.) Chinese
syntax iii: local dependencies and constraints: selection, agreement and case
marking 137
Notice that two of these nouns have been obtained from transitive
verbs. The rule in English is that the noun can take any of the ar-
guments that the verb used to take (though they are now optional).
However, the subject and the object must now appear in the geni-
tive. The PPs on the other hand are taken over as is. For example,
/destroy/ is transitive, so the noun /destruction/ can take two geni-
Map of Hungary. (From Wikipedia)
tives, one for the subject and one for the object. The verb /talk/ takes
a subject, an indirect object and subject matter, expressed by a PP
headed by /about/. The latter two are inherited as is by the noun
/talk/, while the subject is put in the genitive. Alternatively, it can
be expressed by a PP[by].
Subject-Verb Agreement
English displays a phenomenon called subject-verb agreement. This
means that the form of the verb depends on the grammatical features
of the subject. The agreement system is very rudimentary; the only
138 an open introduction to linguistics
contrast that exists is that between singular 3rd and the rest:
Notice that since the verb does agree in person with the subject it has
to make a choice for DPs that are not pronominal. It turns out that
the choice it makes is that ordinary DPs trigger 3rd agreement:
This applies even when the DP actually refers to the speaker! So, So if Jennifer is lecturing in class and
agreement in person is (at least in English) not only a matter of what she (rather oddly) refers to herself
with a full DP, e.g. “Jennifer is lec-
is actually talked about, but it is also a syntactic phenomenon. There turing”, she still uses the third person
are rules which have to be learned. singular form of the verb to be, e.g.
*“Jennifer am lecturing”.
Other languages have more elaborate agreement systems. Let us
look at Hungarian. The verbal root is /lát/ ‘to see’.
The subject is the same in both sentences, but the object is indefinite
in the first (a bird) and definite in the second (the bird). When the
object is indefinite, the form /látok/ is used, to be glossed roughly as
‘I see’, while if the object is definite, then the form /látom/ is used, to
be glossed ‘I see it’. Hungarian additionally has a form to be used when
subject is first person singular and the direct object is 2nd person:
Én lát-lak. (11.32)
I see-sub:1sg.ob:2sg
Obligatoriness
Morphology is obligatory. A noun has a case in a language with cases,
it has a number in a language with numbers, and so on. This creates
a certain predicament about which I have spoken earlier in connection
with gender. Let me here discuss it in connection with number. The
intuitive approach is this. An English noun is either singular or plural.
If it is singular it means ‘one of a kind’ and if it is plural it means
‘several of a kind’. So, /car/ means ‘one car’, and /cars/ means
‘several cars’. But this immediately raises questions.
 What if the speaker cannot (or does not want to) communicate the
number of things involved?
Feature Percolation
This is the same also for the projection of N and V. There is in general
no connection between the features on the YP and that of the D, with
one exception: the determiner passes the number feature on to the NP
complement. It is the same also with gender features, and with case.
Might this therefore be a general feature of the X-bar syntax?
First, recall the rules (10.17) and (10.18). What we did not establish
then was the identity of the V0 with respect to its transitivity. We use
the following diagnostic: in a coordination only constituents which
have the same features can be coordinated. If this is so then the
following sentence would be ungrammatical if we called /kicked the
ball/ as a transitive V0 .
follows.
cat :V cat :V
→ proj:0 (11.48)
proj:1
trs :− trs :−
cat :V cat :V " #
cat :D
→ proj:0 (11.49)
proj:1
proj:2
trs :− trs :+
Study Questions
1. Identify the head of the NP in the following DPS.
3. Recall that selectional features are not passed from head to phrase,
while agreement features are. Show with AVSs like (10.47) and
(10.48) which features would percolate (or not) for a Dutch DP
with neuter gender, like het kind.
12
Syntax IV: Movement and Non-Local Dependencies
Movement
We have learned that in English the transitive verb requires its direct
object immediately to its right. This rule has a number of exceptions.
The first sentence below, (12.1), displays a construction known as top-
icalisation, the second, (12.2), is a simple question using a question
word.
Pilots Harry admires. (12.1)
Which country have you visited? (12.2)
We could of course give up the idea that the direct object is to the
right of the verb. It would be possible to argue that, in the case
of topicalisation, the sentence simply takes the Object-Subject-Verb
order. But the facts are quite complex, especially when we turn to
questions. For example, for question words, no matter what kind of
constituent the question word replaces (subject, object, indirect object
and so on), it occurs sentence initially, even if it is not the subject.
Alice has visited Madrid University in spring to learn Spanish.
(12.3)
What has Alice visited in spring to learn Spanish?
(12.4)
Who has visited Madrid in spring to learn Spanish?
(12.5)
When has Alice visited Madrid University to learn Spanish?
(12.6)
Why has Alice visited Madrid University in spring?
(12.7)
We see that the sentences involving question words (interrogative sen-
tences) differ from their declarative counterparts, (12.3), in that the
144 an open introduction to linguistics
question word is in the initial position, and the verb is in second place.
There is a way to arrive at an interrogative sentence in the following
way. First, insert the question word where it ought to belong according
to our previous rules. Next, move it to first position. Now move the
auxiliary (/has/) into second place. We can represent this as follows,
marking removed elements in red, and newly arrived ones in blue: In the example to the left, Where
has been moved from the position to
(1) Alice has travelled where in spring to learn Spanish? the right of travelled to the sentence-
initial position. Then, it has has been
moved from the position to the left of
travelled to the second position.
(2) Where Alice has visited where in spring to learn Spanish?
(3) Where has Alice has visited where in spring to learn Spanish?
(The underscore just helps you to see where the word came from. It
is typically neither visible nor audible.) The first notation is more
explicit in showing you which element came from where (assuming
they are all different). However, neither notation reveals the order in
which the movements have applied. It turns out, though, that this is
irrelevant.
The plus side of this proposal is that it is actually very simple.
However, we need to be clear about what exactly we are doing. Super-
ficially, it seems that what we are doing is deriving one surface sentence
of English from another surface sentence of English. (This was in fact
the way transformations were originally thought of.) However, we are
actually going to claim that the process is more substantial. We in-
stead will claim that every surface sentence of English is derived in a
two stage process. First, we generate a structure in which every head
projects its phrase, satisfying its requirements (such as verbs project-
ing a VP consisting of a subject and, if transitive, an object, and the
necessary PPs). After that is taken care of, we apply a few transfor-
mations to the structure to derive the actual surface string. Recall
that this is what we did with phonological representations. First we
generated deep representations and then we changed them according
to certain rules. Thus, we say that the context free grammar generates
deep syntactic representations, but that the rules just considered
operate on them to give a final output, the surface syntactic rep-
resentation. The surface syntactic representation is what speakers
actually say. The rules that turn deep structures into surface struc-
tures are the (syntactic) transformations.
syntax iv: movement and non-local dependencies 145
Wh-Movement
When talking about the transformations responsible for putting ques-
tion words at the front of a sentence we usually talk about a type
of movement, called Wh-Movement. Question words are also re-
ferred to as wh-words, since they all start with /wh/ (/who/, /what/,
/where/, /why/, etc.). At first blush one would think that syntactic
transformations operate on strings; but this is not so. Suppose the
original sentence was not (12.3) but
Wh-Movement I. (Preliminary)
Only phrases can be moved by Wh-Movement. What moves is
the smallest phrase containing a given wh-word. It moves to the
beginning of a clause (= CP).
Mary thinks you ought to see what city? (deep struct.) (12.15)
Here the wh-phrase moves to end of the main sentence (and notice
that something strange happens to the verb too):
However, some verbs dislike being passed over. In that case the wh-
phrase ducks under; it goes to the left end of the lower sentence.
∗
What city does Mary wonder you have seen? (12.17)
Mary wonders what city you have seen. (12.18)
We shall soon see below that this is not a good way of putting things,
since it refers to linear order and not to hierarchical structure.
Verb Second
Wh-movement is a movement of phrases. In addition to this there
is also movement of zero level projections, or heads. (It is therefore
called head movement.) A particular example is verb movement.
Many languages display a phenomenon called Verb Second or V2.
German is one of them. Looking at individual word occurrences, the
verb is not always in second place. But if we look at constituent
structures in a tree, we will see that the verb does appear in the second
place. Here is a pair of sentences with word-to-word translation.
Only the auxiliary (/will/) is found in the second place in the main
clause. Based on a number of additional facts, this can be explained
as follows. The verb is at the end of the clause in deep structure.
In a subordinate clause it stays there. Otherwise it moves to second
position.
syntax iv: movement and non-local dependencies 147
[[∅ [[will [John [see Alice]V0 ]VP ]T0 ]TP ]C0 ]CP (12.24) “Please help me. I am a prince!”
[[∅ [[John will [ see Alice]VP ]T0 ]TP ]C0 ]CP (12.26)
[[∅ [[s [John [see Alice]V0 ]VP ]T0 ]TP ]C0 ]CP (12.27)
c-commands
1 −
2 3, 6, 7, 8, 9
3 2, 4, 5
4 5
(12.40)
5 4
6 7, 8, 9
7 6
8 9
9 8
2 • • 3
4 • • 5 6 • • 7
a fortune
CP
•
Figure 12.2: Before Movement (Step
0)
• C0
C0 • • TP
T0
Q •
T0 • • VP
will DP • • V0
Simon • V0 • DP
like
which cat
CP
•
Figure 12.3: After Movement (Step 1)
DP • • C0
• C0 • TP
which cat
T0
Q •
T0 • • VP
will DP • • V0
Simon • V0 • DP
like ∅
syntax iv: movement and non-local dependencies 151
[CP [VP which salesperson wonders [CP [TP [T0 should [VP he
(12.45)
promote which product]]]]]]
To exclude this, we require first that every wh-phrase can only move to
the next CP. If it is unavailable, the wh-phrase stays in place. Second,
we shall say that /wonders/ blocks the movement of the wh-phrase.
X0 → Y 0 X0 (12.49)
CP
•
Figure 12.4: After Head Movement
DP •
(Step 2)
• C0
which cat
• C0 • TP
T0
T0 • C0 • •
will Q • T0 • VP
∅ DP • • V0
Simon • V0 • DP
like ∅
which cat
• C0 • TP
T0
T •
0
C •
0 DP • •
will Q Simon T0 • • VP
∅ DP • • V0
∅ • V0 • DP
like ∅
Notes
It should be clear that a proper formulation of a seemingly simple
idea such as moving a wh-phrase to the front of a sentence needs a
lot of attention to detail. And this is not only because the facts are
unpredictable. It is also because what looks like a simple proposal can
154 an open introduction to linguistics
become quite complex once we start looking harder. This applies for
example to programming, where something that starts out as a simple
program can become quite long because we need to make sure it works
properly in all cases.
Study Questions
1. What is the Deep Syntactic Structure according to the account
given in this chapter before all movement has taken place:
2. For the Deep Structures you gave above, what is the string after
wh-movement has applied (e.g. at Step 1)?
13
Syntax V: How to Draw Syntactic Trees
V DP
destroy it
The head of this phrase must be the verb trying, so the entire phrase is
thus a verb phrase. This verb then must take the TP phrase to destroy
it as its complement.
VP
V TP
trying T VP
to V DP
destroy it
syntax v: how to draw syntactic trees 157
We have already made quite a long phrase. But we have not yet
analyzed a complete sentence. Consider the expanded dialogue here:
Speaker A: What is the fellowship of the ring trying to do about the ring?
(13.7)
Speaker B: They are trying to destroy it. (13.8)
If we assume that clauses and sentences are formed the same way
phrases are, we can also assume that the same type of merge operation
can be applied to form the full clause. So we should expect that we
can first merge the auxiliary verb are with the VP trying to destroy it,
and then merge They with the result.
Auxiliary verbs like are belong to the category T, just like the in-
finitive particle to. This would seem to suggest that trying to destroy
it is a projection of the auxiliary verb are and the entire sequence is a
TP, an auxiliary phrase.
But there is a problem. Are trying to destroy it isn’t actually a
constituent. Check for yourself that it doesn’t pass any constituent
tests. This means it cannot be a TP. Speaker A: What is the fellowship try-
If Are trying to destroy it is not a constituent, then it must be some ing to do? Speaker B: *Are trying to
destroy it.
kind of incomplete phrase. In what way does it seem to be incomplete?
It seems that auxiliaries require a subject, and are trying to destroy it
does not have a subject. Consider other auxiliary verbs like can and
would. These are also not constituents. Other examples of T-bar structures:
Following previous syntactic analysis, we will assume that a lexical am leaving on a jet plane.
has gone.
item from category T, like the auxiliary expression are, can merge can congratulate the president
with a VP to form an incomplete auxiliary expression. We will call
this incomplete expression ‘T-bar’, also written as T, but also as T’,
for easier typesetting. When the T’ structure merges with a subject
DP we can form a complete auxiliary phrase. This means we will get
the following structure:
TP
DP T
They T VP
are V TP
trying T VP
to V DP
destroy it
Again, we will describe the auxiliary are as the head of a phrase, and
158 an open introduction to linguistics
CP
C TP
That DP T
they T VP
are V TP
trying T VP
to V DP
destroy it
Now we see that we have a sentence with a CP at the top. Actually,
all sentences are CPs, not just subordinate clauses. However, the C
position is empty in declarative main clauses in English. In English
yes-no questions, the CP head position is filled with the tensed head,
moved there via head-movement. For example:
syntax v: how to draw syntactic trees 159
CP
C TP
Will DP T
you T VP
twill V DP
Notes on X-bar
A major point at the end of chapter 3 in the textbook was that all
phrases have a uniform structure. We called this X-bar structure, and
it looks like this:
XP Recall that there are three types
of DPs: Names, which are dia-
grammed as nouns with an empty de-
Specifier X terminer, Pronouns, which are deter-
miners without an empty noun phrase
complement, and full NPs, which are
X YP generally made up of a determiner plus
a common noun, and where we might
Head Complement get adjective adjuncts.
All phrases follow this X-bar structure, except for the Determiner
Phrase, DP, which is a special exception. The X-bar expansion of
a DP does not have a D-bar. Instead, in the place where you would
expect to find the D, there is an NP. This NP is special too, as it al-
ways has D as its specifier. This D is the head of the whole Determiner
Phrase. A complete, empty DP thus looks like this:
DP
Specifier D
D NP
head Specifier N
N Complement
An example of a phrase that follows the regular X-bar structure is the
PP below:
160 an open introduction to linguistics
PP
Specifier P
P Complement
However, you should know that this structure is not limited to one
X’. Sometimes, you need multiple X-bar levels to be able to correctly
diagram your sentence. This allows for one or more extra arguments
in between the specifier and the head, so-called adjuncts. In the
following tree, ZP is in the adjunct position:
XP
Specifier X
ZP X
adjunct X YP
Head Complement
You may have noticed that in the tree derivations in the previous
sections, we only use this extensive X-bar structure in one position:
for the TP headed by are. This is in fact the only place in the above
derivation where using the entire X-bar expansion is actually necessary.
We could have added the expansions everywhere, and if we had, we
would have gotten trees like the following:
syntax v: how to draw syntactic trees 161
CP
TP
DP T
D T VP
D NP are V
They ... V TP
trying T
T VP
to V
V DP
destroy D
D NP
it ...
As you can see, there are numerous empty specifiers and complements.
These positions could contain lexical items, but for the sentences we
will be diagramming, they most often will not. What kind of items
could go in there then? Well consider this, for example:
162 an open introduction to linguistics
CP
TP
DP T
D T VP
D NP are V
They ... V TP
trying AP T
hard T VP
to AP V
quickly V DP
destroy QP D
all D NP
the N
rings
When the spec position is empty, which is frequently the case for many
DPs and TPs, the X-bar level is sometimes left, letting the head to
appear as a sister node to the complement, both directly under the
XP:
XP
X YP
Head Complement
Sometimes, when the phrase has no complement either, you can use
the even shorter form with the triangle notation, as we have done in
previous trees, mostly for DPs. For example:
syntax v: how to draw syntactic trees 163
CP
TP
DP T
They T VP
are running
If an exam or homework question says that you should diagram using
(full) X-bar structures, then you should not use the shorthand form.
Otherwise, the shorthand form is fine (except of course, when you need
the X-bar structure to correctly diagram the sentence).
One final note: We have focused on learning to diagram tree struc-
tures for a small number of phrase and sentence types, without much
variation. But of course, X-bar structure must be applicable to all
possible structure in language. One question that has come up is how
to deal with multiple modifiers of the verb. Consider the sentence :
DP V
Han V PP
V PP on Tatooine
V DP in the cafe
read a book
164 an open introduction to linguistics
Study Questions
1. What words could appear in the Spec-VP position between are
and trying in the example sentence They are trying hard to quickly
destroy the rings.
2. The sentences Han read a book in the cafe on Tatooine has (at
least) two possible syntactic structures. Diagram the VP so that on
Tatooine modifies the cafe rather than the reading event.
14
Syntax VI: Binding
Pronouns
In this chapter we look at a phenomenon that is most intimately con-
nected with c-command, namely binding. Binding is, as we shall see,
as much of a semantic phenomenon as a syntactic one, but we shall
ignore its semantic aspects as much as we can. What is important here
is that there are several different kinds of DPs: ordinary DPs, names
and pronouns. Pronouns can be either reflexive (like /myself/,
/yourself/ etc.) or not (/he/, /she/, etc.). In addition, there are
also reciprocals (/each other/). In generative grammar, the term
anaphor is used to refer to both reciprocals and reflexives but this is
not the usage elsewhere, where an anaphor is an expression that refers
to something that another expression already refers to (and therefore
also non-reflexive pronouns generally count as anaphors).
In English, the reflexive pronouns exist only in the accusative. There
is no ∗ /hisself/, for example. There a number of other important The pronoun /her/ can both be gen-
pronouns, such as demonstratives (/this/, /that/) and relative itive and accusative, so /herself/
doesn’t clearly illustrate this fact. On
pronouns (/who/, /what/, /which/), which are used to open a rela- the other hand, it is /ourselves/ and
tive clause. not ∗ /usselves/. English has not
been completely consistent in arrang-
I could not find the CD [which you told me about]. (14.1) ing its case system.
The enclosed constituents are called relative clauses. They are used
to modify nouns, for example. In that respect they are like adjectives,
but they follow the noun in English rather than preceding it, e.g.:
∗
I could not find the which you told me about CDs. (14.3)
∗
There is a new in which no one ever talks film. (14.4)
ment. One should strictly distinguish the relative pronoun from the
complementizer /that/ as found in sentential complements (/believe
that/). The complementizer occupies a different position (C0 , Spec-
CP). Some relative clauses can be opened by /that/.
3rd person pronouns, /is/ (‘he’), and /se/ (‘himself’). The reflex-
ive exists in all cases but the nominative. It is the same both in the
singular and the plural, see below:
nom −
gen sui
dat sibi (14.9)
acc se
abl se
Latin also has another set of pronouns called possessive. They are used
to denote possession. They are just like adjectives. For example,
Binding
These different kinds of expressions each have their distinct behaviour.
Look at the following three sentences.
However, in (14.12) the two occurrences of /John/ are not meant to In the books you often find letters i,
point to the same individual. If they are supposed to point to different j and k in place of concrete numbers,
but this not a good idea since it sug-
individuals, we write different numbers: gests that plain numbers are not ab-
stract enough. But in fact they are.
John1 votes for John2 in the election. (14.19)
Principle A.
A reflexive pronoun must be bound by a constituent in the same
CP (or DP).
Principle B.
A non-reflexive pronoun must not be bound by a constituent
inside the same CP.
Principle C.
A name must not be bound.
For example, (14.12) can be used successfully to talk about two in-
dividuals named John. So, the rules alone cannot tell us whether a
sentence is grammatical, but the rules do tell us sometimes what the
possible meanings are.
To understand the role of c-command, we need to look at the struc-
ture of some sentences. The subject c-commands the object of the same
syntax vi: binding 169
verb, since smallest constituent containing the subject is the VP, which
also contains the object. However, the object does not c-command the
subject, since the smallest constituent containing it is only the V0 . So
the following is illegitimate no matter what indices we assign:
∗
Himself voted for John. (14.21)
Notice that the principles not only say that these sentences are fine,
they also tell us about the assignment of indices. They claim that
(14.24) is fine but (14.25) is not.
∗
The queen1 was never quite sure about her1 . (14.28)
∗
The students1 were all talking to them1 . (14.29)
∗
The queen1 was never quite sure about herself2 . (14.30)
∗
The students1 were all talking to themselves2 . (14.31)
Here, different numbers mean that the expressions are meant to refer
to different individuals or groups. Principle B talks about binding
170 an open introduction to linguistics
inside the same CP. It does not exclude that a pronoun is bound in a
higher CP. This is borne out.
∗
John1 ’s lack of knowledge worried himself1 . (14.34)
Here, /John/ does not c-command the pronoun. But the reflexive must
be bound by something that c-commands it inside the clause. This
can only be /John’s lack of knowledge/. But if we think that, we
would have to say /itself/ rather than /himself/. Next, why is
(14.35) fine?
Movement Again
We have argued earlier that an element can only move into a position
that c-commands the earlier position. It seems that binding theory
could be used in this situation. However, in order to use binding
theory, we need to have an element that is being bound. Recall from
previous discussions that when an element moves it leaves behind an
empty element. But why is that so? To see this, look at movement
out of a subordinate clause.
The fact that agreement holds between binder and pronoun means
that certain indexations are not licit. Here is an example.
∗
Her1 lack of knowledge worried John1 . (14.47)
∗
Their disinterest in syntax1 bothered John1 . (14.48)
Also, we have said that a pronoun cannot be bound inside the same
CP. But it can be bound outside of it:
In the previous example, both indices are licit, since binding by /John/
or by /boss/ is compatible with agreement. However, in the following
sentence one of the options is gone.
Hungarian does have a case for being on top of, but in general the
situation is quite analogous to Finnish.
syntax vi: binding 173
We also find a group of verbs that form the perfect by adding /u/.
The difference is due to modern spelling. The letters /u/ and /v/ were
originally not distinct and denoted the same sound, namely [u], which
became a bilabial approximant in between vowels. The difference is
thus accounted for by phonemic rules.
Other verbs add an /s/. The combination /gs/ is written /x/.
There are verbs where the perfect is signalled by lengthening the vowel
(lengthening was not written in Roman times, we add it here for illus-
tration):
Now, things are made difficult by the fact that certain verbs use a
combination of these. The verb /frangere/ (‘to break’) is an example
of this: it uses loss of nasal, accompanied by vowel lengthening, plus
ablaut (vowel change):
The lengthening of the vowel could be attributed to the loss of the nasal
(in which case it is called compensatory lengthening). However,
the next verb shows that this need not always be the case.
The verb /tangere/ uses a combination of reduplication, loss of
nasal, and ablaut, without lengthening the vowel):
and that the verbal root cannot be pronounced as such, the solution is
to move the verb upwards, and form a complex head [C [V think][C s]].
Study Questions
1. Add indices to the referential expressions in the following sentences
to show binding. Clearly identify any potential ambiguities.
What is meaning?
What is meaning? What does it mean to understand a word, or a
statement that someone makes? How do we know that we understand
the same things when we hear the word blue? When I remark that
the sky is very blue today, and you agree, how do we then know that
we both have the same understanding of the word “blue”? These are
fundamental questions about what meaning is.
But in order to analyze meaning and deal with it in our applications,
we have to also consider how we can represent it.
These are all questions dealt with in semantics, the study of mean-
ing. Humans have attempted to understand what meaning is for cen-
turies. The type of semantic theory that is most relevant for compu-
tational applications is called model theoretic truth-conditional seman-
tics. This is also where most formal semantic research is carried out.
In this chapter, we’ll introduce the fundamental ideas of this theory
and how it represents meaning.
First let’s consider how we understand a simple word. Take a con-
crete noun like apple. How do we understand that someone is referring
to an apple when they say the word “apple”, and not, for example, a
pear or a chair? One (rejected) theory of linguistic meaning is that
when you hear a word, it conjures up a picture of what that word can
refer to in your mind. Now for the word apple, this doesn’t seem to be
such a crazy idea. But what about verbs? For the word eat, we might
also be able to envision in our mind what action is involved, and if our
speaker says to us, I ate an apple, we might see that action occurring
in our minds and thus be able to understand it.
This theory of meaning suggests that when we listen to someone,
we see a stream of pictures in our mind, like watching a film that is
constantly playing. For speech production, this theory claims we rely
178 an open introduction to linguistics
on visual images to link our thoughts to the words we want to say. But
the hypothesis that meaning is related to visual pictures is problematic
when we examine it further. First, how do we know that we are all
using the words to mean the same thing? My experience with apples
may be different than yours, so does that mean that the picture of
an apple I have in my mind will be different than yours? And if that
is the case, will it affect our ability to use language to communicate?
Maybe for apples it won’t matter so much, but what if the meaning
of the word ‘red’ to me is what you call the color ‘orange’? At what
point do divergences in our mental imagery start to impede smooth
communication? Second, how did we learn the meaning of words? We
are not born knowing the words in our language or the concepts that
they apply to, and in fact, depending on the language we are exposed
to as children, we learn different words (and perhaps different concepts,
but more on that later). We know that we learn the meaning of most
concrete words through our interaction with the world: mothers point
to or give their children apples and say apple or ringo or pommes
and children then interact with the world and the object and learn
it. So even if we did rely on mental imagery to understand sentences,
we would still need the world around us to mediate our acquisition of
the words and their associated images. Note though, that this solves Use vs. Mention
part of the problem of determining whether or not different individuals When you use a word with its mean-
ing, we don’t have to mark it in any
use words to mean the same things: because we all live in the same special way. But when we talk about
physical world, and are introduced to words via our interaction with the word itself, we say that we are
mentioning it, and then we need to
that world, we can be reassured that despite different experiences, the signal this. That’s why in the text
meanings we ascribe to many words will be similar among speakers of when I talk about the word “Jan”, I
put quotes around it, or sometimes we
the same language.
use italics to signal mention.
But a visual imagery theory of meaning ignores the fact that we can
learn the meaning of new words by the context in which they occur,
without any visual imagery necessary. If I say to you “I have a pet
wug”, without any imagery you can still understand something about
wugs (that they can be pets, that they are a kind of animal) without
any visual imagery. Clearly the relationships between known words,
and constraints they signal about the meaning of unknown words, are
also part of meaning that doesn’t seem to be handled well with visual
imagery alone.
A visual imagery theory of meaning also ignores the fact that peo-
ple have very strong intuitions about whether or not a given sentence
correctly describes a situation in the world or not. Surely this re-
lationship between the world and language must be relevant for our
understanding of language.
A final problem with a theory of meaning based on mental images
is defining the meaning of abstract words like confidence, nothing, the,
above and loud. What about negation, in sentences like There is no
such thing as a pink elephant or I never eat breakfast. For these and
other reasons (for example, blind people also learn to speak their native
language, so visual imagery is not a necessity for language) this simple
idea was rejected years ago. Hopefully I’ve convinced you that the
idea that visual imagery moderates linguistic meaning becomes more
semantics i: fundamentals of meaning and its representation 179
theory.
In our model, example (1) will be true when the individual named
John (so an individual member) is a member of the set of those who
walked. We often write it in a form of predicate logic, e.g. walked(john).
For example (2), we’ll simply say shoots(han). For (3), we can use con-
nectors from logic to create talks(j) ∧ walks(p) ∨ talks(p) (ignoring for
the moment that we need to specify the scope of each operator).
What I have now been doing is describing, in set-theoretic terms,
how the model of the world has to look for the examples statements
to be true. So actually, in addition to the idea of reference, and the
model organized with set theory, I’m also relying on a theory of truth.
At this point we really should introduce some terminology and be-
come more consistent with how we talk about different things. Sen-
tences, or more precisely, statements, are linguistic strings. They
communicate something (in a context) that we will call a proposi-
tion. We need to make a distinction between the sounds or words
strung together that makes a statement, and the proposition it com-
municates. We can now say something about what makes a good
translation from one language to another: good translations commu-
nicate the same propositions. So the statement in (4) in English and
the statement in Dutch in (5) express the same proposition and will
thus be true in the same models and false in the same models.
The same statement can communicate more than one proposition, be-
cause a statement can be ambiguous.
How many imperial soldiers were captured? There are two possible
interpretations, or propositions, which can be paraphrased as the fol-
lowing:
We can also write rules for a simple copula sentence with an adjec-
tive phrase ascribing a feature to an individual. For example:
(14) The Dutch government has proof that aliens have visited earth.
Compositionality
We’ve talked about the meanings of words, and how they can (at least
for a lot of simple words) be interpreted in set theory, but we haven’t
paid much attention to how this happens. Consider the following two
statements.
These two statements, even though they are made up of the same
three lexical items, actually express very different propositions. That’s
because the subject position, which without further information can
be interpreted as the position to the left of the verb in English, is the
agent that carries out the action. The verb to bite is transitive, so it
subcategorizes for an object argument, and that appears to the right
of the verb in a simple statement. These two examples thus differ
syntactically as to which name is used as a subject and which as an
object. These syntactic differences are meaningful because these two
examples also differ in who is the biter and who is bitten.
This illustrates a further point: that the meaning of a proposition
is not simply derived by adding up the meanings of the component
words, but the way in which they are combined is also relevant to
determining the meaning. The semantics of natural language adheres
to what is called the Principle of Compositionality, first proposed
by the German philosopher, Gottlieb Frege.
In the famous Cantina scene from Star Wars (Episode IV: A New
Hope), Han Solo is sitting across the table from Greedo, and there is
some ambiguity about what happened. Did Greedo shoot and miss, Han Solo and Greedo in the original
making it possible for Han then to shoot? Or did Han simply shoot movie.
first? (17) could be used to describe the two possible situations, but
without a comma or some intonation to disambiguate the syntactic
structure associated with this reading, the sentences is also ambiguous.
The structural ambiguity arises from the difference between conjunc-
tions and disjunctions, and whether you take the conjunction and to
be the main conjunct, or the disjunctive or to be the main conjunct,
even when the truth value of the component propositions is kept the
same. Greedo shot’ and Greedo missed’ is true (represented by ‘1’ in
the table), while Han shot(first)’ is false (represented by ‘0’):
The first sentence (1) probably seems quite silly. It’s clearly a tautol- The morning star.
ogy, because everything is clearly themselves. But it’s also true: the
object in the world picked out by the name ‘Hesperus’ is the same
object picked out by the same name (and in fact this seems to be quite
a basic prerequisite we have to agree on to even have an existential
theory of meaning!). But the second statement, within an existential
theory of meaning, is also a tautology. It’s saying that the object that
is referred to with ‘Hesperus’ is the object referred to by ‘Phosphorus’.
Since both names refer to the planet Venus, it’s again basically saying
that Venus is Venus.
But, this is problematic. Clearly we don’t want our semantic theory
to treat scientific discoveries as tautologies. The Greeks before 350
B.C., who didn’t know that their morning and evening stars were both
Venus, had a different relationship to the same object. They considered
the object by one name to have different features than the same object
when called by another name (because they didn’t know it was the
same object). Frege’s conclusion was then that actually, there must be
more to the meaning of a word than it’s ability to identify objects in
the world. How we relate to the object, what features we ascribe to
it, how we learned about it, how it appears in a certain context, are
an additional part of it’s meaning. In order to deal with this, Frege
added another layer of meaning to each word: it’s intension.
Under an intensional theory of word meaning, words have an ex-
tension, which identifies what they pick out in the world. But words
also have an intension, which describes its features and how it is un-
derstood. This new aspect of meaning lets us analyze Hesperus and
Phosphorus. For the ancient Greeks, these two words have quite dif-
ferent intensions. One appears in the morning, one in the evening, etc.
We know they have the same extension, but the ancient Greeks did
not. When they discovered that both referred to Venus, this specified
the extension.
The introduction of intensions also solves another problem related
to the set-theoretic interpretation of words. Under and extensional
theory, the meaning of the words zombie, elf and ewok are all members
of the empty set, because they are things that do not exist in the
world. There is therefore no difference between them in extensional
meaning. However, an ewok and a zombie have very different features,
and clearly need to be distinguished!
So let’s again take stock. What are the features of our semantic
theory?
• Truth conditions which tell us what our model has to look like for
a proposition to be true or to be false
Entailments
Entailment can be defined formally as follows: A entails B, if and only
if in every situation in which A is true, B is also true. Consider the
semantics i: fundamentals of meaning and its representation 187
following sentences.
This is clearly the case. If (20) is true, given the nature of conjunction,
(21) must also be true.
Now consider the following:
If (22) is true, does it mean that (23) and (24) is true? That does seem
to be the case. In any world where (22) is true, that the individuals
are also dating has to be true. We can then say that (22) entails (using
the symbol for entailment: |= ) examples (23) and (24). Note however,
that entailment is not symmetric (though it sometimes can be): if Han
is dating, that does not mean that Han and Leia are dating. And note
as well, that most people reading (22) will assume that the speaker
means that Han is dating Leia and vice versa. However, the sentence
doesn’t actually say that, and the sentence is also consistent with a
situation where e.g. Han is dating Betty and Leia is dating Barney.
If example (25) is true, then example (26) must also be true. If you
accept (25) as true then the rules of language, the rules of word mean-
ings and how they combine when a conjunction is used, what this then
entails, force you to agree that (26) must also be true. The above
examples showed entailment relationships that arise in part because of
the the conjunction and. Let’s now have a look at some more entail-
ment relationships with adjectives.
Han Solo was born on the planet Corellia. The sentence (27) entails
the simpler, less informative statement in (28). This is the same for
many adjectives. Confirm for yourself that the following statements
entail the simpler statement without the modifying adjective:
that kittens are young cats, and that men have the feature of being
male„ and if someone dies they are dead (zombies and vampires ex-
cluded of course). These are lexical entailments: in all possible worlds
where (38) is true, (39) must also be true. If you don’t accept these
things then you are not speaking English. These have to do with the
basic features of meaning of these words which all native speakers have
learned and accept. So in interpreting any given sentence, there is a lot
of lexical ‘baggage’ that comes with it in the form of the information
we know about how the concepts signalled by the words we use relate
to each other in the world.
Consider the following sentences:
Example (44) is not necessarily true. In fact, the king of the Nether-
lands, William-Alexander, is not a dictator because the Netherlands
is a constitutional monarchy. You might accept the statement Histor-
ically, all kings are dictators, but it is not necessarily true, but true
because of our understanding of the world and the history of our world.
Study Questions
1. Basics of meaning
Consider the following three claims. For each claim, criticize or de-
fend the claim. (State at the beginning which you are doing!) and
then motivate your argument with real examples from natural lan-
guage. Be sure to use correct terminology and mention any relevant
190 an open introduction to linguistics
principles.
2. Entailment
Consider the following sentences
4. Adjective meaning
semantics i: fundamentals of meaning and its representation 191
5. One of the things we have not really touched upon here is how we
assign meaning to words. Implicitly, we have assumed that words
like /John/, /think/ and /yellow/ have meaning because they re-
fer to something in the real world. This idea is called a ‘referential
theory of meaning’. It works well in most cases, but it becomes prob-
lematic when we talk about things that are not part of our world,
such as fictional characters, magical creatures or a Napoleonic vic-
tory in the battle of Waterloo. A solution to this is to distinguish
between extension, i.e. that what is referred to, and intension,
i.e. what is referred to in the mind of the speaker. Now, consider
the following true statements and answer the questions:
Every day may not be good but there is something good in every
day.
-Anonymous
Quantifiers
But this way of interpreting quantifiers doesn’t work well when ex-
tended to the entire range of natural language quantifiers. For exam-
ple, consider most in (6). Can we simply make a quantifier M, and
then treat it similarly to either a universal or an existential quantifier?
The answer is no, but let’s look why. Clearly we need to relate things
(the jedi and the brown-wearers) so we need to use either a conjunction
or an implication.
Let’s consider what interpretations each conjunct would lead to with
most. We’ll consider the truth of (6) in one of three possible models
presented in the Table below.
First, confirm for yourself that the natural language sentence in (6)
is true in each of the three models. Next, we have to have some def-
inition of the quantity meant with most. Let’s take a rather simple
definition and simply say that most means more than half. With this
definition, we can then consider what would happen if most as a quanti-
fier M would be interpreted similarly to how existential quantifiers are
interpreted in predicate logic, relying on conjunction to relate the sets.
This would mean that something like: M x(jedi(x) & wear_brown(x)).
But this means that more than half the individuals in the model are
Jedi that wear brown. In m1 this would be true, but in m2 and m3 , this
sentence would be false, because there are more individuals that are
semantics ii: quantification: beyond reference 195
• no A B = True iff (A ∩ B) = ∅ .
Total Ewoks died in battle survived Table 16.1: Example situations outlin-
ing Ewok casualties
100 20 80
100 40 60
100 49 51
1,000 499 501
1,000,000 200,000 800,000
We can give a definition that simply says that if more than half of the
restrictor set is also a member of the scopal set, then the sentence is
true, and indeed, this is also a very common way to define most in
generalized quantifier theory:
Conservativity
Crosslinguistically, the same inventory of quantifiers seems to be present
in most languages. Further, there are number of semantic features of
quantifiers that we also find cross-linguistically. Because quantifiers in
all languages even those with very different grammatical structures,
have these same features, there are two intriguing possibilities. First,
it could be that we are pre-programmed to have quantifiers with these
features, and this is part of our genetic linguistic endowment. Alter-
natively, it could be that these features are easier to learn, or maybe
sentences with these features easier to verify (determine whether or not
they are true in a given context). One of the most interesting features
is conservativity which we will examine in more detail. Consider the
following definition:
The example (15) is of the form Every N VP. Every N VP is true iff
[[N]] ⊂ [[VP]]. Example (15) is thus true iff [[Ewok]] is a subset of the
set [[danced]], and false if there are members of the set of Ewoks that
are not also members of the set of those who danced. To understand
this sentence, you do not have to know anything about individuals
who danced that were not Ewoks. Every is thus conservative, because
Every Ewok danced means that same thing as Every Ewok is an Ewok
that danced. Let’s consider example (8), (At least five Jedi knit.) This
sentence is true iff the intersection between the set of individuals that
are Jedi, and the set of knitters, has a cardinality of at least five. To
determine whether or not this sentence is true in any given context,
it is not necessary to consider any knitters that are not Jedi. The
example is thus synonymous with At least five Jedi are Jedi who knit.
This means it is conservative.
Now consider the meaning of a quantifier I made up: Korgat, given
in set theory notation below.
Korgat AB is true iff ¬ (B ⊆ A)
Let’s consider the following two pictures while we discuss the mean-
ing of this quantifier.
semantics ii: quantification: beyond reference 199
Example (17) is true w.r.t. Picture B above: only chicks are sledding.
But to confirm this, you have to examine all the members of the set
of sledders, the scopal set. Only is therefore not conservative. As an
adverb, only is not syntactically modifying the noun chicks, either.
Only also associates with focus, a hard to define phonetic feature
in spoken language. Consider example (18). This sentence first limits
interpretation to the chicks, and then says that they are only doing one
activity, as members of the set of sledders, not other activity sets. It
therefore compares activities and singles out which ones the chicks are
doing. That’s quite different from (17), which says that among those
that are sledding, they are all chicks. Example (17) thus zooms in on
200 an open introduction to linguistics
the sledders, and then says that being a sledder only applies to chicks.
Example (17) will be wrong if you have a non-chick sledder, whereas
example (18) will be wrong if you have a chick doing something else
(skiing?).
You can make other differences in meaning even without moving only
around in the sentence (it’s actually an adverb, so it can occur in any
position that an adverb can occur in). Consider the following:
(19) Obi-wan only gave Luke a lightsaber. (He didn’t teach him
how to use it.)
(20) Obi-wan only gave Luke a lightsaber. (He didn’t give one to
Han or anyone else.)
(21) Obi-wan only gave Luke a lightsaber. (He didn’t give him
anything else.)
The quantifiers all and most are not monotone decreasing in their
second argument. They are in fact monotone increasing, which can
be seen in the next examples:
This means we can make inferences from subsets to supersets for the
second argument of quantifiers all and most.
We can write formal definitions of these monotonicity properties:
mon ↓ Det is monotone decreasing in its right argument:=
for all A, B, B’: if B’ ⊆ B, and Det A B, then Det A B’.
(31) All farmers ate steak. |= All Dutch farmers ate steak.
(32) Some farmers ate steak. 6|= Some Dutch farmers ate steak.
(33) Some Dutch farmers ate steak. |= Some farmers ate steak.
(38) No child who has ever seen Claude has smiled at him.
(39) Every child who has ever seen Claude has smiled at him.
(40) *Several children who have ever seen Claude have smiled at
him.
(41) No child has ever smiled at Claude.
(42) *Every child has ever smiled at Claude.
Study Questions
1. If we look closely at the semantics of a sentence, there is often more
than one possible way to interpret it. Nevertheless, communication
problems due to ambiguity in meaning rarely occurs in real life. Can
you think of some reasons for why this is? How does the field of
pragmatics relate to this issue?
2. One of the tasks that researchers within the field of Natural Lan-
guage Processing (NLP) have tried to do automatically is called
paraphrase detection, i.e. the automatic filtering of paraphrases
from a text collection or summary, for example. How could you
use what you know about semantics and propositions to solve this
problem?
3. Just as the sentence /Peter and Sue get married/ has multiple
possible interpretations, so can the sentence /I love Sachertor-
te/. Why, then, do we call the first sentence ambiguous, but not
the second?
The things that exist come in different forms. There are objects,
time points, events, truth values, and so on. These properties are
fundamental. Truth values are different things than events which
in turn differ from objects. What basic elements are necessary
in order to account for meaning? This is a question semantics is
also concerned with.
On What There Is
With language we can seemingly talk about everything that exists.
The range of things that we can talk about is immense. However, this
vast expressive power creates challenges of its own. For to the human
mind, the things around us are not all of the same nature. There are
physical objects, people, places, times, properties, events and so on.
Languages reflect this categorisation of things in different ways. For
example: many languages classify things into various groups, called
genders or (noun) classes. This classification proceeds along several
lines. Many languages distinguish animates from inanimates, for ex-
ample. And within the animate category, make a further distinction
between people and animals, and within the people between the male
and the female, and so on. There is no objective reason for any particu-
lar division. Why should we care to distinguish humans from animals,
or animals from stones? Clearly, the answer is that language is shaped
to serve the needs of humans, not dogs or stones. And furthermore:
language answers to the need of a society and its culture.
The division of things into genders or classes is realised in different
ways in different languages. For some languages it is part of the mor-
phology (French, for example), while for others it is irrelevant (e.g.
Hungarian). People have attached a lot of importance to that dif-
ference; the so-called Sapir-Whorf-Thesis states that the language
we speak has a profound influence on the way we think. The reason-
ing is that, if gender distinctions are obligatory, then we need to pay
more attention to them and consequently consider this a feature worth
attending to.
Nowadays, it is thought that the influence of language on percep-
tion has been overstated. The fact that the Hopi language does not
distinguish past from future does not mean that monolingual Hopi
206 an open introduction to linguistics
speakers cannot tell the difference or are in any way less likely to
understand it. Nevertheless, there are areas where the thesis has its
merits, for example in the categorical perception of sounds. Recall
that distinctions not in your native languages require effort to hear.
Another example is spatial language: it has been claimed that there
are languages where there are only absolute spatial prepositions of the
type /north/, /south/, and that the speakers of these languages are
far better at tracking absolute directions than speakers of languages
which also have things like /left/ and /right/.
Here, we are more concerned with how the classification of things
is relevant for language in general and semantics in particular. The
elements that seem to be necessary are reflected in the basic semantic
types that most semantic theories recognize. We will discuss a few of
them in this chapter.
Paul is eating.
The neighbouring cat is eating. (17.7)
∴ Paul and the neighbouring cat are eating.
A Vs.
A and B V. A and B V.
B Vs. (17.8)
∴ A Vs. ∴ B Vs.
∴ A and B V.
Now, if there are groups of people, are there also groups of groups?
There are! This causes some interesting ambiguities. For example, you
can get married as a group of two people by getting married to each
other. This involves a one-time event that makes you married to each
other. A group of two people can also get married by each marrying
someone else. The sentence (17.9) can be read in these two ways. This
is due to the fact that /get married/ is a verb that can take both
groups and individuals as its subject.
John and Alison and Bill and Sue got married. (17.11)
There is a reading which says that John marries Alison, and Bill mar-
ries Sue. This reading exists in virtue of the following: /John and
Alison/ denotes a group, and so does /Bill and Sue/. The verb
applies to such groups in the sense ‘marrying each other’. Since our
default idea is that of marriage between exactly two people, if several
groups get married to each other, this must means that all groups (of
two) get married separately.
There are also verbs that encourage a group of groups reading:
Time
Properties can change over time. I can be tired in the evening, but the
next morning I might not be tired any more. Some properties have an
inbuilt time dependency. For example, a cat is a kitten only in certain
stages of its life. When it is old enough, by definition, it ceases to be
a kitten, even though it never ceases to be a cat.
Here, we picture time as consisting of a continuum of time points
on a line. We take t and t0 as time points, and we write t < t0 to say
that t is prior to t0 , and t > t0 to say that it is after t0 . For any two
time points t and t0 , either t = t0 , t < t0 or t > t0 . This trichotomy is
reflected in the three basic tenses of English. The present is used for
something that happens now, the past is used for something that has
happened before now, and the future is used for something that will
happen later.
Location
Space is as important in daily experience as time. We grow up think-
ing spatially; how to get furniture through a door, how to catch the
neighbour’s cat, how to not get hit by a car, etc. All these things
require coordination of actions and thinking in time and space. This
is the reason why language is filled with phrases that in one way or
another refer to (points in) space. The most evident expressions are
/here/, which functions in the same way as /now/, and /there/, which
is analogous to /then/. /here/ denotes the space that the speaker is
occupying at the moment of utterance. This concept is also relevant
for the meaning of words such as words /come/ and /go/. If someone
is approaching me right now, I can say:
He is coming. (17.20)
But I would not say that he is going. That would imply he is moving
away from me now. German uses verbal prefixes (/hin/ and /her/)
for a lot of verbs to indicate whether movement is directed towards
the speaker or not.
Unlike time, space is not linear. It is organized differently, and lan-
guage reflects that. Suppose I want to say where a particular building
is located on campus. Typically we phrase this by giving an orienta-
tion and a distance (also known as ‘polar coordinates’). For example,
We have started out by saying that sentences are either true or false.
So, any given sentence such as the following has to be either true or
false:
We can imagine with respect to (17.22) that it is true right now or that
it is false. In fact, we probably do not even know. With (17.23) it is
the same, although if we have learned a little history we will know that
it is true. Still, we might find ourselves thinking ‘what if Napoleon had
actually won the battle of Waterloo ...’. Thus, we are able to imagine
a situation that is different from the actual situation. The technical
term is world. Worlds decide the truth value of every sentence one
way or another. There are worlds in which (17.22) is true and (17.23)
is false, others in which (17.22) is false and (17.23) is true, others in
which both are false, and again others in which both are true. And
one of those worlds is the one (and only) world we live in.
Seemingly then, any sentence can be true or false in some world.
But this is not quite true. Sentences like (17.24) are different, because
they are always true. Of course, you could imagine otherwise, but
this would be tantamount to violating the rules of the language the
sentence is expressed in. If I were to say ‘suppose that kittens are not
young cats but in fact old rats ...’ what I ask you is to change the
way English is understood. Then I am not talking about a different
world, but a different language. Worlds have an independent existence
from the language that is being used. We say then that (17.24) is
necessarily true, just like /4+7=11/.
The denotation of a word like /cat/ in this world is the set of all
beings that are cats. This can change from world to world. We can
imagine a world that has absolutely no cats. (If we go back in time,
there was a time when this was actually true.) Or one that has no
mice. But we do not suppose that, just because there are different sets
of cats in different worlds, the meaning of the word changes—it does
not. That’s why you cannot suppose that kittens are old rats. We say
that the meaning of the word /cat/ is a function cat0 that for each
and every world w gives some set, cat0 (w ). We of course understand
that cat0 (w ) is the set of all cats in w. (Some people use the word
intension (sic) for that function.) Likewise, the intension of the word
/rat/ gives for each world w the set of all rats in w, and likewise for
the word /kitten/. It is a (world-independent) fact of English that:
There are plenty of words that are sensitive not just to their denotation
semantics iii: basic elements of interpretation 211
Events
Both (17.28) and (17.29) can manifest as the same physical activity:
me sitting behind the computer and typing something in. Whether or
not I am fixing the computer by doing so, time will tell. But in theory,
I could be fixing the computer just by typing on the keyboard. The
same goes for writing a book. Thus, one and the same physical activity
can be a manifestation of various different things. In order to capture
this insight, it has been proposed that verbs denote particular objects
called events. There are events of typing, as there are events of fixing
the computer, and events of writing a book. Events are distinct from
the activities that they manifest as.
Events will figure in Chapter 19, so we shall not go into much detail
here. There are a few things worth knowing about events. First, there
are two kinds of events: states and processes. A state is an event
where nothing changes:
The list of participant types is longer, but these ones are some of the
most common ones.
Processes are sometimes meant to finish in some state and some-
times they are not. If you are said to be running, no indication is made
as to when you will stop doing so. If you are said to be running a mile,
then there is an inherent point that defines the end of the activity you
engage in. Some verbs denote the latter kind of events: /arrive/,
/reach/, /pop/, /finish/. The process they denote is finished when
something specific happens.
Study Questions
1. What is the difference between distributive and non-distributive
verbs? What is your natural interpretation of the following verbs,
distributive or non-distributive?
a. fight
b. sleep
c. talk
d. sit
3. How are the terms event, process, state, theme and experiencer re-
lated?
18
Semantics IV: Scope
Scope ambiguities can arise when there are two quantified expres-
sions that can be in different dominance relations. Consider for exam-
ple the following sentences:
Actually, it’s not just quantified expressions, but also other semantic
operators that can enter into scope (read: dominance) relations with
other operators.
Scopal relations can also be entered into between adverbs and quan-
tified expressions, and by quantified phrases and intensional verbs (Ex-
amples from Tunstall, 1981):
(6) Ronny made each of the beds in a short amount of time, but it
took a while to complete the whole task. all the beds > quickly
(7) Sophie thinks a friend will come by tomorrow.
(8) Sophie thinks some friend or other will come by. think > a
friend
(9) There is a particular friend, say Emma, that Sophie thinks will
come by. a friend > think
Negation can enter into different scopal relations with different DPs.
Consider the following example:
Example (11) is ambiguous. It can have the surface scopal reading that
the butler didn’t murder three people, (but perhaps two or one!) or
that there are three people that we can identify, that the butler didn’t
murder. However, the very similar example (12) is not ambiguous as it
does not permit an inverse scope reading, and can only mean that that
the butler did not murder three people (again, maybe two or one).
Formally we can say that if a logical operator Op1 c-commands
another operator, Op2, than the reading where Op1 takes scope over
Op2 is available. That’s the surface scope reading. But the reading
where Op2 takes scope over Op1 is not always available. It turns
out that this relates to monotonicity properties. Non-decreaseing VPs
can usually have inverse scope readings, whereas monotone decreasing
VPs tend not to allow inverse scope readings. This again illustrates
the relevance of monotonicity properties to language.
Different quantifiers therefore have different tendencies to prefer
different scopes. For example, each tends to want to take wide-scope
more than every. In fact we rank different quantifiers in terms of their
preferences for wide scope:
Relative clauses
Let’s consider some other cases where scope makes a difference in in-
terpretation. Look at the difference between (18.1) and (18.2):
The part /which has recently been repaired/ is a clause that func-
tions like an adjective; it is called a relative clause. Unlike adjec-
tives, relative clauses follow the noun. Notice that /we have/ also is
a relative clause, though it is somewhat shortened (we could replace it
/which we have/).
Suppose you go to a car dealer and he utters (18.1). Then he is
basically saying that he has only one car. Moreover, this car has been
under repair recently. Suppose however he utters (18.2). Then he
is only saying that there is a single car that has been under repair
recently, while he may still have tons of others. It is clear from which
dealer you want to buy. To visualize this difference, we indicate the
scope of the operator /the only/:
In the first sentence, /the only/ takes scope over /car we have/. It
has scope over /car we have which has recently been repaired/
in the second sentence. It is not necessary to formalize the semantics
of /the only/. We need only say that something is the only P , if and
only if (a) it is a P , and (b) nothing else is a P . So, if /the only/ takes
scope only over /car we have/, then the car we are talking about is
a car the dealer has (by (a)), and there is no other car that he has (by
(b)). Thus, he has only one car. If the scope is /car we have which
has recently been repaired/, then the car we are talking about
has recently been repaired (by (a)), it is one of the dealer’s cars (also
by (a)), and there is no other car like that in the dealer’s possession.
So, there may be other cars that the dealer has, but they have not
been repaired, and there may be other cars that were not the dealer’s
but that have been repaired.
The difference in structure between the two is signalled by the
comma. If the comma is added, the scope of /the only/ ends there.
218 an open introduction to linguistics
The possessors (Simon, Paul) are DPs themselves, so they are put in
the specifier position of the DP, quite like the determiner /the/. Now, Notice that even though the possessive
the actual structure we are proposing is this: and the determiner cannot co-occur in
English this is not due to the fact that
they compete for the same position.
[DP [DP Paul][D0 [D ’s][NP three recent attacks on squirrels]]] In Hungarian, for example, they can
co-occur:
(18.8)
Mari-nak a cipő-je (18.7)
Now, take a DP which has a non-restrictive relative clause: Mary-dat the shoe-poss3sg
Mary’s shoe
Simon’s only bedtime story, which he listens to (18.9)
Literally translated, this means
carefully Mary’s the shoe her. The possessive
/Marinak/ (in the dative!) occurs
before the actual determiner.
This DP is identical to what we saw earlier, except that we have Notice that the ’s is analysed here as
/Simon’s only/ . In 18.9, /Simon’s only/ takes /bedtime story/ the D-head; there are other possibili-
ties, for example taking it to be a case
as its scope. It can only do so if the relative clause is an adjunct to marker (which I prefer)
the DP /Simon’s only bedtime story/.
Likewise one may wonder about the place of restrictive relative
clauses. It is clear that they can be neither adjuncts to DP nor ad-
juncts to D0 (because then /the only/ cannot take scope over them).
Restrictive relative clauses are therefore adjuncts of either a N0 or a
NP.
There are at least three ways of reading this sentence. Under one
interpretation, it means that there is something that John searches
for, and he thinks that is the holy grail. Under another interpretation,
it means that John is looking for something, which might or might
not be the holy grail, but which John believes to be the holy grail.
Under the third interpretation, it is unclear whether there is a holy
grail at all, but John believes that there is, and he is searching for it.
This particular case is interesting because people are divided over the
issue whether or not the holy grail existed (or exists, for that matter).
Additionally, it is not clear what it actually was. So, John might find
it and not know that he has found it. We may paraphrase the readings
as follows:
John believes that there is such a thing as the holy grail, but it is not
specified whether it actually exists.
Such differences in meaning are quite real. There are people who
do in fact search for the holy grail. Some of them see in it the same
magical object as which it is described in the epics, while others do not
think the object has or ever had the powers associated to it. Rather,
for them the importance is that the knights of King Arthur’s court
possessed or tried to possess it. So, they regard the medieval stories as
myths, that nevertheless tell of something real (as Schliemann believed
that Homer’s Iliad may not have been about the Greeks and their gods,
but at least about the fall of a real city, Troy, one that he later found).
Now, whether or not these differences can be related back to differ-
ences in structures corresponding to (18.14) remains to be seen, since
such a claim is difficult to substantiate.
Study Questions
1. Syntactic, structural differences are a main cause of different inter-
pretations, but they are not the only one. Can you think of some
other causes and give examples?
2. Consider the sentence /I saw the man on the hill with the telescope./.
What are the three possible interpretations of this sentence? Are
these caused by syntactic differences? Hint: think about the position of the
PP /with the telescope/.
3. We have considered three different interpretations of sentence (18.14),
in sentences (18.15)-(18.17). The differences between these inter-
pretations are somewhat difficult to express clearly. Try to give a
clearer expression of the different interpretations, using the notions
of intension and extension (see page 191).
4. A Negative Polarity Item (NPI) is a lexical item that can only occur
in negative contexts. Using this fact and what you already know
about scope, c-commanding and (non-)reflexive pronouns, answer
the following questions:
We have said that nouns denote objects and that verbs denote events.
It has been observed, however, that some categorisations that have
been made in the domain of objects carry over to events and vice
versa. They ultimately relate to an underlying structure that is similar
in both of them.
the plural /mice/ denotes the set of all subsets of M . (Maybe just the
subsets that contain more that one element, but that is not important
here.) Let us then say this: /A is a B/ is true if what /A/ denotes
is an element of the denotation of /B/. Then /Paul is a mouse./ is
true if and only if Paul is a member of M , that is, if Paul is a mouse.
We shall say the same about /A are B/: it is true if and only if what
/A/ denotes is in /B/. Therefore /Paul and Doug are mice./ is true
if and only if the denotation of /Paul and Doug/ is in the set of all
subsets of M , that is, if it is a subset of M . The denotation of /Paul
and Doug/ is a subset of M exactly if both Paul and Doug are mice.
Let us return to the issue of dividing objects of a kind. It seems
clear that the division into smaller and smaller units must stop at some
point. A train cannot be divided into smaller and smaller trains. At
some point, in fact very soon, it stops to be a train. This does not
apply to some other things, e.g. water. Although science tells us that
things are otherwise, in everyday life water is divisible to any degree
we like. We can take an arbitrary quantity of water and divide it as
we please—the parts will still be water. Thus, /water/ is not a count
noun; it is instead called a mass noun.
One problem remains, though. We have not talked about mass
things, but instead we have consistently talked about mass or count
nouns. We said, for example, that /water/ is a mass noun, not that
the substance water itself is a mass substance. In this case, it is easy to
confuse the two. But there are occasions where the two are different.
The word /furniture/, for example, turns out to be a mass noun,
but is not a mass substance, since what it denotes clearly cannot be
indefinitely divided into parts that are also called furniture.
But how do we know that something is a mass noun if we cannot
ultimately rely on our intuitions about the world? There are a number
of tests that establish this. First, mass nouns do not occur in the plu-
ral. We do not find ∗ /furnitures/ or ∗ /courages/. On the surface,
/waters/ seems to be an exception. However, in this case, /water/
is used in a different sense, where it denotes clearly defined patches of
water, such as rivers or lakes. A better test is this one: mass nouns
freely occur with so-called classifiers, while count nouns do not, for
example:
a glass of water (19.1)
a piece of furniture (19.2)
∗ ∗
a glass/ piece of bus (19.3)
Notice that one does not say ∗ /a glass of furniture/, nor a ∗ /piece
of water/. The kind of classifier that goes with a mass noun depends
on what it actually denotes. Some can be used across the board, like
/a lot/. Notice that classifiers must be used without the indefinite
determiner of the noun. So, with respect to (19.3), we can say /a
piece of a bus/ but then /piece/ is no longer a classifier.
There is an obvious distinction between whether or not objects of
a kind are divisible into objects of the same kind, i.e. whether they
are mass substances, and whether a noun denoting these objects in a
semantics v: cross-categorial parallelism 223
He is drawing. (19.6)
These are all legitimate ways of describing what is or was going on.
Unfortunately, the director has decided to cut a few of the scenes at the
end. So we now have a different film. Now we are not in a position to
truthfully utter (19.5) on the basis of the film any more. This is because
even though what that person began to draw looks like the beginning
of a circle, that circle may actually never have been completed, or
224 an open introduction to linguistics
Here, it does not matter where John started out from, as long it was
somewhere away from the station. We can cut parts of the beginning
of the film, and it will still be a film about John’s going to the station.
Telic events are directed towards a goal (that gave them their name;
in Ancient Greek, ‘telos’ meant ‘goal’.). As it appears, the majority of
non-divisible events are telic. A different one is:
Here, cutting out parts of the film anywhere will result in making
(19.9) false, because John did not write the novel in its entirety.
Now, how do we test for divisibility (= atelicity)? The standard test
is to see whether /for an hour/ is appropriate as opposed to /in an
hour/. Divisible events can occur with /for an hour/, but not with
/in an hour/. With indivisible events it is the other way around:
∗
Alex was drinking in an hour. (19.14)
Alex was drinking for an hour. (19.15)
When combined with a mass noun it remains atelic, but when com-
bined with a count noun (or a non-divisible substance) it is telic.
∗
Alex drank water in an hour. (19.16)
Alex drank water for an hour. (19.17)
Alex drank a beer in an hour. (19.18)
∗
Alex drank a beer for an hour. (19.19)
∗
Alex was drinking beers in an hour. (19.20)
Alex was drinking beers for an hour. (19.21)
the latter reorders the lexical elements. Likewise, the deep phonologi-
cal representation becomes apparent only after we have computed the
surface morphological form. Thus, the parallel model should be re-
placed by a sequential model, which has been advocated for by Igor
Mel’čuk (in his Meaning-to-Text theory). In this model, the levels are
not parallel, but rather they are ordered sequentially. Thus, when we
want to say something, we do this by first organising the semantic
representation, then the words on the basis of that representation and
the syntactic rules, then the morphological representation on the basis
of the lexical, and the phonological on the basis of the morphologi-
cal representation. Listening and understanding involves the reversed
sequence.
An alternative paradigm is that of generative grammar. Generative
grammar assumes a generative process, which is essentially indepen-
dent of all levels. It runs by itself, but it interfaces with the phonology
and the meaning at certain points. The transformations are not taken
to be operations that are actually executed, but rather as ways to or-
ganize syntactic (and linguistic) knowledge. This makes the empirical
assessment of this theory very difficult, because it is difficult to say
what sort of evidence is evidence for or against this model.
Another option is to use other models of syntax, for example ones
that try to eliminate the distinction between deep and surface struc-
ture. In Generalized Phrase Structure Grammar (GPSG), for example,
the question words are generated directly in sentence-initial position;
there simply is no underlying structure that first puts the object ad-
jacent to the verb, only to move it to the beginning of the sentence
later.
There are probably as many theories as there are linguists. But
even though the discussion surrounding the architecture of linguistics
has been very fashionable in the last decades, some things still remain
that are problematic for most theories.
Let’s just consider one very irritating fact. We have seen earlier
that Malay uses reduplication for the plural. This poses a number
of problems: first of all the plural sign has no substance: there is
no actual string that signals the plural (like the English /s/). What
signals the plural is the reduplication. This is essentially a function ρ
that turns a string x into that string concatenated with itself in the
following way:
ρ ( x ) : = xa - a x (19.22)
1.5in
• “in one minute” and “for one minute” can be used with the same
predicates in Japanese
• Suggests these are cognitively basic ideas that are then language
specifically realized
228 an open introduction to linguistics
Study Questions
1. How do the terms mass noun, mass substance, count noun and
count substance relate to each other? Try to define each term.
2. What are telic and atelic verbs? How are these related to the
notions of event, process and state?
4. For each of the following nouns, say whether they are mass or count
nouns and whether they refer to mass or count substances.
a. lego block
b. iron ore
c. jeans
d. cutlery
e. lake
Pragmatics is the study of meaning in context. You can say the same
sentence in different contexts, and the place in time in which you say
it, the people that you say it to, the information that both you and
the hearer share, all influence how the sentence is interpreted. Major
topics include topic structure, focus, speech acts, and presupposition
and conversational implicatures. In this short introduction to prag-
matics we’ll only cover the latter two topics, presuppositions and
conversational implicatures.
Presupposition
Example (1) asserts that John, at the time of the utterance, is not a
smoker. But this sentence also communicates additional information.
It also communicates that John used to smoke at a time earlier than
utterance time. This second piece of information is called a presup-
position. In order to understand the asserted information we have to
first accept the presupposed information as ‘background’ information.
(It would be very odd to say that John stopped smoking if John had
never been a smoker.) Example (2) also has a presupposition. This
sentence asserts that Jayne communicated with John via electronic
mail, and presupposes that Jayne also emailed someone else. Note
that the phonetic focus on JOHN in this example helps remove some
ambiguity, and Jayne EMAILED John too would instead presuppose
that Jayne also did something else to John (but will still assert the
same thing as (2)).
Presupposed information is semantically different from asserted in-
formation, and we see this clearly when we examine how both types
of information behave when in the scope of logical operators. Logi-
cal operators are lexical items that communicate logical functions, like
negation. Consider the negated version of (1):
230 an open introduction to linguistics
1. S (simple),
2. not-S (denial),
3. S? (yes-no question),
5. perhaps-S (possibility).
You should verify for yourself that for each sentence in the S-family
of tests, the presuppositional inference (that John used to smoke at
a time previous to the utterance time) goes through, but that the
assertion (that John does not smoke now) gets modified in all but the
first statement.
pragmatics 231
Presupposition Triggers
(11) John realized Sara was the best programmer in the group.
(12) P: Sara is the best programmer in the group.
(13) P: John exists.
(14) P: Sara exists.
(15) Sara was surprised John noticed her.
(16) P: John noticed Sara
(17) P: John exists.
(18) P: Sara exists.
(19) It was Jack who found the hidden village.
(20) P: someone found the hidden village
(21) P: Jack exists.
(22) P: there is a hidden village
For all of the above examples, you can check if the presupposition is
as claimed by applying tests from the S-family of tests.
232 an open introduction to linguistics
Presuppositional Projection
(29) John doesn’t have a son. But if John had a son, his son is
bald.
(30) ??John has a son. John doesn’t have a son. But if John had a
son, his son is bald.
(31) ??John doesn’t have a son. But if John has a son & if John had
a son, his son is bald.
(32) John doesn’t have a son. But if John had a son, John has a
son and he is bald.
Now let’s look at another type meaning that sentences can commu-
nicate that goes beyond the simple semantic meaning, but that is
based on different principles and behaves differently than presuppo-
sition: conversational implicatures.
Sentences have basic meanings, and this is what semantics studies. But
in actual communicative settings we can use sentences to mean more
than their basic semantic meaning. This is because our expectations
about how the world works, in particular how we expect people to
behave to each other, allows us to get a richer interpretation from
what people say. Much of our communication is therefore not literal,
e.g. someone who says I could eat a horse doesn’t actually mean that
the speaker is able to eat a large animal.
The first linguist to look at this systematically was Paul Grice. He
234 an open introduction to linguistics
Maxim of Quality
Maxim of Relevance
• Be relevant.
pragmatics 235
Maxim of Quantity
Both (41) and (42) illustrate how we can derive additional meaning be-
yond what is actually said by considering how the speaker’s seemingly
indirect response actually might be very informative if the speaker is
following the Maxim of Quantity.
Below are some more examples from the acquisition study of Surian
et al.
Examples (43) and (44) both illustrate the Quantity maxim’s re-
quirement that speakers should be informative. Example (45) illus-
trates the Quantity’s maxim’s requirement that speaker’s avoid re-
dundancy. Prolixity:
Some quantity implicatures are so frequent that many researchers the fact of using too many words and
therefore being boring or difficult to
believe they are no longer calculated on the fly by adults, but their read or listen to: Despite all its ab-
meanings have become conventional. The most well-studied quantity surd prolixity, this is one of the great
books in the English language. (From
implicatures is the some-not all implicature. This is the implicature Cambridge Dictionary Online)
calculated on the basis of the quantity maxim whenever the words
some is used. Consider:
Anyone who has taken an introductory logic course knows that some
semantically seems to just mean at least one. But often in normal
conversation we instead interpret it as meaning ‘not all’, even though
it is compatible with all. This interpretation is analyzed as being the
result of a conversational implicature based on the Maxim of Quantity:
A teacher who utters (46) would say that all the students passed the
exam if that was the case (why not share this amazing success!). Sim-
ilarly, if all the windows have lights on, the speaker should say this. If
they do not, it conversationally implies that not all windows had lights
on.
Maxim of Manner
• Be clear.
• Avoid ambiguity.
• Be orderly.
1. The maxim of Quantity has two parts that require the coop-
erative speaker to say as much but no more than is required
for his particular purposes in the “talk exchange”
2. The maxim of Quality also has two parts, and demands that
the speaker say only what he believes to be true and on that
for which he has sufficient evidence;
So powerful is each of the maxims that at times they vie for the
privilege of explaining the same facts. It is not clear to me, for
example, how one could both be relevant and say either less or
more than is required. In what way is the avoidance of prolixity
different from saying only as much as is required? If a particular
contribution is obscure does it not also lack relevance? While
it is perhaps possible to eliminate some of the redundancy in
the maxims, I feel that the extreme power of the system is in
fact an unavoidable characteristic.
(51) John and Mary went to the party. ?? Actually, John didn’t
go.
(52) John stopped smoking for stoptober. ?? Actually, he never
smoked.
(53) I’m 1000 years old. Actually, I’m just kidding.
(54) lucy: Do you have a car ?
tom: I have a really fast bike. Actually, I also have a car.
(55) Some of the students passed the exam. Actually, all of them.
(56) Dr.Z is almost always on time. Actually, he’s always on time.
I was just kidding.
After presenting the user with the example dialog, the user is
asked to rate the system performance on a Likert scale from 1
to 5 for 4 questions:
3. Are the system responses relevant, i.e., does the system re-
spond with things pertinent to the discussion? (Maxim of
Relation)
It’s helpful to see some example dialogues and how they were rated.
Here are two sample dialogues, with an excerpt from the paper:
Dialog 5 (sample):
Dialog 6 (sample):
Summary
There is more to meaning than the semantic meaning we’ve examined
until now. Words and expressions carry with them prerequisites on
the context in which they are used. One class of these prerequisites is
called presuppositions, a class of inference that distinguishes itself
from other information by being unaffected by logical operators. An-
other type of enriched meaning is that of conversational implicatures.
This meaning is derived from our expectation of what ideal human
communication should be like. Paul Grice identified features of suc-
cessful, normal communication, calling it the cooperative principles.
He further specified the cooperative principle into four maxims. Using
pragmatics 241
Study Questions
Presupposition, Entailment or Conversational Implicature
For all of the sentence pairs (or conversation-sentence pairs) below:
1. Say whether the second sentence (the one labelled (ii)) is an entail-
ment, a presupposition or a conversational implicature of the first
sentence or conversation (i);
2. Explain how you know this (based on what tests, e.g. negation,
deniability).
Answer: Sentence (i) presupposes sentence (ii). You can see this by
negating (i) and seeing that (ii) would still hold.
1. The Doctor has a sonic screwdriver and the Doctor has psychic
paper.
Learning sounds
Newborn children are immediately able to cry. Interestingly, it’s clear
both to them and us that crying is not a meaningful speech sound.
Actually, newborns seem to know without experience that certain dis-
tinctions are not important in human languages. For example, babies
only a few days old do not respond differently if the speaker is male,
female, young, old or if the articulation is louder or softer. Instead
they seem to immediately focus on linguistically relevant contrasts. Two paradigms for studying the lin-
Soon after birth, infants start practicing making speech sounds, us- guistic development of babies are HAS
ing both passive and active articulators. Initially most sounds are (High Amplitude Sucking) and Pref-
erential Looking. Watch the following
vowels, because babies are unable to control the pressure, closing and video about how babies learning
opening necessary to produce plosives. Gradually they gain more con- English develop their recognition
trol, and will begin ‘babbling’, that is: producing all speech sounds of English and non-English speech
sounds. https://www.youtube.com/
present in the world’s languages. At this stage children are also able watch?v=WvM5bqUsbu8
to hear contrasts between sounds that are not present in their own
language. For example, English doesn’t have unaspirated intial conso-
nants, and a native English speaking adult does not hear this contrast
if you play audio files for them. However, Hindi, and many other
languages do have this contrast. Experimental studies of very young
babies perception (under six months) at the universal babbling stage
have found that even children being raised in an English speaking en-
vironment are able to hear the Hindi distinction.
Eventually, at around 9-11 months, babbling becomes language spe-
cific, and children only produce sounds that are present in their native
language. . At this point, children also lose their ability to distinguish Please note that everyone should
between all the sound contrasts in the worlds languages. They instead watch the TED talk from Dr. Patri-
cia Kuhl about comprehension at this
become specialized in hearing relevant contrasts in their own language. stage of language learning.
This changes has an advantage: hearing only relevant distinctions (e.g.
ones that lead to minimal pairs in your language) makes it easier to
learn words.
At about 12 months, children begin to say their first words. Often
the first words is mama or papa. Please note, that both these words
use bilabials. Cross-linguistically, words for mothers and fathers often
are made up of sounds that are easy to see visually, and this seems to
244 an open introduction to linguistics
be related to the fact that the child will be able to see how the parent
produces these words, allowing them to mirror their own production
on the parents. (But do we say mama for mother because that is
just an arbitrary word we associated with mothers, or do we choose to
let mama mean mother because these are sounds that the child (can)
produce?)
Learning words
Children are amazingly good at learning words once they start. Usu-
ally a handful of words are produced around 12 months, but later
children learn multiple new words a day. Keep in mind as well that
when learning words, children are also learning concepts, and how to
categorize the world. Word learning is much more than simply learning
to name things. First, getting the concepts right is sometimes hard.
Second, initially children do not know that many words, yet they still
attempt to communicate, which leads to a number of ‘shortcomings’
(errors?) in their production of words.
For both these reasons, cross-linguistically children make two com-
mon errors in learning the reference of words. Overextension errors
are when a child applies a word too widely, calling all men pappa, or
all meals breakfast, for example. Underextension errors are when chil-
dren interpret a word too specifically, such as only calling chocolate
cupcakes ‘cupcake’ or only their own car is a car.
Even thought very young children know very few words, they are
very adept at making them work for them. At this age children often From Vocabulary development in
use a single word to mean an entire sentence. So cookie can mean Give children with Down syndrome: Longi-
tudinal cross-sectional data (Zampini
me a cookie or Here is your cookie. This is called the holophrastic phase & D’Odorico, 2013)
in word learning, where a single word can be interpreted as a whole Comparison of vocabulary size be-
tween 18-30 months in Typically
(holo-) sentence (phrastic). Developing children in different per-
centiles, compared to children with
Age Active, productive vocabulary Down’s syndrome in the 50 percentile.
15 months 10 words
18 months 15 words
18+ 30-50 words in one month
From this example we can see that children can learn quite com-
plex information about a word, including whether it is a transitive or
intransitive verb, just by hearing it used. This example also shows
that children seem to really be able to ‘fastmap’ verbs even when they
are not provided with any visual cues, retaining that knowledge across
a time span of a day or two between the initial presentation of the
verbs and the testing(!). This clearly shows that small children seem
to be exceptionally good at learning new words and word meanings.
But how do they move on from the one word stage to more complex
utterances?
Children were then asked questions like Point to the second green
ball. Note that the NP here, second green ball, is analyzed as the
feature of being second applying to the feature of being a green ball,
because second scopes over green ball in a hierarchical tree structure
(see margin example). However if children interpret it as having a flat,
non-hierarchical structure, just as in e.g. the big red ball, then being
second and being green will both be interpreted as features of the ball,
248 an open introduction to linguistics
as if they were two intersective adjectives taking scope over the ball
only.
For the neutral arrays children and adults correct identified the ball.
(See the arrows). However for the biased array, 54.5 % of responses
were incorrect, with children incorrectly choosing the second ball that
was green. This suggests that the children interpreted second green ball
as the ball that is second and is green, i.e. with a flat, non-hierarchical
structure. Because in the neutral array the second ball was red, this
interpretation was not possible.
These results were taken as evidence that children do not under-
stand hierarchical structure. However, children were correct almost
half the time, so this suggests that the error was not really systematic.
For this reason Matthei investigated whether children’s errors could
be explained by the complexity of the syntax used in the sentences
being too complex. Children had to interpret a noun phrase with two
modifiers. If the syntax of the DPs could be simplified they might
Structures assumed in syntax at the
perform better. They did this in Experiment 2, using a single modifier time of Matthei’s experiment. Note
and two different types of objects. You can see an example of a biased the three branches for flat structures.)
experimental item used in Experiment 2 below.
exactly what Crain and colleagues would expect: children gave non-
adult answers because the way in which they were tested was not
appropriate for them.
So a child that says ‘milk’ when being given a glass of milk will
be praised by the caregiver for that utterance. The desire to get that
praise will increase the association between the word milk and the ob-
ject over time, and the child will learn that the vocalization of milk
refers to the white liquid. A child who mispronounces the word papa,
will react to the negative reinforcement that the father will not rec-
ognize and respond, because they didn’t understand what the child
said. This will encourage the child to improve their pronunciation.
For Skinner, and other behaviorists, the child is (compared to adults)
extremely reactive to how linguistic stimulus is reinforced.
One problem with this idea is that it’s hard to see how the learning
can be extended to links between stimuli and responses as the links
become more complex, and more subtle. Let’s examine how this would
then work for language learning.
Setting aside the challenge of extending the account to more com-
plex and abstract words, and giving a more detailed account of the
actual learning process, we can instead examine its correctness in a
different way. If this account is true, what predictions follow? If chil-
dren are learning to react to linguistic stimuli from examples in the
environment, via some stimulus-outcome system, then they:
Consider also the following dialogue between Pien (4;10) and her mother,
where Pien uses an incorrect regularized form of the past-tense verb,
eette (‘*eated’), instead of heb opgegeten:
The children could not have learned these sequences or these forms
from an adult. In fact, we see systematic errors in children’s speech at
similar ages across languages, so this suggests that there is a blueprint
to the acquisition process. In fact, the regularized form of the past-
tense error, like the form that Pien made in the dialogue above, is
a process seen in most languages: cross-linguistically, children often
make errors in acquiring distinctions between regular and irregular
morphological forms. These examples all suggest that forms cannot A U-shaped learning curve that shows
accuracy for irregular forms in En-
be simply memorized from what children hear from adult speakers.
glish. At time 1, children have simply
Let’s look at the case of the English past-tense morpheme in more memorized all forms, and produce
detail. This example is very important not just in linguistics but also them correctly because of this. At
time 2, accuracy drops, because
in cognitive science. A large body of evidence suggests that children children begin to incorrectly apply a
don’t simply memorize the past-tense forms in English. past-tense formation rule meant for
regular forms. At time 3, children
First recall that we can distinguish between two different types of
have realized that irregular forms
past-tense forms: regular and irregular. In English, and other lan- differ from regular forms, and they
guages as well, young children often go through a phase where they reach the adult stage.
mom: nee, je zegt het bord (no, you say the(het) plate.)
eva: mag ik de bord? (can I have the(de) plate?)
mom: het bord (the(het) plate)
eva: het bord. mag ik nou de bord? (the(het) plate. Can I have
the(de) plate now?)
These examples, and many more, suggest that children are not sen-
sitive to explicit negative evidence. Given the infrequency of caretakers
correcting children, for correction to play a major role in language ac-
quisition we would have to assume that children are very sensitive to
implicit corrections, but that seems rather unlikely.
(4) Marie n’a pas invité Léa ou Jean à dîner. Marie has not invited
Lea or Jean for dinner
(5) Mary didn’t invite Lucy or she didn’t invite John for dinner.
(6) ??Neither Lucy nor John were invited to dinner by Mary.
As you can see, the English statement (1) is ambiguous. The disjunc-
tion can either take narrow scope ((3)) or wide scope ((2)) over the
negation, leading to two possible readings. Other languages like Ger-
man and Korean also allow both scopal interpretations. However, in
French, only the interpretation where the disjunction has wide scope
over the negation is natural. An interpretation where the negation
gets narrow scope, e.g. (6), is highly degraded. The same pattern
is found in Turkish, Russian, Mandarin and Japanese. Strikingly, all
children begin with the assumption that their language allows both
scopal interpretations. Acquisitionists who believe in innate structures
therefore see this as evidence that children must have initial universal
assumptions, in their LAD, that all scopal relations are possible.
Further evidence comes from children’s errors with wh-questions.
One common error in young children is a failure to invert the subject
and verb. Consider the following example in English and Italian:
der. Situation (b) this would correspond to a child who makes some
errors in question formation, but can form some questions correctly.
Situation (c), where the child’s hypothesis set H is a subset of the adult
target hypothesis T, is the case where the child correctly uses a subset
of target forms, never using an incorrect form. But this child does not
yet produce the full range of questions. Finally, in situation (d), H is a
superset of T. In this case the child uses some forms correctly, but then
also produces incorrect forms. Pinker argued that it would be impos-
sible to go from the superset situation (d) to the correct hypothesis,
because children do not receive or respond to negative evidence. This
is called the superset problem. But only negative evidence in the
form of corrections would inform at child in situation (d) that they are
incorrectly assuming a superset of the target grammar. Because this
is a logically possible situation, Pinker and others have argued that
the only way to explain how children always converge on the target
adult system is to assume that there is innate knowledge in the form of
constraints on the form of grammar that prevents children from ever
ending up in the superset situation in the first place. This view of
language also makes some additional predictions. If the innateness of
language is there in part to prevent children from ending up in a su-
perset grammar, then we should predict that it will encourage children
to always assume a subset grammar. From a subset grammar, positive
evidence in the form of what structures adults around the child use,
are sufficient to extend the child’s hypothesis until it is the same as the
target grammar. This then predicts that if there are variations within
forms cross-linguistically, then children will initially seem to assume
the form that is a subset of the other form, regardless of the correct
forms in the adult language in that particular language.
What does recursion look like when sentential constituents are embed-
ded in other sentence constituents? consider the following examples: You can watch a video from the
hit television series ‘Friends’ where
(11) Joey knows Monica and Chandler are in love. they use multiple recursive em-
beddings here: https://youtu.be/
(12) Rachel knows that Joey knows that Monica and Chandler are LUN2YN0bOi8?t=1m14s
in love.
(13) Monica knows that Rachel knows that Joey knows that Monica
and Chandler are in love.
260 an open introduction to linguistics
Syntax will allow this to go on for ever; the only limits come from our
processing abilities.
Because recursion is such an important aspect of language, it is also
a candidate for innateness. So an important question to ask is: what
do children know about recursion? Interestingly, young children do not
seem to understand recursive structures. For example, in the Childes
database we find evidence of children’s comprehension difficulties (from
Roeper, 2011):
After being introduced to cookie monster and his sister, the exper-
imenter asked children “Can you show me Cookie Monster’s sisters’
picture?”. The correct answer is of course picture A, but one third of
the children actually chose picture B, e.g. “cookie monster and sister”.
If very young children do not understand recursion, does this mean
that it cannot be innate? Not necessarily. Many biologically deter-
mined developmental milestones only appear later in children’s devel-
opment, e.g. molars, puberty. Language may also have innate aspects
that are not initially present. Further, to be feasible, a theory of lan-
guage acquisition must also incorporate realistic assumptions about
children’s executive functioning, e.g. working memory, ability to sup-
press distracting information, etc. This means that there may be as-
pects of language development like the development of the ability to
deal with recursive structures that are dependent on processing abili-
ties, and thus recursion may not be initially available because children
lack e.g. sufficient working memory.
Summary
Exercises
1. L1 acquisition and Nativism (8 pts)
Consider the following dialogue and answer the questions below.
child: A ball!
(a) How do you explain the use of the word ‘ball’ to refer to ‘apple’?
(b) How do Nativists believe that children learn their native lan-
guages?
(c) Give two arguments that support the Nativist view of language
learning.
(d) How do non-Nativists believe that children learn their native
languages?
(e) Would non-Nativists use the dialogue above as proof to support
their view of how children learn their native languages? Explain
your answer.
(f) The speech of the child in the example above is also character-
istic of a particular stage of U-shaped learning. Explain what
U-shaped learning is, and in which stage of U-shaped learning
this child is.
(g) Is there any evidence that this dialogue is not authentic?
2. Stages in Acquisition
3. Vocabulary acquisition
In early diary studies, where parent linguists recorded their own
children’s speech, we can see evidence of a common process in early
referent learning discussed in the chapter. The table below presents
the words that children used initially for the first referent, and then
what they used that word to refer to later in subsequent references:
(a) What was the purpose of Matthei’s experiments? I.e. what was
he investigating ?
(b) Let’s say you are going to recreate Matthei’s experiment. What
kind of question would you ask children to answer given the array
of objects above?
language acquisition 263
(c) What answer would you expect from adults who have full gram-
matical understanding? (identify the animal by letter).
(d) What answer would you expect many of the children to give
if they responded similarly to the children in Matthei’s study?
(Identify the animal by letter).
(e) What did Matthei conclude from his results?
(f) Can we be sure of this conclusion? Elaborate on any potential
alternative explanations.
22
Language Families and History of Languages
This list can be made a lot longer, and after doing so, it turns out
that the correspondences between languages are largely systematic.
It can be observed, for example, that word-initial [p] in Dutch and
English corresponds to German [pf], that initial [s] becomes [S] before
[t] and [p], and so on. This has lead to two things: the postulation of
a language (of which we have no record!), called Germanic, a set of
words together with morphology and syntax for this language, and a
set of rules which show how the language developed into its daughter
languages. In the present case, the fact that Dutch [p] corresponds to
266 an open introduction to linguistics
Often, one simply writes ∗ p > pf for the sound change. The similarity
to phonological rules is not accidental; there, they were taken to mean
a sequential process, a development from a sound to another in an
environment. Here a similar interpretation is implied, only that the
time span in which this is supposed to have taken place is much longer,
approximately two thousand years!
The list of Germanic languages is long. Apart from the ones just
listed, Danish, Swedish, Norwegian, Faroese, Icelandic, Frisian and
Gothic also belong to that group. Gothic is interesting because it is
a language of which we only have written records; we do not know
exactly how it sounded.
As with the Germanic languages, similarities can be observed be-
tween French, Spanish, Italian, Romanian and Portuguese, too. In
fact, all these languages come from a language which we know very
well: Latin. The development of Latin into these languages is well
documented, compared to other language groups. This is important,
since it allows to assert the existence of a parent language and changes
with certainty, whereas in most cases the parent language has to be
constructed from the daughter languages.
This is so, for example, with Celtic, from which Gaelic, Irish,
Welsh, and Breton descended. Cornish and Manx are also Celtic, but
became extinct in the 19th and 20th century, respectively. Another This is not the end of the story for
Celtic language, Gaulish, was spoken in the whole of France, but it Manx and Cornish, however. In the
late 20th century, both langauges have
was completely superseded by Latin. The only records we have of been revived by language enthusiasts.
Gaulish are in names of people and places. For example, we know of Speakers of the ‘resurrected’ forms of
both languages number in the hun-
the Gaulish king Vercingetorix through the writings of Caesar. The dreds, and nowadays, Manx and Cor-
name is a Celtic name for sure. nish even have some native speakers
Throughout the 19th century, it became apparent that there were again!
with people that lived in the region of the Balkan and the Ukraine in
the 5th millennium BC, some believe they originate further north in
Russia, while others equate them with the Kurgan culture, 4400–2900
BC, in the south of Russia (near the Caspian sea). From there, they
are believed to have spread into the Indian subcontinent, Persia and
large parts of Europe. The first to arrive in central Europe and Britain
were the Celts who established a large empire only to be topped by
the Romans and later by the Germans.
unaspirated aspirated
voiceless voiced voiceless voiced
velar ∗k ∗g ∗ kh ∗ gh
∗ ∗ ∗ k̂h ∗ ĝh
(22.4)
palatal k̂ ĝ
apico-dental ∗ t ∗d ∗ th ∗ dh
labial ∗p ∗b ∗ ph ∗ bh
English /ten/.
Indo-European is believed to have had not only singular and plural
verb conjugations, but also a dual (for two people). The dual was lost
in Latin, but retained in Greek and Sanskrit. Let’s look at an example,
268 an open introduction to linguistics
How Do We Know?
The reconstruction of a language which is no longer spoken (‘extinct’)
is a difficult task. There are two main methods: comparison between
languages, and internal reconstruction. The latter is used in absence of
comparative evidence. In that case, one observes certain irregularities
in the language and proposes an explanation by referring to possible
earlier developments in the language.
It has been observed, for example, that irregular inflection is older
than regular inflection. For example, in English there are plurals in
/en/ (/oxen/, /vixen/) and plurals formed by vowel change (/women/,
/mice/). These are explained by internal reconstruction as reflecting
an earlier state of the language where the plural was formed by addition
of /en/ and vowel change. The plural /s/ was a later development.
Likewise, this method predicts that the comparative in English was
once formed using /er/ and /est/, but at some point was (partly) re-
placed by forms involving /more/ and /most/. In both cases, German
reflects the earlier stage of English.
Notice, however, that the proposed changes are based on present-
day English. How, then, can we make certain that we are right?
First and foremost, there are written documents. We have transla-
tions of the bible into numerous languages, including medieval Geor-
gian (a Caucasian language), Old English (King Alfred’s bible), and
Gothic, for example. Latin and Greek literature has been preserved
to this day thanks to the effort of thousands of monks in monaster-
ies (copying was a very honourable and time consuming task in those
days). Other languages that have been preserved are, among others,
Avestan, Sanskrit and Hittite (written mostly in cuneiform). Most
language families and history of languages 269
other languages were written down only from the early middle ages
onwards, mostly in the form of biblical texts and legal documents.
Now, this provides us with the written languages, but it does not
tell us how they were spoken. In the case of Hittite, the problem is
very obvious: the writing system was totally different from ours and it
had to be discovered how it was to be read. For Sanskrit we know from
the writings of the linguists of those days (among which Pan.ini (500
BC) is probably one of the latest) how the language was spoken. This
is because they gave us explicit descriptions of how the sounds were
produced. For Latin and Greek matters are less easy. The Greeks, for
example, did not realize the distinction between voiced and voiceless
consonants (they knew they were different but could not say what the
difference was). In the middle ages all learned people spoke Latin, but
the Latin they spoke was very different from classical Latin, both in
vocabulary and pronunciation. By that time, people did not know how
things were pronounced in the classical times. So how come we know?
One answer is: through mistakes people made when writing Latin.
Inscriptions in Pompeii and other sites provide telling examples. One
specific example is that /m/ after vowels was either completely lost
or it just nasalized the preceding vowel. This can be inferred from
inscriptions where one finds /ponte/ in place of what should have
been /pontem/. People who made the mistakes simply could not hear
the difference. (Which is not to say that there is none; only that it
was too small to be noticeable.) Also, in verse dating from that time
the endings in vowel plus /m/ counted as non-existent for the metre.
(This in turn we know for sure because we know what the metre was.)
All this is strong evidence that in classical times the final /m/ was not
pronounced.
Another method is looking at borrowings into other languages. The
name /Caesar/ (and the title that derived from it) was borrowed into
many languages, and appears in the form of /Kaiser/ in German,
taken from Gothic /kaisar/, and /césar/ [seza:r] in French. So, at
the time the Goths borrowed the word the letter /c/ was pronounced
[k]. And since French descends from Latin we must conclude that the
[k]-pronunciation is older.
A third source is the alphabet itself. The Romans did not distin-
guish /v/ from /u/. They wrote /v/ regardless. This shows that the
two were not felt to be distinct. It is unlikely that /v/ was pronounced
[v] (as in English /vase/). Rather, it was originally a bilabial approxi-
mant (the non-syllabic ∗ w mentioned above), and became a labiodental
fricative only later.
A problem is caused by the fact that languages consist of tens
of thousands of words, but a lot of them are not indigenous words.
Many words that we use in the scientific context, for example, come
from Latin and/or Greek. Moreover, they may have been borrowed
from these languages at any moment in time. The linguistic terms
(/phoneme/, /lexeme/) is a telling example. These words have been
artificially created from Greek source words, so learned words like these
have to be discarded.
270 an open introduction to linguistics
other words come into sight: Latin /gustare/ ‘to taste’ (from which,
via French, English got /disgusting/), Greek /geuomai/ ‘I taste’,
Sanskrit /juśati/ ‘he likes’, Old Irish /do-goa/ ‘to choose’—they all
look like cognates to this one. Indeed, from all these words one has
reconstructed the Indo-European root ∗ /geus/ ‘to choose’. The Latin
word presents the zero-grade ∗ /gus/, and in West Germanic, we find Roots typically had /e/. The zero
grade is the version of the root that
/kuzi/ ‘vote, choice’.
has no /e/. Similarly, ∗ /dik/ is the
But do not think that historical linguistics is always this straight- zero grade of ∗ /deik/ ‘to show’.
forward. Sometimes it can be tricky and deceiving. For example,
French /choisir/ ‘to choose’ does not come from Latin /gustare/
(although it does come from the same Indo-European root)—there is
no way to explain this with known sound laws. Known sound laws
predict /gustare/ to develop into /goûter/, and this is what we find.
Instead /choisir/ was taken from—Gothic! Indeed, French has taken
loanwords from Germanic, being occupied/inhabited in large parts by
Germanic tribes. The name /France/ itself derives from the name of
a tribe: /Frankon/.
With respect to Dutch and German, the reconstruction is actually
easy, since the languages split around the 17th century. Certain di-
alects of North Germany still have [p] where others have [pf]. Often,
it happens that a certain word is not attested in all languages of a
certain group. For example, English /horse/ corresponds to Dutch
/paard/ and German /Pferd/, with same meaning. The sound laws
allow to assert that /paard/ and /Pferd/ descend from the same word,
but for English /horse/ this is highly implausible. There is a German
word, /Ross/ ‘horse’, and a Dutch /ros/, which sound quite simi-
lar, but they are far less frequent. What we have to explain first is
why the words (/horse/, /ros/, /Ross/) are different. Now, we are
lucky to have sources confirming that the word was sometimes spelt
as /hros/ and sometimes as /ros/ in Old German. In Icelandic, it
is still /hross/. The loss of /h/ before /r/ is attested in other cases,
like the word /ring/. But the change happened in English, too, and
the only reason why the /h/ was preserved in /horse/ is that the /r/
changed places with /o/.
Second, where did Dutch and German get their main words from?
Both words come from the same source, but it is not Germanic, it is
Medieval Latin /paraveredus/ ‘post-horse’ (200 - 900 AD), which in
turn is /para/ + /veredus/. /para/ comes from Greek (!) /para/
‘aside’, and /veredus/ is Celtic. It is composed from /ve/ ‘under’
and /reda/ ‘coach’. In Welsh, there is a word /gorwydd/ ‘horse’. So,
/paraveredus/ had a more special meaning: it was the horse that
was running on the side. (The coach had two horses, one on the
right and one on the left; the one on the left was saddled, and the
one on the right was the ‘side horse’.) English has a word /palfrey/,
which denotes a kind of riding horse. The Indo-European root that has
been reconstructed is ∗ ekw os, compare Latin /equus/, Greek /hippos/
and Sanskrit /aśvah ./. Had we not known that the Latin word is
/equus/, we would have had to guess from French /cheval/ and Span-
ish /caballo/.
272 an open introduction to linguistics
Dutch, 16 Kham, 14
open, 59 Xhosa, 15
penultimate, 65
weight, 66 Yiddish, 20
Syllable Equivalence, 76
Syllable Structure I, 59 zero grade, 271
Syllable Structure II, 61 Zulu, 15
syntactic representation
deep, 144
surface, 144
tense, 115
tenseness, 20
tension, 38
theme, 212
tier, 60
topicalisation, 143
trace, 171
trachea, 12
transfix, 95
transformation, 144
tree, 107
root, 107
two word stage, 247
UR, 75
V2, 146
value, 36
value range, 36
verb, 115
distributive, 207
transitive, 116
verb second, 146
vocal cords, 12
Voice Agreement Principle, 70
voiced, 13
voiceless, 13
voicing, 13
VOT, 22, 28
vowel chart, 19
vowels, 18