19
19
All
rights reserved. Draft of September 21, 2021.
CHAPTER
Sometime between the 7th and 4th centuries BCE, the Indian grammarian Pān.ini1
wrote a famous treatise on Sanskrit grammar, the As.t.ādhyāyı̄ (‘8 books’), a treatise
that has been called “one of the greatest monuments of hu-
man intelligence” (Bloomfield, 1933, 11). The work de-
scribes the linguistics of the Sanskrit language in the form
of 3959 sutras, each very efficiently (since it had to be
memorized!) expressing part of a formal rule system that
brilliantly prefigured modern mechanisms of formal lan-
guage theory (Penn and Kiparsky, 2012). One set of rules
describes the kārakas, semantic relationships between a
verb and noun arguments, roles like agent, instrument, or
destination. Pān.ini’s work was the earliest we know of
that modeled the linguistic realization of events and their
participants. This task of understanding how participants relate to events—being
able to answer the question “Who did what to whom” (and perhaps also “when and
where”)—is a central question of natural language processing.
Let’s move forward 2.5 millennia to the present and consider the very mundane
goal of understanding text about a purchase of stock by XYZ Corporation. This
purchasing event and its participants can be described by a wide variety of surface
forms. The event can be described by a verb (sold, bought) or a noun (purchase),
and XYZ Corp can be the syntactic subject (of bought), the indirect object (of sold),
or in a genitive or noun compound relation (with the noun purchase) despite having
notionally the same role in all of them:
• XYZ corporation bought the stock.
• They sold the stock to XYZ corporation.
• The stock was bought by XYZ corporation.
• The purchase of the stock by XYZ corporation...
• The stock purchase by XYZ corporation...
In this chapter we introduce a level of representation that captures the common-
ality between these sentences: there was a purchase event, the participants were
XYZ Corp and some stock, and XYZ Corp was the buyer. These shallow semantic
representations , semantic roles, express the role that arguments of a predicate take
in the event, codified in databases like PropBank and FrameNet. We’ll introduce
semantic role labeling, the task of assigning roles to spans in sentences, and selec-
tional restrictions, the preferences that predicates express about their arguments,
such as the fact that the theme of eat is generally something edible.
1 Figure shows a birch bark manuscript from Kashmir of the Rupavatra, a grammatical textbook based
on the Sanskrit grammar of Panini. Image from the Wellcome Collection.
2 C HAPTER 19 • S EMANTIC ROLE L ABELING
In this representation, the roles of the subjects of the verbs break and open are
deep roles Breaker and Opener respectively. These deep roles are specific to each event; Break-
ing events have Breakers, Opening events have Openers, and so on.
If we are going to be able to answer questions, perform inferences, or do any
further kinds of semantic processing of these events, we’ll need to know a little more
about the semantics of these arguments. Breakers and Openers have something in
common. They are both volitional actors, often animate, and they have direct causal
responsibility for their events.
thematic roles Thematic roles are a way to capture this semantic commonality between Break-
agents ers and Openers. We say that the subjects of both these verbs are agents. Thus,
AGENT is the thematic role that represents an abstract idea such as volitional causa-
tion. Similarly, the direct objects of both these verbs, the BrokenThing and OpenedThing,
are both prototypically inanimate objects that are affected in some way by the action.
theme The semantic role for these participants is theme.
Although thematic roles are one of the oldest linguistic models, as we saw above,
their modern formulation is due to Fillmore (1968) and Gruber (1965). Although
there is no universally agreed-upon set of roles, Figs. 19.1 and 19.2 list some the-
matic roles that have been used in various computational papers, together with rough
definitions and examples. Most thematic role sets have about a dozen roles, but we’ll
see sets with smaller numbers of roles with even more abstract meanings, and sets
with very large numbers of roles that are specific to situations. We’ll use the general
semantic roles term semantic roles for all sets of roles, whether small or large.
19.2 • D IATHESIS A LTERNATIONS 3
It turns out that many verbs allow their thematic roles to be realized in various
syntactic positions. For example, verbs like give can realize the THEME and GOAL
arguments in two different ways:
4 C HAPTER 19 • S EMANTIC ROLE L ABELING
These multiple argument structure realizations (the fact that break can take AGENT,
INSTRUMENT , or THEME as subject, and give can realize its THEME and GOAL in
verb either order) are called verb alternations or diathesis alternations. The alternation
alternation
dative we showed above for give, the dative alternation, seems to occur with particular se-
alternation
mantic classes of verbs, including “verbs of future having” (advance, allocate, offer,
owe), “send verbs” (forward, hand, mail), “verbs of throwing” (kick, pass, throw),
and so on. Levin (1993) lists for 3100 English verbs the semantic classes to which
they belong (47 high-level classes, divided into 193 more specific classes) and the
various alternations in which they participate. These lists of verb classes have been
incorporated into the online resource VerbNet (Kipper et al., 2000), which links each
verb to both WordNet and FrameNet entries.
that the argument can be labeled a PROTO - AGENT. The more patient-like the proper-
ties (undergoing change of state, causally affected by another participant, stationary
relative to other participants, etc.), the greater the likelihood that the argument can
be labeled a PROTO - PATIENT.
The second direction is instead to define semantic roles that are specific to a
particular verb or a particular group of semantically related verbs or nouns.
In the next two sections we describe two commonly used lexical resources that
make use of these alternative versions of semantic roles. PropBank uses both proto-
roles and verb-specific semantic roles. FrameNet uses semantic roles that are spe-
cific to a general semantic idea called a frame.
The PropBank semantic roles can be useful in recovering shallow semantic in-
formation about verbal arguments. Consider the verb increase:
(19.13) increase.01 “go up incrementally”
Arg0: causer of increase
Arg1: thing increasing
Arg2: amount increased by, EXT, or MNR
Arg3: start point
Arg4: end point
A PropBank semantic role labeling would allow us to infer the commonality in
the event structures of the following three examples, that is, that in each case Big
Fruit Co. is the AGENT and the price of bananas is the THEME, despite the differing
surface forms.
(19.14) [Arg0 Big Fruit Co. ] increased [Arg1 the price of bananas].
(19.15) [Arg1 The price of bananas] was increased again [Arg0 by Big Fruit Co. ]
(19.16) [Arg1 The price of bananas] increased [Arg2 5%].
PropBank also has a number of non-numbered arguments called ArgMs, (ArgM-
TMP, ArgM-LOC, etc.) which represent modification or adjunct meanings. These
are relatively stable across predicates, so aren’t listed with each frame file. Data
labeled with these modifiers can be helpful in training systems to detect temporal,
location, or directional modification across predicates. Some of the ArgM’s include:
TMP when? yesterday evening, now
LOC where? at the museum, in San Francisco
DIR where to/from? down, to Bangkok
MNR how? clearly, with much enthusiasm
PRP/CAU why? because ... , in response to the ruling
REC themselves, each other
ADV miscellaneous
PRD secondary predication ...ate the meat raw
NomBank While PropBank focuses on verbs, a related project, NomBank (Meyers et al.,
2004) adds annotations to noun predicates. For example the noun agreement in
Apple’s agreement with IBM would be labeled with Apple as the Arg0 and IBM as
the Arg2. This allows semantic role labelers to assign labels to arguments of both
verbal and nominal predicates.
19.5 FrameNet
While making inferences about the semantic commonalities across different sen-
tences with increase is useful, it would be even more useful if we could make such
inferences in many more situations, across different verbs, and also between verbs
and nouns. For example, we’d like to extract the similarity among these three sen-
tences:
(19.17) [Arg1 The price of bananas] increased [Arg2 5%].
(19.18) [Arg1 The price of bananas] rose [Arg2 5%].
(19.19) There has been a [Arg2 5%] rise [Arg1 in the price of bananas].
Note that the second example uses the different verb rise, and the third example
uses the noun rather than the verb rise. We’d like a system to recognize that the
19.5 • F RAME N ET 7
price of bananas is what went up, and that 5% is the amount it went up, no matter
whether the 5% appears as the object of the verb increased or as a nominal modifier
of the noun rise.
FrameNet The FrameNet project is another semantic-role-labeling project that attempts
to address just these kinds of problems (Baker et al. 1998, Fillmore et al. 2003,
Fillmore and Baker 2009, Ruppenhofer et al. 2016). Whereas roles in the PropBank
project are specific to an individual verb, roles in the FrameNet project are specific
to a frame.
What is a frame? Consider the following set of words:
reservation, flight, travel, buy, price, cost, fare, rates, meal, plane
There are many individual lexical relations of hyponymy, synonymy, and so on
between many of the words in this list. The resulting set of relations does not,
however, add up to a complete account of how these words are related. They are
clearly all defined with respect to a coherent chunk of common-sense background
information concerning air travel.
frame We call the holistic background knowledge that unites these words a frame (Fill-
more, 1985). The idea that groups of words are defined with respect to some back-
ground information is widespread in artificial intelligence and cognitive science,
model where besides frame we see related works like a model (Johnson-Laird, 1983), or
script even script (Schank and Abelson, 1977).
A frame in FrameNet is a background knowledge structure that defines a set of
frame elements frame-specific semantic roles, called frame elements, and includes a set of predi-
cates that use these roles. Each word evokes a frame and profiles some aspect of the
frame and its elements. The FrameNet dataset includes a set of frames and frame
elements, the lexical units associated with each frame, and a set of labeled exam-
ple sentences. For example, the change position on a scale frame is defined as
follows:
This frame consists of words that indicate the change of an Item’s posi-
tion on a scale (the Attribute) from a starting point (Initial value) to an
end point (Final value).
Some of the semantic roles (frame elements) in the frame are defined as in
core roles Fig. 19.3. Note that these are separated into core roles, which are frame specific, and
non-core roles non-core roles, which are more like the Arg-M arguments in PropBank, expressing
more general properties of time, location, and so on.
Here are some example sentences:
(19.20) [I TEM Oil] rose [ATTRIBUTE in price] [D IFFERENCE by 2%].
(19.21) [I TEM It] has increased [F INAL STATE to having them 1 day a month].
(19.22) [I TEM Microsoft shares] fell [F INAL VALUE to 7 5/8].
(19.23) [I TEM Colon cancer incidence] fell [D IFFERENCE by 50%] [G ROUP among
men].
(19.24) a steady increase [I NITIAL VALUE from 9.5] [F INAL VALUE to 14.3] [I TEM
in dividends]
(19.25) a [D IFFERENCE 5%] [I TEM dividend] increase...
Note from these example sentences that the frame includes target words like rise,
fall, and increase. In fact, the complete frame consists of the following words:
8 C HAPTER 19 • S EMANTIC ROLE L ABELING
Core Roles
ATTRIBUTE The ATTRIBUTE is a scalar property that the I TEM possesses.
D IFFERENCE The distance by which an I TEM changes its position on the scale.
F INAL STATE A description that presents the I TEM’s state after the change in the ATTRIBUTE’s
value as an independent predication.
F INAL VALUE The position on the scale where the I TEM ends up.
I NITIAL STATE A description that presents the I TEM’s state before the change in the AT-
TRIBUTE ’s value as an independent predication.
I NITIAL VALUE The initial position on the scale from which the I TEM moves away.
I TEM The entity that has a position on the scale.
VALUE RANGE A portion of the scale, typically identified by its end points, along which the
values of the ATTRIBUTE fluctuate.
Some Non-Core Roles
D URATION The length of time over which the change takes place.
S PEED The rate of change of the VALUE.
G ROUP The G ROUP in which an I TEM changes the value of an
ATTRIBUTE in a specified way.
Figure 19.3 The frame elements in the change position on a scale frame from the FrameNet Labelers
Guide (Ruppenhofer et al., 2016).
Recall that the difference between these two models of semantic roles is that
FrameNet (19.27) employs many frame-specific frame elements as roles, while Prop-
Bank (19.28) uses a smaller number of numbered argument labels that can be inter-
preted as verb-specific labels, along with the more general ARGM labels. Some
examples:
[You] can’t [blame] [the program] [for being unable to identify it]
(19.27)
COGNIZER TARGET EVALUEE REASON
[The San Francisco Examiner] issued [a special edition] [yesterday]
(19.28)
ARG 0 TARGET ARG 1 ARGM - TMP
parse ← PARSE(words)
for each predicate in parse do
for each node in parse do
featurevector ← E XTRACT F EATURES(node, predicate, parse)
C LASSIFY N ODE(node, featurevector, parse)
NP-SBJ = ARG0 VP
issued DT JJ NN IN NP
noon yesterday
Figure 19.5 Parse tree for a PropBank sentence, showing the PropBank argument labels. The dotted line
shows the path feature NP↑S↓VP↓VBD for ARG0, the NP-SBJ constituent The San Francisco Examiner.
Global Optimization
The classification algorithm of Fig. 19.5 classifies each argument separately (‘lo-
cally’), making the simplifying assumption that each argument of a predicate can be
labeled independently. This assumption is false; there are interactions between argu-
ments that require a more ‘global’ assignment of labels to constituents. For example,
constituents in FrameNet and PropBank are required to be non-overlapping. More
significantly, the semantic roles of constituents are not independent. For example
PropBank does not allow multiple identical arguments; two constituents of the same
verb cannot both be labeled ARG 0 .
Role labeling systems thus often add a fourth step to deal with global consistency
across the labels in a sentence. For example, the local classifiers can return a list of
possible labels associated with probabilities for each constituent, and a second-pass
Viterbi decoding or re-ranking approach can be used to choose the best consensus
label. Integer linear programming (ILP) is another common way to choose a solution
that conforms best to multiple constraints.
Softmax
concatenate
with predicate
ENCODER
As with all the taggers, the goal is to compute the highest probability tag se-
quence ŷ, given the input sequence of words w:
ŷ = argmax P(y|w)
y∈T
Fig. 19.6 shows a sketch of a standard algorithm from He et al. (2017). Here each
input word is mapped to pretrained embeddings, and then each token is concatenated
with the predicate embedding and then passed through a feedforward network with
a softmax which outputs a distribution over each SRL label. For decoding, a CRF
layer can be used instead of the MLP layer on top of the biLSTM output to do global
inference, but in practice this doesn’t seem to provide much benefit.
Selectional restrictions vary widely in their specificity. The verb imagine, for
example, imposes strict requirements on its AGENT role (restricting it to humans
and other animate entities) but places very few semantic requirements on its THEME
role. A verb like diagonalize, on the other hand, places a very specific constraint
on the filler of its THEME role: it has to be a matrix, while the arguments of the
adjectives odorless are restricted to concepts that could possess an odor:
(19.33) In rehearsal, I often ask the musicians to imagine a tennis game.
(19.34) Radon is an odorless gas that can’t be detected by human senses.
(19.35) To diagonalize a matrix is to find its eigenvalues.
These examples illustrate that the set of concepts we need to represent selectional
restrictions (being a matrix, being able to possess an odor, etc) is quite open ended.
This distinguishes selectional restrictions from other features for representing lexical
knowledge, like parts-of-speech, which are quite limited in number.
With this representation, all we know about y, the filler of the THEME role, is that
it is associated with an Eating event through the Theme relation. To stipulate the
selectional restriction that y must be something edible, we simply add a new term to
that effect:
When a phrase like ate a hamburger is encountered, a semantic analyzer can form
the following kind of representation:
Sense 1
hamburger, beefburger --
(a fried cake of minced beef served on a bun)
=> sandwich
=> snack food
=> dish
=> nutriment, nourishment, nutrition...
=> food, nutrient
=> substance
=> matter
=> physical entity
=> entity
Figure 19.7 Evidence from WordNet that hamburgers are edible.
Selectional Association
selectional
One of the most influential has been the selectional association model of Resnik
preference (1993). Resnik defines the idea of selectional preference strength as the general
strength
amount of information that a predicate tells us about the semantic class of its argu-
ments. For example, the verb eat tells us a lot about the semantic class of its direct
objects, since they tend to be edible. The verb be, by contrast, tells us less about
its direct objects. The selectional preference strength can be defined by the differ-
ence in information between two distributions: the distribution of expected semantic
classes P(c) (how likely is it that a direct object will fall into class c) and the dis-
tribution of expected semantic classes for the particular verb P(c|v) (how likely is
it that the direct object of the specific verb v will fall into semantic class c). The
greater the difference between these distributions, the more information the verb
is giving us about possible objects. The difference between these two distributions
relative entropy can be quantified by relative entropy, or the Kullback-Leibler divergence (Kullback
KL divergence and Leibler, 1951). The Kullback-Leibler or KL divergence D(P||Q) expresses the
difference between two probability distributions P and Q
X P(x)
D(P||Q) = P(x) log (19.38)
x
Q(x)
The selectional preference SR (v) uses the KL divergence to express how much in-
formation, in bits, the verb v expresses about the possible semantic class of its argu-
ment.
SR (v) = D(P(c|v)||P(c))
X P(c|v)
= P(c|v) log (19.39)
c
P(c)
selectional Resnik then defines the selectional association of a particular class and verb as the
association
relative contribution of that class to the general selectional preference of the verb:
1 P(c|v)
AR (v, c) = P(c|v) log (19.40)
SR (v) P(c)
The selectional association is thus a probabilistic measure of the strength of asso-
ciation between a predicate and a class dominating the argument to the predicate.
Resnik estimates the probabilities for these associations by parsing a corpus, count-
ing all the times each predicate occurs with each argument word, and assuming
that each word is a partial observation of all the WordNet concepts containing the
word. The following table from Resnik (1996) shows some sample high and low
selectional associations for verbs and some WordNet semantic classes of their direct
objects.
Direct Object Direct Object
Verb Semantic Class Assoc Semantic Class Assoc
read WRITING 6.80 ACTIVITY -.20
write WRITING 7.26 COMMERCE 0
see ENTITY 5.79 METHOD -0.01
predicate verb, directly modeling the strength of association of one verb (predicate)
with one noun (argument).
The conditional probability model can be computed by parsing a very large cor-
pus (billions of words), and computing co-occurrence counts: how often a given
verb occurs with a given noun in a given relation. The conditional probability of an
argument noun given a verb for a particular relation P(n|v, r) can then be used as a
selectional preference metric for that pair of words (Brockmann and Lapata 2003,
Keller and Lapata 2003):
(
C(n,v,r)
P(n|v, r) = C(v,r) if C(n, v, r) > 0
0 otherwise
The inverse probability P(v|n, r) was found to have better performance in some cases
(Brockmann and Lapata, 2003):
(
C(n,v,r)
P(v|n, r) = C(n,r) if C(n, v, r) > 0
0 otherwise
componential
analysis elements or features, called primitive decomposition or componential analysis,
has been taken even further, and focused particularly on predicates.
Consider these examples of the verb kill:
(19.41) Jim killed his philodendron.
(19.42) Jim did something to cause his philodendron to become not alive.
There is a truth-conditional (‘propositional semantics’) perspective from which these
two sentences have the same meaning. Assuming this equivalence, we could repre-
sent the meaning of kill as:
(19.43) KILL(x,y) ⇔ CAUSE(x, BECOME(NOT(ALIVE(y))))
thus using semantic primitives like do, cause, become not, and alive.
Indeed, one such set of potential semantic primitives has been used to account
for some of the verbal alternations discussed in Section 19.2 (Lakoff 1965, Dowty
1979). Consider the following examples.
(19.44) John opened the door. ⇒ CAUSE(John, BECOME(OPEN(door)))
(19.45) The door opened. ⇒ BECOME(OPEN(door))
(19.46) The door is open. ⇒ OPEN(door)
The decompositional approach asserts that a single state-like predicate associ-
ated with open underlies all of these examples. The differences among the meanings
of these examples arises from the combination of this single predicate with the prim-
itives CAUSE and BECOME.
While this approach to primitive decomposition can explain the similarity be-
tween states and actions or causative and non-causative predicates, it still relies on
having a large number of predicates like open. More radical approaches choose to
break down these predicates as well. One such approach to verbal predicate decom-
position that played a role in early natural language systems is conceptual depen-
conceptual
dependency dency (CD), a set of ten primitive predicates, shown in Fig. 19.8.
Primitive Definition
ATRANS The abstract transfer of possession or control from one entity to
another
P TRANS The physical transfer of an object from one location to another
M TRANS The transfer of mental concepts between entities or within an
entity
M BUILD The creation of new information within an entity
P ROPEL The application of physical force to move an object
M OVE The integral movement of a body part by an animal
I NGEST The taking in of a substance by an animal
E XPEL The expulsion of something from an animal
S PEAK The action of producing a sound
ATTEND The action of focusing a sense organ
Figure 19.8 A set of conceptual dependency primitives.
Below is an example sentence along with its CD representation. The verb brought
is translated into the two primitives ATRANS and PTRANS to indicate that the waiter
both physically conveyed the check to Mary and passed control of it to her. Note
that CD also associates a fixed set of thematic roles with each primitive to represent
the various participants in the action.
(19.47) The waiter brought Mary the check.
18 C HAPTER 19 • S EMANTIC ROLE L ABELING
19.9 Summary
• Semantic roles are abstract models of the role an argument plays in the event
described by the predicate.
• Thematic roles are a model of semantic roles based on a single finite list of
roles. Other semantic role models include per-verb semantic role lists and
proto-agent/proto-patient, both of which are implemented in PropBank,
and per-frame role lists, implemented in FrameNet.
• Semantic role labeling is the task of assigning semantic role labels to the
constituents of a sentence. The task is generally treated as a supervised ma-
chine learning task, with models trained on PropBank or FrameNet. Algo-
rithms generally start by parsing a sentence and then automatically tag each
parse tree node with a semantic role. Neural models map straight from words
end-to-end.
• Semantic selectional restrictions allow words (particularly predicates) to post
constraints on the semantic properties of their argument words. Selectional
preference models (like selectional association or simple conditional proba-
bility) allow a weight or probability to be assigned to the association between
a predicate and an argument word or class.
Each verb then had a set of rules specifying how the parse should be mapped to se-
mantic roles. These rules mainly made reference to grammatical functions (subject,
object, complement of specific prepositions) but also checked constituent internal
features such as the animacy of head nouns. Later systems assigned roles from pre-
built parse trees, again by using dictionaries with verb-specific case frames (Levin
1977, Marcus 1980).
By 1977 case representation was widely used and taught in AI and NLP courses,
and was described as a standard of natural language processing in the first edition of
Winston’s 1977 textbook Artificial Intelligence.
In the 1980s Fillmore proposed his model of frame semantics, later describing
the intuition as follows:
“The idea behind frame semantics is that speakers are aware of possi-
bly quite complex situation types, packages of connected expectations,
that go by various names—frames, schemas, scenarios, scripts, cultural
narratives, memes—and the words in our language are understood with
such frames as their presupposed background.” (Fillmore, 2012, p. 712)
The word frame seemed to be in the air for a suite of related notions proposed at
about the same time by Minsky (1974), Hymes (1974), and Goffman (1974), as
well as related notions with other names like scripts (Schank and Abelson, 1975)
and schemata (Bobrow and Norman, 1975) (see Tannen (1979) for a comparison).
Fillmore was also influenced by the semantic field theorists and by a visit to the Yale
AI lab where he took notice of the lists of slots and fillers used by early information
extraction systems like DeJong (1982) and Schank and Abelson (1977). In the 1990s
Fillmore drew on these insights to begin the FrameNet corpus annotation project.
At the same time, Beth Levin drew on her early case frame dictionaries (Levin,
1977) to develop her book which summarized sets of verb classes defined by shared
argument realizations (Levin, 1993). The VerbNet project built on this work (Kipper
et al., 2000), leading soon afterwards to the PropBank semantic-role-labeled corpus
created by Martha Palmer and colleagues (Palmer et al., 2005).
The combination of rich linguistic annotation and corpus-based approach in-
stantiated in FrameNet and PropBank led to a revival of automatic approaches to
semantic role labeling, first on FrameNet (Gildea and Jurafsky, 2000) and then on
PropBank data (Gildea and Palmer, 2002, inter alia). The problem first addressed in
the 1970s by handwritten rules was thus now generally recast as one of supervised
machine learning enabled by large and consistent databases. Many popular features
used for role labeling are defined in Gildea and Jurafsky (2002), Surdeanu et al.
(2003), Xue and Palmer (2004), Pradhan et al. (2005), Che et al. (2009), and Zhao
et al. (2009). The use of dependency rather than constituency parses was introduced
in the CoNLL-2008 shared task (Surdeanu et al., 2008). For surveys see Palmer
et al. (2010) and Màrquez et al. (2008).
The use of neural approaches to semantic role labeling was pioneered by Col-
lobert et al. (2011), who applied a CRF on top of a convolutional net. Early work
like Foland, Jr. and Martin (2015) focused on using dependency features. Later work
eschewed syntactic features altogether; Zhou and Xu (2015) introduced the use of
a stacked (6-8 layer) biLSTM architecture, and (He et al., 2017) showed how to
augment the biLSTM architecture with highway networks and also replace the CRF
with A* decoding that make it possible to apply a wide variety of global constraints
in SRL decoding.
Most semantic role labeling schemes only work within a single sentence, fo-
cusing on the object of the verbal (or nominal, in the case of NomBank) predicate.
20 C HAPTER 19 • S EMANTIC ROLE L ABELING
However, in many cases, a verbal or nominal predicate may have an implicit argu-
implicit
argument ment: one that appears only in a contextual sentence, or perhaps not at all and must
be inferred. In the two sentences This house has a new owner. The sale was finalized
10 days ago. the sale in the second sentence has no A RG 1, but a reasonable reader
would infer that the Arg1 should be the house mentioned in the prior sentence. Find-
iSRL ing these arguments, implicit argument detection (sometimes shortened as iSRL)
was introduced by Gerber and Chai (2010) and Ruppenhofer et al. (2010). See Do
et al. (2017) for more recent neural models.
To avoid the need for huge labeled training sets, unsupervised approaches for
semantic role labeling attempt to induce the set of semantic roles by clustering over
arguments. The task was pioneered by Riloff and Schmelzenbach (1998) and Swier
and Stevenson (2004); see Grenager and Manning (2006), Titov and Klementiev
(2012), Lang and Lapata (2014), Woodsend and Lapata (2015), and Titov and Khod-
dam (2014).
Recent innovations in frame labeling include connotation frames, which mark
richer information about the argument of predicates. Connotation frames mark the
sentiment of the writer or reader toward the arguments (for example using the verb
survive in he survived a bombing expresses the writer’s sympathy toward the subject
he and negative sentiment toward the bombing. See Chapter 20 for more details.
Selectional preference has been widely studied beyond the selectional associa-
tion models of Resnik (1993) and Resnik (1996). Methods have included clustering
(Rooth et al., 1999), discriminative learning (Bergsma et al., 2008), and topic mod-
els (Séaghdha 2010, Ritter et al. 2010), and constraints can be expressed at the level
of words or classes (Agirre and Martinez, 2001). Selectional preferences have also
been successfully integrated into semantic role labeling (Erk 2007, Zapirain et al.
2013, Do et al. 2017).
Exercises
Exercises 21
Agirre, E. and D. Martinez. 2001. Learning class-to-class Fillmore, C. J. 2012. Encounters with language. Computa-
selectional preferences. CoNLL. tional Linguistics, 38(4):701–718.
Baker, C. F., C. J. Fillmore, and J. B. Lowe. 1998. The Berke- Fillmore, C. J. and C. F. Baker. 2009. A frames approach to
ley FrameNet project. COLING/ACL. semantic analysis. In Bernd Heine and Heiko Narrog, ed-
Bergsma, S., D. Lin, and R. Goebel. 2008. Discriminative itors, The Oxford Handbook of Linguistic Analysis, pages
learning of selectional preference from unlabeled text. 313–340. Oxford University Press.
EMNLP. Fillmore, C. J., C. R. Johnson, and M. R. L. Petruck. 2003.
Bloomfield, L. 1933. Language. University of Chicago Background to FrameNet. International journal of lexi-
Press. cography, 16(3):235–250.
Foland, Jr., W. R. and J. H. Martin. 2015. Dependency-
Bobrow, D. G., R. M. Kaplan, M. Kay, D. A. Norman,
based semantic role labeling using convolutional neural
H. Thompson, and T. Winograd. 1977. GUS, A frame
networks. *SEM 2015.
driven dialog system. Artificial Intelligence, 8:155–173.
Gale, W. A., K. W. Church, and D. Yarowsky. 1992. Work on
Bobrow, D. G. and D. A. Norman. 1975. Some principles
statistical methods for word sense disambiguation. AAAI
of memory schemata. In Daniel G. Bobrow and Allan
Fall Symposium on Probabilistic Approaches to Natural
Collins, editors, Representation and Understanding. Aca-
Language.
demic Press.
Gerber, M. and J. Y. Chai. 2010. Beyond nombank: A study
Brockmann, C. and M. Lapata. 2003. Evaluating and com-
of implicit arguments for nominal predicates. ACL.
bining approaches to selectional preference acquisition.
EACL. Gildea, D. and D. Jurafsky. 2000. Automatic labeling of se-
mantic roles. ACL.
Carreras, X. and L. Màrquez. 2005. Introduction to
the CoNLL-2005 shared task: Semantic role labeling. Gildea, D. and D. Jurafsky. 2002. Automatic labeling of se-
CoNLL. mantic roles. Computational Linguistics, 28(3):245–288.
Chambers, N. and D. Jurafsky. 2010. Improving the use Gildea, D. and M. Palmer. 2002. The necessity of syntactic
of pseudo-words for evaluating selectional preferences. parsing for predicate argument recognition. ACL.
ACL. Goffman, E. 1974. Frame analysis: An essay on the organi-
Che, W., Z. Li, Y. Li, Y. Guo, B. Qin, and T. Liu. 2009. Mul- zation of experience. Harvard University Press.
tilingual dependency-based syntactic and semantic pars- Grenager, T. and C. D. Manning. 2006. Unsupervised dis-
ing. CoNLL. covery of a statistical verb lexicon. EMNLP.
Collobert, R., J. Weston, L. Bottou, M. Karlen, Gruber, J. S. 1965. Studies in Lexical Relations. Ph.D. thesis,
K. Kavukcuoglu, and P. Kuksa. 2011. Natural language MIT.
processing (almost) from scratch. JMLR, 12:2493–2537. He, L., K. Lee, M. Lewis, and L. Zettlemoyer. 2017. Deep
DeJong, G. F. 1982. An overview of the FRUMP system. In semantic role labeling: What works and what’s next.
Wendy G. Lehnert and Martin H. Ringle, editors, Strate- ACL.
gies for Natural Language Processing, pages 149–176. Hendrix, G. G., C. W. Thompson, and J. Slocum. 1973. Lan-
LEA. guage processing via canonical verbs and semantic mod-
Do, Q. N. T., S. Bethard, and M.-F. Moens. 2017. Improv- els. Proceedings of IJCAI-73.
ing implicit semantic role labeling by predicting semantic Hirst, G. 1987. Semantic Interpretation and the Resolution
frame arguments. IJCNLP. of Ambiguity. Cambridge University Press.
Dowty, D. R. 1979. Word Meaning and Montague Grammar. Hymes, D. 1974. Ways of speaking. In Richard Bauman and
D. Reidel. Joel Sherzer, editors, Explorations in the ethnography of
Erk, K. 2007. A simple, similarity-based model for selec- speaking, pages 433–451. Cambridge University Press.
tional preferences. ACL. Johnson-Laird, P. N. 1983. Mental Models. Harvard Univer-
Fillmore, C. J. 1966. A proposal concerning English prepo- sity Press, Cambridge, MA.
sitions. In Francis P. Dinneen, editor, 17th annual Round Katz, J. J. and J. A. Fodor. 1963. The structure of a semantic
Table, volume 17 of Monograph Series on Language and theory. Language, 39:170–210.
Linguistics, pages 19–34. Georgetown University Press. Keller, F. and M. Lapata. 2003. Using the web to obtain fre-
Fillmore, C. J. 1968. The case for case. In Emmon W. Bach quencies for unseen bigrams. Computational Linguistics,
and Robert T. Harms, editors, Universals in Linguistic 29:459–484.
Theory, pages 1–88. Holt, Rinehart & Winston. Kipper, K., H. T. Dang, and M. Palmer. 2000. Class-based
Fillmore, C. J. 1985. Frames and the semantics of under- construction of a verb lexicon. AAAI.
standing. Quaderni di Semantica, VI(2):222–254. Kullback, S. and R. A. Leibler. 1951. On information and
Fillmore, C. J. 2003. Valency and semantic roles: the sufficiency. Annals of Mathematical Statistics, 22:79–86.
concept of deep structure case. In Vilmos Ágel, Lud- Lakoff, G. 1965. On the Nature of Syntactic Irregularity.
wig M. Eichinger, Hans Werner Eroms, Peter Hellwig, Ph.D. thesis, Indiana University. Published as Irregularity
Hans Jürgen Heringer, and Henning Lobin, editors, De- in Syntax. Holt, Rinehart, and Winston, New York, 1970.
pendenz und Valenz: Ein internationales Handbuch der
Lang, J. and M. Lapata. 2014. Similarity-driven semantic
zeitgenössischen Forschung, chapter 36, pages 457–475.
role induction via graph partitioning. Computational Lin-
Walter de Gruyter.
guistics, 40(3):633–669.
22 Chapter 19 • Semantic Role Labeling
Levin, B. 1977. Mapping sentences to case frames. Techni- Schank, R. C. and R. P. Abelson. 1975. Scripts, plans, and
cal Report 167, MIT AI Laboratory. AI Working Paper knowledge. Proceedings of IJCAI-75.
143. Schank, R. C. and R. P. Abelson. 1977. Scripts, Plans, Goals
Levin, B. 1993. English Verb Classes and Alternations: A and Understanding. Lawrence Erlbaum.
Preliminary Investigation. University of Chicago Press. Schütze, H. 1992. Context space. AAAI Fall Symposium on
Levin, B. and M. Rappaport Hovav. 2005. Argument Real- Probabilistic Approaches to Natural Language.
ization. Cambridge University Press. Séaghdha, D. O. 2010. Latent variable models of selectional
Marcus, M. P. 1980. A Theory of Syntactic Recognition for preference. ACL.
Natural Language. MIT Press. Shi, P. and J. Lin. 2019. Simple BERT models for relation
Màrquez, L., X. Carreras, K. C. Litkowski, and S. Steven- extraction and semantic role labeling. ArXiv.
son. 2008. Semantic role labeling: An introduction to the Simmons, R. F. 1973. Semantic networks: Their compu-
special issue. Computational linguistics, 34(2):145–159. tation and use for understanding English sentences. In
Meyers, A., R. Reeves, C. Macleod, R. Szekely, V. Zielinska, Roger C. Schank and Kenneth Mark Colby, editors, Com-
B. Young, and R. Grishman. 2004. The nombank project: puter Models of Thought and Language, pages 61–113.
An interim report. NAACL/HLT Workshop: Frontiers in W.H. Freeman and Co.
Corpus Annotation. Sloan, M. C. 2010. Aristotle’s Nicomachean Ethics as the
Minsky, M. 1974. A framework for representing knowledge. original locus for the Septem Circumstantiae. Classical
Technical Report 306, MIT AI Laboratory. Memo 306. Philology, 105(3):236–251.
Surdeanu, M., S. Harabagiu, J. Williams, and P. Aarseth.
Nash-Webber, B. L. 1975. The role of semantics in auto-
2003. Using predicate-argument structures for informa-
matic speech understanding. In Daniel G. Bobrow and
tion extraction. ACL.
Allan Collins, editors, Representation and Understand-
ing, pages 351–382. Academic Press. Surdeanu, M., R. Johansson, A. Meyers, L. Màrquez, and
J. Nivre. 2008. The CoNLL 2008 shared task on joint
Palmer, M., D. Gildea, and N. Xue. 2010. Semantic role
parsing of syntactic and semantic dependencies. CoNLL.
labeling. Synthesis Lectures on Human Language Tech-
nologies, 3(1):1–103. Swier, R. and S. Stevenson. 2004. Unsupervised semantic
role labelling. EMNLP.
Palmer, M., P. Kingsbury, and D. Gildea. 2005. The proposi-
tion bank: An annotated corpus of semantic roles. Com- Tannen, D. 1979. What’s in a frame? Surface evidence for
putational Linguistics, 31(1):71–106. underlying expectations. In Roy Freedle, editor, New Di-
rections in Discourse Processing, pages 137–181. Ablex.
Penn, G. and P. Kiparsky. 2012. On Pān.ini and the gen-
Tesnière, L. 1959. Éléments de Syntaxe Structurale. Librairie
erative capacity of contextualized replacement systems.
C. Klincksieck, Paris.
COLING.
Titov, I. and E. Khoddam. 2014. Unsupervised induction of
Pradhan, S., A. Moschitti, N. Xue, H. T. Ng, A. Björkelund,
semantic roles within a reconstruction-error minimization
O. Uryupina, Y. Zhang, and Z. Zhong. 2013. Towards
framework. NAACL HLT.
robust linguistic analysis using OntoNotes. CoNLL.
Titov, I. and A. Klementiev. 2012. A Bayesian approach to
Pradhan, S., W. Ward, K. Hacioglu, J. H. Martin, and D. Ju-
unsupervised semantic role induction. EACL.
rafsky. 2005. Semantic role labeling using different syn-
tactic views. ACL. Wilks, Y. 1973. An artificial intelligence approach to ma-
chine translation. In Roger C. Schank and Kenneth Mark
Resnik, P. 1993. Semantic classes and syntactic ambiguity. Colby, editors, Computer Models of Thought and Lan-
Proceedings of the workshop on Human Language Tech- guage, pages 114–151. W.H. Freeman.
nology.
Wilks, Y. 1975a. Preference semantics. In Edward L.
Resnik, P. 1996. Selectional constraints: An information- Keenan, editor, The Formal Semantics of Natural Lan-
theoretic model and its computational realization. Cogni- guage, pages 329–350. Cambridge Univ. Press.
tion, 61:127–159.
Wilks, Y. 1975b. A preferential, pattern-seeking, seman-
Riloff, E. and M. Schmelzenbach. 1998. An empirical ap- tics for natural language inference. Artificial Intelligence,
proach to conceptual case frame acquisition. Proceedings 6(1):53–74.
of the Sixth Workshop on Very Large Corpora.
Winston, P. H. 1977. Artificial Intelligence. Addison Wesley.
Ritter, A., O. Etzioni, and Mausam. 2010. A latent dirichlet Woodsend, K. and M. Lapata. 2015. Distributed representa-
allocation method for selectional preferences. ACL. tions for unsupervised semantic role labeling. EMNLP.
Rooth, M., S. Riezler, D. Prescher, G. Carroll, and F. Beil. Xue, N. and M. Palmer. 2004. Calibrating features for se-
1999. Inducing a semantically annotated lexicon via EM- mantic role labeling. EMNLP.
based clustering. ACL.
Zapirain, B., E. Agirre, L. Màrquez, and M. Surdeanu. 2013.
Ruppenhofer, J., M. Ellsworth, M. R. L. Petruck, C. R. John- Selectional preferences for semantic role classification.
son, C. F. Baker, and J. Scheffczyk. 2016. FrameNet II: Computational Linguistics, 39(3):631–663.
Extended theory and practice.
Zhao, H., W. Chen, C. Kit, and G. Zhou. 2009. Multilingual
Ruppenhofer, J., C. Sporleder, R. Morante, C. F. Baker, dependency learning: A huge feature engineering method
and M. Palmer. 2010. Semeval-2010 task 10: Linking to semantic dependency parsing. CoNLL.
events and their participants in discourse. 5th Interna-
Zhou, J. and W. Xu. 2015. End-to-end learning of semantic
tional Workshop on Semantic Evaluation.
role labeling using recurrent neural networks. ACL.